Merge pull request #1221 from borglab/hybrid/hybrid-factor-graph
commit
ff20a50163
|
|
@ -219,6 +219,16 @@ class DiscreteBayesTree {
|
||||||
};
|
};
|
||||||
|
|
||||||
#include <gtsam/discrete/DiscreteLookupDAG.h>
|
#include <gtsam/discrete/DiscreteLookupDAG.h>
|
||||||
|
|
||||||
|
class DiscreteLookupTable : gtsam::DiscreteConditional{
|
||||||
|
DiscreteLookupTable(size_t nFrontals, const gtsam::DiscreteKeys& keys,
|
||||||
|
const gtsam::DecisionTreeFactor::ADT& potentials);
|
||||||
|
void print(
|
||||||
|
const std::string& s = "Discrete Lookup Table: ",
|
||||||
|
const gtsam::KeyFormatter& keyFormatter = gtsam::DefaultKeyFormatter) const;
|
||||||
|
size_t argmax(const gtsam::DiscreteValues& parentsValues) const;
|
||||||
|
};
|
||||||
|
|
||||||
class DiscreteLookupDAG {
|
class DiscreteLookupDAG {
|
||||||
DiscreteLookupDAG();
|
DiscreteLookupDAG();
|
||||||
void push_back(const gtsam::DiscreteLookupTable* table);
|
void push_back(const gtsam::DiscreteLookupTable* table);
|
||||||
|
|
|
||||||
|
|
@ -119,11 +119,36 @@ void GaussianMixture::print(const std::string &s,
|
||||||
"", [&](Key k) { return formatter(k); },
|
"", [&](Key k) { return formatter(k); },
|
||||||
[&](const GaussianConditional::shared_ptr &gf) -> std::string {
|
[&](const GaussianConditional::shared_ptr &gf) -> std::string {
|
||||||
RedirectCout rd;
|
RedirectCout rd;
|
||||||
if (!gf->empty())
|
if (gf && !gf->empty())
|
||||||
gf->print("", formatter);
|
gf->print("", formatter);
|
||||||
else
|
else
|
||||||
return {"nullptr"};
|
return {"nullptr"};
|
||||||
return rd.str();
|
return rd.str();
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* *******************************************************************************/
|
||||||
|
void GaussianMixture::prune(const DecisionTreeFactor &decisionTree) {
|
||||||
|
// Functional which loops over all assignments and create a set of
|
||||||
|
// GaussianConditionals
|
||||||
|
auto pruner = [&decisionTree](
|
||||||
|
const Assignment<Key> &choices,
|
||||||
|
const GaussianConditional::shared_ptr &conditional)
|
||||||
|
-> GaussianConditional::shared_ptr {
|
||||||
|
// typecast so we can use this to get probability value
|
||||||
|
DiscreteValues values(choices);
|
||||||
|
|
||||||
|
if (decisionTree(values) == 0.0) {
|
||||||
|
// empty aka null pointer
|
||||||
|
boost::shared_ptr<GaussianConditional> null;
|
||||||
|
return null;
|
||||||
|
} else {
|
||||||
|
return conditional;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
auto pruned_conditionals = conditionals_.apply(pruner);
|
||||||
|
conditionals_.root_ = pruned_conditionals.root_;
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace gtsam
|
} // namespace gtsam
|
||||||
|
|
|
||||||
|
|
@ -21,6 +21,7 @@
|
||||||
|
|
||||||
#include <gtsam/discrete/DecisionTree-inl.h>
|
#include <gtsam/discrete/DecisionTree-inl.h>
|
||||||
#include <gtsam/discrete/DecisionTree.h>
|
#include <gtsam/discrete/DecisionTree.h>
|
||||||
|
#include <gtsam/discrete/DecisionTreeFactor.h>
|
||||||
#include <gtsam/discrete/DiscreteKey.h>
|
#include <gtsam/discrete/DiscreteKey.h>
|
||||||
#include <gtsam/hybrid/HybridFactor.h>
|
#include <gtsam/hybrid/HybridFactor.h>
|
||||||
#include <gtsam/inference/Conditional.h>
|
#include <gtsam/inference/Conditional.h>
|
||||||
|
|
@ -30,11 +31,14 @@ namespace gtsam {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @brief A conditional of gaussian mixtures indexed by discrete variables, as
|
* @brief A conditional of gaussian mixtures indexed by discrete variables, as
|
||||||
* part of a Bayes Network.
|
* part of a Bayes Network. This is the result of the elimination of a
|
||||||
|
* continuous variable in a hybrid scheme, such that the remaining variables are
|
||||||
|
* discrete+continuous.
|
||||||
*
|
*
|
||||||
* Represents the conditional density P(X | M, Z) where X is a continuous random
|
* Represents the conditional density P(X | M, Z) where X is the set of
|
||||||
* variable, M is the selection of discrete variables corresponding to a subset
|
* continuous random variables, M is the selection of discrete variables
|
||||||
* of the Gaussian variables and Z is parent of this node
|
* corresponding to a subset of the Gaussian variables and Z is parent of this
|
||||||
|
* node .
|
||||||
*
|
*
|
||||||
* The probability P(x|y,z,...) is proportional to
|
* The probability P(x|y,z,...) is proportional to
|
||||||
* \f$ \sum_i k_i \exp - \frac{1}{2} |R_i x - (d_i - S_i y - T_i z - ...)|^2 \f$
|
* \f$ \sum_i k_i \exp - \frac{1}{2} |R_i x - (d_i - S_i y - T_i z - ...)|^2 \f$
|
||||||
|
|
@ -118,7 +122,7 @@ class GTSAM_EXPORT GaussianMixture
|
||||||
/// Test equality with base HybridFactor
|
/// Test equality with base HybridFactor
|
||||||
bool equals(const HybridFactor &lf, double tol = 1e-9) const override;
|
bool equals(const HybridFactor &lf, double tol = 1e-9) const override;
|
||||||
|
|
||||||
/* print utility */
|
/// Print utility
|
||||||
void print(
|
void print(
|
||||||
const std::string &s = "GaussianMixture\n",
|
const std::string &s = "GaussianMixture\n",
|
||||||
const KeyFormatter &formatter = DefaultKeyFormatter) const override;
|
const KeyFormatter &formatter = DefaultKeyFormatter) const override;
|
||||||
|
|
@ -128,6 +132,15 @@ class GTSAM_EXPORT GaussianMixture
|
||||||
/// Getter for the underlying Conditionals DecisionTree
|
/// Getter for the underlying Conditionals DecisionTree
|
||||||
const Conditionals &conditionals();
|
const Conditionals &conditionals();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Prune the decision tree of Gaussian factors as per the discrete
|
||||||
|
* `decisionTree`.
|
||||||
|
*
|
||||||
|
* @param decisionTree A pruned decision tree of discrete keys where the
|
||||||
|
* leaves are probabilities.
|
||||||
|
*/
|
||||||
|
void prune(const DecisionTreeFactor &decisionTree);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @brief Merge the Gaussian Factor Graphs in `this` and `sum` while
|
* @brief Merge the Gaussian Factor Graphs in `this` and `sum` while
|
||||||
* maintaining the decision tree structure.
|
* maintaining the decision tree structure.
|
||||||
|
|
|
||||||
|
|
@ -51,7 +51,7 @@ GaussianMixtureFactor GaussianMixtureFactor::FromFactors(
|
||||||
void GaussianMixtureFactor::print(const std::string &s,
|
void GaussianMixtureFactor::print(const std::string &s,
|
||||||
const KeyFormatter &formatter) const {
|
const KeyFormatter &formatter) const {
|
||||||
HybridFactor::print(s, formatter);
|
HybridFactor::print(s, formatter);
|
||||||
std::cout << "]{\n";
|
std::cout << "{\n";
|
||||||
factors_.print(
|
factors_.print(
|
||||||
"", [&](Key k) { return formatter(k); },
|
"", [&](Key k) { return formatter(k); },
|
||||||
[&](const GaussianFactor::shared_ptr &gf) -> std::string {
|
[&](const GaussianFactor::shared_ptr &gf) -> std::string {
|
||||||
|
|
|
||||||
|
|
@ -22,7 +22,7 @@
|
||||||
|
|
||||||
#include <gtsam/discrete/DecisionTree.h>
|
#include <gtsam/discrete/DecisionTree.h>
|
||||||
#include <gtsam/discrete/DiscreteKey.h>
|
#include <gtsam/discrete/DiscreteKey.h>
|
||||||
#include <gtsam/hybrid/HybridFactor.h>
|
#include <gtsam/hybrid/HybridGaussianFactor.h>
|
||||||
#include <gtsam/linear/GaussianFactor.h>
|
#include <gtsam/linear/GaussianFactor.h>
|
||||||
|
|
||||||
namespace gtsam {
|
namespace gtsam {
|
||||||
|
|
@ -108,7 +108,7 @@ class GTSAM_EXPORT GaussianMixtureFactor : public HybridFactor {
|
||||||
bool equals(const HybridFactor &lf, double tol = 1e-9) const override;
|
bool equals(const HybridFactor &lf, double tol = 1e-9) const override;
|
||||||
|
|
||||||
void print(
|
void print(
|
||||||
const std::string &s = "HybridFactor\n",
|
const std::string &s = "GaussianMixtureFactor\n",
|
||||||
const KeyFormatter &formatter = DefaultKeyFormatter) const override;
|
const KeyFormatter &formatter = DefaultKeyFormatter) const override;
|
||||||
/// @}
|
/// @}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -10,7 +10,132 @@
|
||||||
* @file HybridBayesNet.cpp
|
* @file HybridBayesNet.cpp
|
||||||
* @brief A bayes net of Gaussian Conditionals indexed by discrete keys.
|
* @brief A bayes net of Gaussian Conditionals indexed by discrete keys.
|
||||||
* @author Fan Jiang
|
* @author Fan Jiang
|
||||||
|
* @author Varun Agrawal
|
||||||
|
* @author Shangjie Xue
|
||||||
* @date January 2022
|
* @date January 2022
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#include <gtsam/hybrid/HybridBayesNet.h>
|
#include <gtsam/hybrid/HybridBayesNet.h>
|
||||||
|
#include <gtsam/hybrid/HybridValues.h>
|
||||||
|
#include <gtsam/hybrid/HybridLookupDAG.h>
|
||||||
|
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
/// Return the DiscreteKey vector as a set.
|
||||||
|
static std::set<DiscreteKey> DiscreteKeysAsSet(const DiscreteKeys &dkeys) {
|
||||||
|
std::set<DiscreteKey> s;
|
||||||
|
s.insert(dkeys.begin(), dkeys.end());
|
||||||
|
return s;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
HybridBayesNet HybridBayesNet::prune(
|
||||||
|
const DecisionTreeFactor::shared_ptr &discreteFactor) const {
|
||||||
|
/* To Prune, we visitWith every leaf in the GaussianMixture.
|
||||||
|
* For each leaf, using the assignment we can check the discrete decision tree
|
||||||
|
* for 0.0 probability, then just set the leaf to a nullptr.
|
||||||
|
*
|
||||||
|
* We can later check the GaussianMixture for just nullptrs.
|
||||||
|
*/
|
||||||
|
|
||||||
|
HybridBayesNet prunedBayesNetFragment;
|
||||||
|
|
||||||
|
// Functional which loops over all assignments and create a set of
|
||||||
|
// GaussianConditionals
|
||||||
|
auto pruner = [&](const Assignment<Key> &choices,
|
||||||
|
const GaussianConditional::shared_ptr &conditional)
|
||||||
|
-> GaussianConditional::shared_ptr {
|
||||||
|
// typecast so we can use this to get probability value
|
||||||
|
DiscreteValues values(choices);
|
||||||
|
|
||||||
|
if ((*discreteFactor)(values) == 0.0) {
|
||||||
|
// empty aka null pointer
|
||||||
|
boost::shared_ptr<GaussianConditional> null;
|
||||||
|
return null;
|
||||||
|
} else {
|
||||||
|
return conditional;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
// Go through all the conditionals in the
|
||||||
|
// Bayes Net and prune them as per discreteFactor.
|
||||||
|
for (size_t i = 0; i < this->size(); i++) {
|
||||||
|
HybridConditional::shared_ptr conditional = this->at(i);
|
||||||
|
|
||||||
|
GaussianMixture::shared_ptr gaussianMixture =
|
||||||
|
boost::dynamic_pointer_cast<GaussianMixture>(conditional->inner());
|
||||||
|
|
||||||
|
if (gaussianMixture) {
|
||||||
|
// We may have mixtures with less discrete keys than discreteFactor so we
|
||||||
|
// skip those since the label assignment does not exist.
|
||||||
|
auto gmKeySet = DiscreteKeysAsSet(gaussianMixture->discreteKeys());
|
||||||
|
auto dfKeySet = DiscreteKeysAsSet(discreteFactor->discreteKeys());
|
||||||
|
if (gmKeySet != dfKeySet) {
|
||||||
|
// Add the gaussianMixture which doesn't have to be pruned.
|
||||||
|
prunedBayesNetFragment.push_back(
|
||||||
|
boost::make_shared<HybridConditional>(gaussianMixture));
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Run the pruning to get a new, pruned tree
|
||||||
|
GaussianMixture::Conditionals prunedTree =
|
||||||
|
gaussianMixture->conditionals().apply(pruner);
|
||||||
|
|
||||||
|
DiscreteKeys discreteKeys = gaussianMixture->discreteKeys();
|
||||||
|
// reverse keys to get a natural ordering
|
||||||
|
std::reverse(discreteKeys.begin(), discreteKeys.end());
|
||||||
|
|
||||||
|
// Convert from boost::iterator_range to KeyVector
|
||||||
|
// so we can pass it to constructor.
|
||||||
|
KeyVector frontals(gaussianMixture->frontals().begin(),
|
||||||
|
gaussianMixture->frontals().end()),
|
||||||
|
parents(gaussianMixture->parents().begin(),
|
||||||
|
gaussianMixture->parents().end());
|
||||||
|
|
||||||
|
// Create the new gaussian mixture and add it to the bayes net.
|
||||||
|
auto prunedGaussianMixture = boost::make_shared<GaussianMixture>(
|
||||||
|
frontals, parents, discreteKeys, prunedTree);
|
||||||
|
|
||||||
|
// Type-erase and add to the pruned Bayes Net fragment.
|
||||||
|
prunedBayesNetFragment.push_back(
|
||||||
|
boost::make_shared<HybridConditional>(prunedGaussianMixture));
|
||||||
|
|
||||||
|
} else {
|
||||||
|
// Add the non-GaussianMixture conditional
|
||||||
|
prunedBayesNetFragment.push_back(conditional);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return prunedBayesNetFragment;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
GaussianMixture::shared_ptr HybridBayesNet::atGaussian(size_t i) const {
|
||||||
|
return boost::dynamic_pointer_cast<GaussianMixture>(factors_.at(i)->inner());
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
DiscreteConditional::shared_ptr HybridBayesNet::atDiscrete(size_t i) const {
|
||||||
|
return boost::dynamic_pointer_cast<DiscreteConditional>(
|
||||||
|
factors_.at(i)->inner());
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
GaussianBayesNet HybridBayesNet::choose(
|
||||||
|
const DiscreteValues &assignment) const {
|
||||||
|
GaussianBayesNet gbn;
|
||||||
|
for (size_t idx = 0; idx < size(); idx++) {
|
||||||
|
GaussianMixture gm = *this->atGaussian(idx);
|
||||||
|
gbn.push_back(gm(assignment));
|
||||||
|
}
|
||||||
|
return gbn;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* *******************************************************************************/
|
||||||
|
HybridValues HybridBayesNet::optimize() const {
|
||||||
|
auto dag = HybridLookupDAG::FromBayesNet(*this);
|
||||||
|
return dag.argmax();
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
|
||||||
|
|
@ -17,8 +17,11 @@
|
||||||
|
|
||||||
#pragma once
|
#pragma once
|
||||||
|
|
||||||
|
#include <gtsam/discrete/DecisionTreeFactor.h>
|
||||||
#include <gtsam/hybrid/HybridConditional.h>
|
#include <gtsam/hybrid/HybridConditional.h>
|
||||||
|
#include <gtsam/hybrid/HybridValues.h>
|
||||||
#include <gtsam/inference/BayesNet.h>
|
#include <gtsam/inference/BayesNet.h>
|
||||||
|
#include <gtsam/linear/GaussianBayesNet.h>
|
||||||
|
|
||||||
namespace gtsam {
|
namespace gtsam {
|
||||||
|
|
||||||
|
|
@ -36,6 +39,39 @@ class GTSAM_EXPORT HybridBayesNet : public BayesNet<HybridConditional> {
|
||||||
|
|
||||||
/** Construct empty bayes net */
|
/** Construct empty bayes net */
|
||||||
HybridBayesNet() = default;
|
HybridBayesNet() = default;
|
||||||
|
|
||||||
|
/// Prune the Hybrid Bayes Net given the discrete decision tree.
|
||||||
|
HybridBayesNet prune(
|
||||||
|
const DecisionTreeFactor::shared_ptr &discreteFactor) const;
|
||||||
|
|
||||||
|
/// Add HybridConditional to Bayes Net
|
||||||
|
using Base::add;
|
||||||
|
|
||||||
|
/// Add a discrete conditional to the Bayes Net.
|
||||||
|
void add(const DiscreteKey &key, const std::string &table) {
|
||||||
|
push_back(
|
||||||
|
HybridConditional(boost::make_shared<DiscreteConditional>(key, table)));
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get a specific Gaussian mixture by index `i`.
|
||||||
|
GaussianMixture::shared_ptr atGaussian(size_t i) const;
|
||||||
|
|
||||||
|
/// Get a specific discrete conditional by index `i`.
|
||||||
|
DiscreteConditional::shared_ptr atDiscrete(size_t i) const;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Get the Gaussian Bayes Net which corresponds to a specific discrete
|
||||||
|
* value assignment.
|
||||||
|
*
|
||||||
|
* @param assignment The discrete value assignment for the discrete keys.
|
||||||
|
* @return GaussianBayesNet
|
||||||
|
*/
|
||||||
|
GaussianBayesNet choose(const DiscreteValues &assignment) const;
|
||||||
|
|
||||||
|
/// Solve the HybridBayesNet by back-substitution.
|
||||||
|
/// TODO(Shangjie) do we need to create a HybridGaussianBayesNet class, and
|
||||||
|
/// put this method there?
|
||||||
|
HybridValues optimize() const;
|
||||||
};
|
};
|
||||||
|
|
||||||
} // namespace gtsam
|
} // namespace gtsam
|
||||||
|
|
|
||||||
|
|
@ -76,7 +76,10 @@ class GTSAM_EXPORT HybridBayesTree : public BayesTree<HybridBayesTreeClique> {
|
||||||
/**
|
/**
|
||||||
* @brief Class for Hybrid Bayes tree orphan subtrees.
|
* @brief Class for Hybrid Bayes tree orphan subtrees.
|
||||||
*
|
*
|
||||||
* This does special stuff for the hybrid case
|
* This object stores parent keys in our base type factor so that
|
||||||
|
* eliminating those parent keys will pull this subtree into the
|
||||||
|
* elimination.
|
||||||
|
* This does special stuff for the hybrid case.
|
||||||
*
|
*
|
||||||
* @tparam CLIQUE
|
* @tparam CLIQUE
|
||||||
*/
|
*/
|
||||||
|
|
|
||||||
|
|
@ -78,7 +78,8 @@ bool HybridFactor::equals(const HybridFactor &lf, double tol) const {
|
||||||
const This *e = dynamic_cast<const This *>(&lf);
|
const This *e = dynamic_cast<const This *>(&lf);
|
||||||
return e != nullptr && Base::equals(*e, tol) &&
|
return e != nullptr && Base::equals(*e, tol) &&
|
||||||
isDiscrete_ == e->isDiscrete_ && isContinuous_ == e->isContinuous_ &&
|
isDiscrete_ == e->isDiscrete_ && isContinuous_ == e->isContinuous_ &&
|
||||||
isHybrid_ == e->isHybrid_ && nrContinuous_ == e->nrContinuous_;
|
isHybrid_ == e->isHybrid_ && continuousKeys_ == e->continuousKeys_ &&
|
||||||
|
discreteKeys_ == e->discreteKeys_;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************ */
|
/* ************************************************************************ */
|
||||||
|
|
@ -88,17 +89,23 @@ void HybridFactor::print(const std::string &s,
|
||||||
if (isContinuous_) std::cout << "Continuous ";
|
if (isContinuous_) std::cout << "Continuous ";
|
||||||
if (isDiscrete_) std::cout << "Discrete ";
|
if (isDiscrete_) std::cout << "Discrete ";
|
||||||
if (isHybrid_) std::cout << "Hybrid ";
|
if (isHybrid_) std::cout << "Hybrid ";
|
||||||
for (size_t c=0; c<continuousKeys_.size(); c++) {
|
std::cout << "[";
|
||||||
|
for (size_t c = 0; c < continuousKeys_.size(); c++) {
|
||||||
std::cout << formatter(continuousKeys_.at(c));
|
std::cout << formatter(continuousKeys_.at(c));
|
||||||
if (c < continuousKeys_.size() - 1) {
|
if (c < continuousKeys_.size() - 1) {
|
||||||
std::cout << " ";
|
std::cout << " ";
|
||||||
} else {
|
} else {
|
||||||
std::cout << "; ";
|
std::cout << (discreteKeys_.size() > 0 ? "; " : "");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for(auto && discreteKey: discreteKeys_) {
|
for (size_t d = 0; d < discreteKeys_.size(); d++) {
|
||||||
std::cout << formatter(discreteKey.first) << " ";
|
std::cout << formatter(discreteKeys_.at(d).first);
|
||||||
|
if (d < discreteKeys_.size() - 1) {
|
||||||
|
std::cout << " ";
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
std::cout << "]";
|
||||||
}
|
}
|
||||||
|
|
||||||
} // namespace gtsam
|
} // namespace gtsam
|
||||||
|
|
|
||||||
|
|
@ -50,8 +50,8 @@ class GTSAM_EXPORT HybridFactor : public Factor {
|
||||||
size_t nrContinuous_ = 0;
|
size_t nrContinuous_ = 0;
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
|
// Set of DiscreteKeys for this factor.
|
||||||
DiscreteKeys discreteKeys_;
|
DiscreteKeys discreteKeys_;
|
||||||
|
|
||||||
/// Record continuous keys for book-keeping
|
/// Record continuous keys for book-keeping
|
||||||
KeyVector continuousKeys_;
|
KeyVector continuousKeys_;
|
||||||
|
|
||||||
|
|
@ -120,10 +120,13 @@ class GTSAM_EXPORT HybridFactor : public Factor {
|
||||||
bool isHybrid() const { return isHybrid_; }
|
bool isHybrid() const { return isHybrid_; }
|
||||||
|
|
||||||
/// Return the number of continuous variables in this factor.
|
/// Return the number of continuous variables in this factor.
|
||||||
size_t nrContinuous() const { return nrContinuous_; }
|
size_t nrContinuous() const { return continuousKeys_.size(); }
|
||||||
|
|
||||||
/// Return vector of discrete keys.
|
/// Return the discrete keys for this factor.
|
||||||
DiscreteKeys discreteKeys() const { return discreteKeys_; }
|
const DiscreteKeys &discreteKeys() const { return discreteKeys_; }
|
||||||
|
|
||||||
|
/// Return only the continuous keys for this factor.
|
||||||
|
const KeyVector &continuousKeys() const { return continuousKeys_; }
|
||||||
|
|
||||||
/// @}
|
/// @}
|
||||||
};
|
};
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,140 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file HybridFactorGraph.h
|
||||||
|
* @brief Hybrid factor graph base class that uses type erasure
|
||||||
|
* @author Varun Agrawal
|
||||||
|
* @date May 28, 2022
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <gtsam/discrete/DiscreteFactor.h>
|
||||||
|
#include <gtsam/hybrid/HybridDiscreteFactor.h>
|
||||||
|
#include <gtsam/hybrid/HybridFactor.h>
|
||||||
|
#include <gtsam/inference/FactorGraph.h>
|
||||||
|
#include <gtsam/inference/Ordering.h>
|
||||||
|
|
||||||
|
#include <boost/format.hpp>
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
using SharedFactor = boost::shared_ptr<Factor>;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Hybrid Factor Graph
|
||||||
|
* -----------------------
|
||||||
|
* This is the base hybrid factor graph.
|
||||||
|
* Everything inside needs to be hybrid factor or hybrid conditional.
|
||||||
|
*/
|
||||||
|
class HybridFactorGraph : public FactorGraph<HybridFactor> {
|
||||||
|
public:
|
||||||
|
using Base = FactorGraph<HybridFactor>;
|
||||||
|
using This = HybridFactorGraph; ///< this class
|
||||||
|
using shared_ptr = boost::shared_ptr<This>; ///< shared_ptr to This
|
||||||
|
|
||||||
|
using Values = gtsam::Values; ///< backwards compatibility
|
||||||
|
using Indices = KeyVector; ///> map from keys to values
|
||||||
|
|
||||||
|
protected:
|
||||||
|
/// Check if FACTOR type is derived from DiscreteFactor.
|
||||||
|
template <typename FACTOR>
|
||||||
|
using IsDiscrete = typename std::enable_if<
|
||||||
|
std::is_base_of<DiscreteFactor, FACTOR>::value>::type;
|
||||||
|
|
||||||
|
/// Check if FACTOR type is derived from HybridFactor.
|
||||||
|
template <typename FACTOR>
|
||||||
|
using IsHybrid = typename std::enable_if<
|
||||||
|
std::is_base_of<HybridFactor, FACTOR>::value>::type;
|
||||||
|
|
||||||
|
public:
|
||||||
|
/// @name Constructors
|
||||||
|
/// @{
|
||||||
|
|
||||||
|
/// Default constructor
|
||||||
|
HybridFactorGraph() = default;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Implicit copy/downcast constructor to override explicit template container
|
||||||
|
* constructor. In BayesTree this is used for:
|
||||||
|
* `cachedSeparatorMarginal_.reset(*separatorMarginal)`
|
||||||
|
* */
|
||||||
|
template <class DERIVEDFACTOR>
|
||||||
|
HybridFactorGraph(const FactorGraph<DERIVEDFACTOR>& graph) : Base(graph) {}
|
||||||
|
|
||||||
|
/// @}
|
||||||
|
|
||||||
|
// Allow use of selected FactorGraph methods:
|
||||||
|
using Base::empty;
|
||||||
|
using Base::reserve;
|
||||||
|
using Base::size;
|
||||||
|
using Base::operator[];
|
||||||
|
using Base::add;
|
||||||
|
using Base::push_back;
|
||||||
|
using Base::resize;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Add a discrete factor *pointer* to the internal discrete graph
|
||||||
|
* @param discreteFactor - boost::shared_ptr to the factor to add
|
||||||
|
*/
|
||||||
|
template <typename FACTOR>
|
||||||
|
IsDiscrete<FACTOR> push_discrete(
|
||||||
|
const boost::shared_ptr<FACTOR>& discreteFactor) {
|
||||||
|
Base::push_back(boost::make_shared<HybridDiscreteFactor>(discreteFactor));
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Add a discrete-continuous (Hybrid) factor *pointer* to the graph
|
||||||
|
* @param hybridFactor - boost::shared_ptr to the factor to add
|
||||||
|
*/
|
||||||
|
template <typename FACTOR>
|
||||||
|
IsHybrid<FACTOR> push_hybrid(const boost::shared_ptr<FACTOR>& hybridFactor) {
|
||||||
|
Base::push_back(hybridFactor);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// delete emplace_shared.
|
||||||
|
template <class FACTOR, class... Args>
|
||||||
|
void emplace_shared(Args&&... args) = delete;
|
||||||
|
|
||||||
|
/// Construct a factor and add (shared pointer to it) to factor graph.
|
||||||
|
template <class FACTOR, class... Args>
|
||||||
|
IsDiscrete<FACTOR> emplace_discrete(Args&&... args) {
|
||||||
|
auto factor = boost::allocate_shared<FACTOR>(
|
||||||
|
Eigen::aligned_allocator<FACTOR>(), std::forward<Args>(args)...);
|
||||||
|
push_discrete(factor);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Construct a factor and add (shared pointer to it) to factor graph.
|
||||||
|
template <class FACTOR, class... Args>
|
||||||
|
IsHybrid<FACTOR> emplace_hybrid(Args&&... args) {
|
||||||
|
auto factor = boost::allocate_shared<FACTOR>(
|
||||||
|
Eigen::aligned_allocator<FACTOR>(), std::forward<Args>(args)...);
|
||||||
|
push_hybrid(factor);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Add a single factor shared pointer to the hybrid factor graph.
|
||||||
|
* Dynamically handles the factor type and assigns it to the correct
|
||||||
|
* underlying container.
|
||||||
|
*
|
||||||
|
* @param sharedFactor The factor to add to this factor graph.
|
||||||
|
*/
|
||||||
|
void push_back(const SharedFactor& sharedFactor) {
|
||||||
|
if (auto p = boost::dynamic_pointer_cast<DiscreteFactor>(sharedFactor)) {
|
||||||
|
push_discrete(p);
|
||||||
|
}
|
||||||
|
if (auto p = boost::dynamic_pointer_cast<HybridFactor>(sharedFactor)) {
|
||||||
|
push_hybrid(p);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -37,6 +37,8 @@ class GTSAM_EXPORT HybridGaussianFactor : public HybridFactor {
|
||||||
using This = HybridGaussianFactor;
|
using This = HybridGaussianFactor;
|
||||||
using shared_ptr = boost::shared_ptr<This>;
|
using shared_ptr = boost::shared_ptr<This>;
|
||||||
|
|
||||||
|
HybridGaussianFactor() = default;
|
||||||
|
|
||||||
// Explicit conversion from a shared ptr of GF
|
// Explicit conversion from a shared ptr of GF
|
||||||
explicit HybridGaussianFactor(GaussianFactor::shared_ptr other);
|
explicit HybridGaussianFactor(GaussianFactor::shared_ptr other);
|
||||||
|
|
||||||
|
|
@ -52,7 +54,7 @@ class GTSAM_EXPORT HybridGaussianFactor : public HybridFactor {
|
||||||
|
|
||||||
/// GTSAM print utility.
|
/// GTSAM print utility.
|
||||||
void print(
|
void print(
|
||||||
const std::string &s = "HybridFactor\n",
|
const std::string &s = "HybridGaussianFactor\n",
|
||||||
const KeyFormatter &formatter = DefaultKeyFormatter) const override;
|
const KeyFormatter &formatter = DefaultKeyFormatter) const override;
|
||||||
|
|
||||||
/// @}
|
/// @}
|
||||||
|
|
|
||||||
|
|
@ -98,6 +98,12 @@ GaussianMixtureFactor::Sum sumFrontals(
|
||||||
} else if (f->isContinuous()) {
|
} else if (f->isContinuous()) {
|
||||||
deferredFactors.push_back(
|
deferredFactors.push_back(
|
||||||
boost::dynamic_pointer_cast<HybridGaussianFactor>(f)->inner());
|
boost::dynamic_pointer_cast<HybridGaussianFactor>(f)->inner());
|
||||||
|
|
||||||
|
} else if (f->isDiscrete()) {
|
||||||
|
// Don't do anything for discrete-only factors
|
||||||
|
// since we want to eliminate continuous values only.
|
||||||
|
continue;
|
||||||
|
|
||||||
} else {
|
} else {
|
||||||
// We need to handle the case where the object is actually an
|
// We need to handle the case where the object is actually an
|
||||||
// BayesTreeOrphanWrapper!
|
// BayesTreeOrphanWrapper!
|
||||||
|
|
@ -106,8 +112,8 @@ GaussianMixtureFactor::Sum sumFrontals(
|
||||||
if (!orphan) {
|
if (!orphan) {
|
||||||
auto &fr = *f;
|
auto &fr = *f;
|
||||||
throw std::invalid_argument(
|
throw std::invalid_argument(
|
||||||
std::string("factor is discrete in continuous elimination") +
|
std::string("factor is discrete in continuous elimination ") +
|
||||||
typeid(fr).name());
|
demangle(typeid(fr).name()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
@ -158,7 +164,7 @@ discreteElimination(const HybridGaussianFactorGraph &factors,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
auto result = EliminateDiscrete(dfg, frontalKeys);
|
auto result = EliminateForMPE(dfg, frontalKeys);
|
||||||
|
|
||||||
return {boost::make_shared<HybridConditional>(result.first),
|
return {boost::make_shared<HybridConditional>(result.first),
|
||||||
boost::make_shared<HybridDiscreteFactor>(result.second)};
|
boost::make_shared<HybridDiscreteFactor>(result.second)};
|
||||||
|
|
@ -178,6 +184,19 @@ hybridElimination(const HybridGaussianFactorGraph &factors,
|
||||||
// sum out frontals, this is the factor on the separator
|
// sum out frontals, this is the factor on the separator
|
||||||
GaussianMixtureFactor::Sum sum = sumFrontals(factors);
|
GaussianMixtureFactor::Sum sum = sumFrontals(factors);
|
||||||
|
|
||||||
|
// If a tree leaf contains nullptr,
|
||||||
|
// convert that leaf to an empty GaussianFactorGraph.
|
||||||
|
// Needed since the DecisionTree will otherwise create
|
||||||
|
// a GFG with a single (null) factor.
|
||||||
|
auto emptyGaussian = [](const GaussianFactorGraph &gfg) {
|
||||||
|
bool hasNull =
|
||||||
|
std::any_of(gfg.begin(), gfg.end(),
|
||||||
|
[](const GaussianFactor::shared_ptr &ptr) { return !ptr; });
|
||||||
|
|
||||||
|
return hasNull ? GaussianFactorGraph() : gfg;
|
||||||
|
};
|
||||||
|
sum = GaussianMixtureFactor::Sum(sum, emptyGaussian);
|
||||||
|
|
||||||
using EliminationPair = GaussianFactorGraph::EliminationResult;
|
using EliminationPair = GaussianFactorGraph::EliminationResult;
|
||||||
|
|
||||||
KeyVector keysOfEliminated; // Not the ordering
|
KeyVector keysOfEliminated; // Not the ordering
|
||||||
|
|
@ -189,7 +208,10 @@ hybridElimination(const HybridGaussianFactorGraph &factors,
|
||||||
if (graph.empty()) {
|
if (graph.empty()) {
|
||||||
return {nullptr, nullptr};
|
return {nullptr, nullptr};
|
||||||
}
|
}
|
||||||
auto result = EliminatePreferCholesky(graph, frontalKeys);
|
std::pair<boost::shared_ptr<GaussianConditional>,
|
||||||
|
boost::shared_ptr<GaussianFactor>>
|
||||||
|
result = EliminatePreferCholesky(graph, frontalKeys);
|
||||||
|
|
||||||
if (keysOfEliminated.empty()) {
|
if (keysOfEliminated.empty()) {
|
||||||
keysOfEliminated =
|
keysOfEliminated =
|
||||||
result.first->keys(); // Initialize the keysOfEliminated to be the
|
result.first->keys(); // Initialize the keysOfEliminated to be the
|
||||||
|
|
@ -229,14 +251,27 @@ hybridElimination(const HybridGaussianFactorGraph &factors,
|
||||||
boost::make_shared<HybridDiscreteFactor>(discreteFactor)};
|
boost::make_shared<HybridDiscreteFactor>(discreteFactor)};
|
||||||
|
|
||||||
} else {
|
} else {
|
||||||
// Create a resulting DCGaussianMixture on the separator.
|
// Create a resulting GaussianMixtureFactor on the separator.
|
||||||
auto factor = boost::make_shared<GaussianMixtureFactor>(
|
auto factor = boost::make_shared<GaussianMixtureFactor>(
|
||||||
KeyVector(continuousSeparator.begin(), continuousSeparator.end()),
|
KeyVector(continuousSeparator.begin(), continuousSeparator.end()),
|
||||||
discreteSeparator, separatorFactors);
|
discreteSeparator, separatorFactors);
|
||||||
return {boost::make_shared<HybridConditional>(conditional), factor};
|
return {boost::make_shared<HybridConditional>(conditional), factor};
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/* ************************************************************************ */
|
/* ************************************************************************
|
||||||
|
* Function to eliminate variables **under the following assumptions**:
|
||||||
|
* 1. When the ordering is fully continuous, and the graph only contains
|
||||||
|
* continuous and hybrid factors
|
||||||
|
* 2. When the ordering is fully discrete, and the graph only contains discrete
|
||||||
|
* factors
|
||||||
|
*
|
||||||
|
* Any usage outside of this is considered incorrect.
|
||||||
|
*
|
||||||
|
* \warning This function is not meant to be used with arbitrary hybrid factor
|
||||||
|
* graphs. For example, if there exists continuous parents, and one tries to
|
||||||
|
* eliminate a discrete variable (as specified in the ordering), the result will
|
||||||
|
* be INCORRECT and there will be NO error raised.
|
||||||
|
*/
|
||||||
std::pair<HybridConditional::shared_ptr, HybridFactor::shared_ptr> //
|
std::pair<HybridConditional::shared_ptr, HybridFactor::shared_ptr> //
|
||||||
EliminateHybrid(const HybridGaussianFactorGraph &factors,
|
EliminateHybrid(const HybridGaussianFactorGraph &factors,
|
||||||
const Ordering &frontalKeys) {
|
const Ordering &frontalKeys) {
|
||||||
|
|
|
||||||
|
|
@ -19,9 +19,12 @@
|
||||||
#pragma once
|
#pragma once
|
||||||
|
|
||||||
#include <gtsam/hybrid/HybridFactor.h>
|
#include <gtsam/hybrid/HybridFactor.h>
|
||||||
|
#include <gtsam/hybrid/HybridFactorGraph.h>
|
||||||
|
#include <gtsam/hybrid/HybridGaussianFactor.h>
|
||||||
#include <gtsam/inference/EliminateableFactorGraph.h>
|
#include <gtsam/inference/EliminateableFactorGraph.h>
|
||||||
#include <gtsam/inference/FactorGraph.h>
|
#include <gtsam/inference/FactorGraph.h>
|
||||||
#include <gtsam/inference/Ordering.h>
|
#include <gtsam/inference/Ordering.h>
|
||||||
|
#include <gtsam/linear/GaussianFactor.h>
|
||||||
|
|
||||||
namespace gtsam {
|
namespace gtsam {
|
||||||
|
|
||||||
|
|
@ -53,10 +56,9 @@ struct EliminationTraits<HybridGaussianFactorGraph> {
|
||||||
typedef HybridBayesNet
|
typedef HybridBayesNet
|
||||||
BayesNetType; ///< Type of Bayes net from sequential elimination
|
BayesNetType; ///< Type of Bayes net from sequential elimination
|
||||||
typedef HybridEliminationTree
|
typedef HybridEliminationTree
|
||||||
EliminationTreeType; ///< Type of elimination tree
|
EliminationTreeType; ///< Type of elimination tree
|
||||||
typedef HybridBayesTree BayesTreeType; ///< Type of Bayes tree
|
typedef HybridBayesTree BayesTreeType; ///< Type of Bayes tree
|
||||||
typedef HybridJunctionTree
|
typedef HybridJunctionTree JunctionTreeType; ///< Type of Junction tree
|
||||||
JunctionTreeType; ///< Type of Junction tree
|
|
||||||
/// The default dense elimination function
|
/// The default dense elimination function
|
||||||
static std::pair<boost::shared_ptr<ConditionalType>,
|
static std::pair<boost::shared_ptr<ConditionalType>,
|
||||||
boost::shared_ptr<FactorType> >
|
boost::shared_ptr<FactorType> >
|
||||||
|
|
@ -72,10 +74,16 @@ struct EliminationTraits<HybridGaussianFactorGraph> {
|
||||||
* Everything inside needs to be hybrid factor or hybrid conditional.
|
* Everything inside needs to be hybrid factor or hybrid conditional.
|
||||||
*/
|
*/
|
||||||
class GTSAM_EXPORT HybridGaussianFactorGraph
|
class GTSAM_EXPORT HybridGaussianFactorGraph
|
||||||
: public FactorGraph<HybridFactor>,
|
: public HybridFactorGraph,
|
||||||
public EliminateableFactorGraph<HybridGaussianFactorGraph> {
|
public EliminateableFactorGraph<HybridGaussianFactorGraph> {
|
||||||
|
protected:
|
||||||
|
/// Check if FACTOR type is derived from GaussianFactor.
|
||||||
|
template <typename FACTOR>
|
||||||
|
using IsGaussian = typename std::enable_if<
|
||||||
|
std::is_base_of<GaussianFactor, FACTOR>::value>::type;
|
||||||
|
|
||||||
public:
|
public:
|
||||||
using Base = FactorGraph<HybridFactor>;
|
using Base = HybridFactorGraph;
|
||||||
using This = HybridGaussianFactorGraph; ///< this class
|
using This = HybridGaussianFactorGraph; ///< this class
|
||||||
using BaseEliminateable =
|
using BaseEliminateable =
|
||||||
EliminateableFactorGraph<This>; ///< for elimination
|
EliminateableFactorGraph<This>; ///< for elimination
|
||||||
|
|
@ -100,7 +108,13 @@ class GTSAM_EXPORT HybridGaussianFactorGraph
|
||||||
|
|
||||||
/// @}
|
/// @}
|
||||||
|
|
||||||
using FactorGraph::add;
|
using Base::empty;
|
||||||
|
using Base::reserve;
|
||||||
|
using Base::size;
|
||||||
|
using Base::operator[];
|
||||||
|
using Base::add;
|
||||||
|
using Base::push_back;
|
||||||
|
using Base::resize;
|
||||||
|
|
||||||
/// Add a Jacobian factor to the factor graph.
|
/// Add a Jacobian factor to the factor graph.
|
||||||
void add(JacobianFactor&& factor);
|
void add(JacobianFactor&& factor);
|
||||||
|
|
@ -113,6 +127,39 @@ class GTSAM_EXPORT HybridGaussianFactorGraph
|
||||||
|
|
||||||
/// Add a DecisionTreeFactor as a shared ptr.
|
/// Add a DecisionTreeFactor as a shared ptr.
|
||||||
void add(boost::shared_ptr<DecisionTreeFactor> factor);
|
void add(boost::shared_ptr<DecisionTreeFactor> factor);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Add a gaussian factor *pointer* to the internal gaussian factor graph
|
||||||
|
* @param gaussianFactor - boost::shared_ptr to the factor to add
|
||||||
|
*/
|
||||||
|
template <typename FACTOR>
|
||||||
|
IsGaussian<FACTOR> push_gaussian(
|
||||||
|
const boost::shared_ptr<FACTOR>& gaussianFactor) {
|
||||||
|
Base::push_back(boost::make_shared<HybridGaussianFactor>(gaussianFactor));
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Construct a factor and add (shared pointer to it) to factor graph.
|
||||||
|
template <class FACTOR, class... Args>
|
||||||
|
IsGaussian<FACTOR> emplace_gaussian(Args&&... args) {
|
||||||
|
auto factor = boost::allocate_shared<FACTOR>(
|
||||||
|
Eigen::aligned_allocator<FACTOR>(), std::forward<Args>(args)...);
|
||||||
|
push_gaussian(factor);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Add a single factor shared pointer to the hybrid factor graph.
|
||||||
|
* Dynamically handles the factor type and assigns it to the correct
|
||||||
|
* underlying container.
|
||||||
|
*
|
||||||
|
* @param sharedFactor The factor to add to this factor graph.
|
||||||
|
*/
|
||||||
|
void push_back(const SharedFactor& sharedFactor) {
|
||||||
|
if (auto p = boost::dynamic_pointer_cast<GaussianFactor>(sharedFactor)) {
|
||||||
|
push_gaussian(p);
|
||||||
|
} else {
|
||||||
|
Base::push_back(sharedFactor);
|
||||||
|
}
|
||||||
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
} // namespace gtsam
|
} // namespace gtsam
|
||||||
|
|
|
||||||
|
|
@ -41,6 +41,7 @@ HybridGaussianISAM::HybridGaussianISAM(const HybridBayesTree& bayesTree)
|
||||||
void HybridGaussianISAM::updateInternal(
|
void HybridGaussianISAM::updateInternal(
|
||||||
const HybridGaussianFactorGraph& newFactors,
|
const HybridGaussianFactorGraph& newFactors,
|
||||||
HybridBayesTree::Cliques* orphans,
|
HybridBayesTree::Cliques* orphans,
|
||||||
|
const boost::optional<Ordering>& ordering,
|
||||||
const HybridBayesTree::Eliminate& function) {
|
const HybridBayesTree::Eliminate& function) {
|
||||||
// Remove the contaminated part of the Bayes tree
|
// Remove the contaminated part of the Bayes tree
|
||||||
BayesNetType bn;
|
BayesNetType bn;
|
||||||
|
|
@ -74,16 +75,21 @@ void HybridGaussianISAM::updateInternal(
|
||||||
std::copy(allDiscrete.begin(), allDiscrete.end(),
|
std::copy(allDiscrete.begin(), allDiscrete.end(),
|
||||||
std::back_inserter(newKeysDiscreteLast));
|
std::back_inserter(newKeysDiscreteLast));
|
||||||
|
|
||||||
// KeyVector new
|
|
||||||
|
|
||||||
// Get an ordering where the new keys are eliminated last
|
// Get an ordering where the new keys are eliminated last
|
||||||
const VariableIndex index(factors);
|
const VariableIndex index(factors);
|
||||||
const Ordering ordering = Ordering::ColamdConstrainedLast(
|
Ordering elimination_ordering;
|
||||||
index, KeyVector(newKeysDiscreteLast.begin(), newKeysDiscreteLast.end()),
|
if (ordering) {
|
||||||
true);
|
elimination_ordering = *ordering;
|
||||||
|
} else {
|
||||||
|
elimination_ordering = Ordering::ColamdConstrainedLast(
|
||||||
|
index,
|
||||||
|
KeyVector(newKeysDiscreteLast.begin(), newKeysDiscreteLast.end()),
|
||||||
|
true);
|
||||||
|
}
|
||||||
|
|
||||||
// eliminate all factors (top, added, orphans) into a new Bayes tree
|
// eliminate all factors (top, added, orphans) into a new Bayes tree
|
||||||
auto bayesTree = factors.eliminateMultifrontal(ordering, function, index);
|
HybridBayesTree::shared_ptr bayesTree =
|
||||||
|
factors.eliminateMultifrontal(elimination_ordering, function, index);
|
||||||
|
|
||||||
// Re-add into Bayes tree data structures
|
// Re-add into Bayes tree data structures
|
||||||
this->roots_.insert(this->roots_.end(), bayesTree->roots().begin(),
|
this->roots_.insert(this->roots_.end(), bayesTree->roots().begin(),
|
||||||
|
|
@ -93,9 +99,61 @@ void HybridGaussianISAM::updateInternal(
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
void HybridGaussianISAM::update(const HybridGaussianFactorGraph& newFactors,
|
void HybridGaussianISAM::update(const HybridGaussianFactorGraph& newFactors,
|
||||||
|
const boost::optional<Ordering>& ordering,
|
||||||
const HybridBayesTree::Eliminate& function) {
|
const HybridBayesTree::Eliminate& function) {
|
||||||
Cliques orphans;
|
Cliques orphans;
|
||||||
this->updateInternal(newFactors, &orphans, function);
|
this->updateInternal(newFactors, &orphans, ordering, function);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
/**
|
||||||
|
* @brief Check if `b` is a subset of `a`.
|
||||||
|
* Non-const since they need to be sorted.
|
||||||
|
*
|
||||||
|
* @param a KeyVector
|
||||||
|
* @param b KeyVector
|
||||||
|
* @return True if the keys of b is a subset of a, else false.
|
||||||
|
*/
|
||||||
|
bool IsSubset(KeyVector a, KeyVector b) {
|
||||||
|
std::sort(a.begin(), a.end());
|
||||||
|
std::sort(b.begin(), b.end());
|
||||||
|
return std::includes(a.begin(), a.end(), b.begin(), b.end());
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
void HybridGaussianISAM::prune(const Key& root, const size_t maxNrLeaves) {
|
||||||
|
auto decisionTree = boost::dynamic_pointer_cast<DecisionTreeFactor>(
|
||||||
|
this->clique(root)->conditional()->inner());
|
||||||
|
DecisionTreeFactor prunedDiscreteFactor = decisionTree->prune(maxNrLeaves);
|
||||||
|
decisionTree->root_ = prunedDiscreteFactor.root_;
|
||||||
|
|
||||||
|
std::vector<gtsam::Key> prunedKeys;
|
||||||
|
for (auto&& clique : nodes()) {
|
||||||
|
// The cliques can be repeated for each frontal so we record it in
|
||||||
|
// prunedKeys and check if we have already pruned a particular clique.
|
||||||
|
if (std::find(prunedKeys.begin(), prunedKeys.end(), clique.first) !=
|
||||||
|
prunedKeys.end()) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add all the keys of the current clique to be pruned to prunedKeys
|
||||||
|
for (auto&& key : clique.second->conditional()->frontals()) {
|
||||||
|
prunedKeys.push_back(key);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Convert parents() to a KeyVector for comparison
|
||||||
|
KeyVector parents;
|
||||||
|
for (auto&& parent : clique.second->conditional()->parents()) {
|
||||||
|
parents.push_back(parent);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (IsSubset(parents, decisionTree->keys())) {
|
||||||
|
auto gaussianMixture = boost::dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
clique.second->conditional()->inner());
|
||||||
|
|
||||||
|
gaussianMixture->prune(prunedDiscreteFactor);
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
} // namespace gtsam
|
} // namespace gtsam
|
||||||
|
|
|
||||||
|
|
@ -48,6 +48,7 @@ class GTSAM_EXPORT HybridGaussianISAM : public ISAM<HybridBayesTree> {
|
||||||
void updateInternal(
|
void updateInternal(
|
||||||
const HybridGaussianFactorGraph& newFactors,
|
const HybridGaussianFactorGraph& newFactors,
|
||||||
HybridBayesTree::Cliques* orphans,
|
HybridBayesTree::Cliques* orphans,
|
||||||
|
const boost::optional<Ordering>& ordering = boost::none,
|
||||||
const HybridBayesTree::Eliminate& function =
|
const HybridBayesTree::Eliminate& function =
|
||||||
HybridBayesTree::EliminationTraitsType::DefaultEliminate);
|
HybridBayesTree::EliminationTraitsType::DefaultEliminate);
|
||||||
|
|
||||||
|
|
@ -59,8 +60,17 @@ class GTSAM_EXPORT HybridGaussianISAM : public ISAM<HybridBayesTree> {
|
||||||
* @param function Elimination function.
|
* @param function Elimination function.
|
||||||
*/
|
*/
|
||||||
void update(const HybridGaussianFactorGraph& newFactors,
|
void update(const HybridGaussianFactorGraph& newFactors,
|
||||||
|
const boost::optional<Ordering>& ordering = boost::none,
|
||||||
const HybridBayesTree::Eliminate& function =
|
const HybridBayesTree::Eliminate& function =
|
||||||
HybridBayesTree::EliminationTraitsType::DefaultEliminate);
|
HybridBayesTree::EliminationTraitsType::DefaultEliminate);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief
|
||||||
|
*
|
||||||
|
* @param root The root key in the discrete conditional decision tree.
|
||||||
|
* @param maxNumberLeaves
|
||||||
|
*/
|
||||||
|
void prune(const Key& root, const size_t maxNumberLeaves);
|
||||||
};
|
};
|
||||||
|
|
||||||
/// traits
|
/// traits
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,76 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file DiscreteLookupDAG.cpp
|
||||||
|
* @date Aug, 2022
|
||||||
|
* @author Shangjie Xue
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <gtsam/discrete/DiscreteBayesNet.h>
|
||||||
|
#include <gtsam/discrete/DiscreteLookupDAG.h>
|
||||||
|
#include <gtsam/discrete/DiscreteValues.h>
|
||||||
|
#include <gtsam/hybrid/HybridBayesNet.h>
|
||||||
|
#include <gtsam/hybrid/HybridConditional.h>
|
||||||
|
#include <gtsam/hybrid/HybridLookupDAG.h>
|
||||||
|
#include <gtsam/hybrid/HybridValues.h>
|
||||||
|
#include <gtsam/linear/VectorValues.h>
|
||||||
|
|
||||||
|
#include <string>
|
||||||
|
#include <utility>
|
||||||
|
|
||||||
|
using std::pair;
|
||||||
|
using std::vector;
|
||||||
|
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/* ************************************************************************** */
|
||||||
|
void HybridLookupTable::argmaxInPlace(HybridValues* values) const {
|
||||||
|
// For discrete conditional, uses argmaxInPlace() method in
|
||||||
|
// DiscreteLookupTable.
|
||||||
|
if (isDiscrete()) {
|
||||||
|
boost::static_pointer_cast<DiscreteLookupTable>(inner_)->argmaxInPlace(
|
||||||
|
&(values->discrete));
|
||||||
|
} else if (isContinuous()) {
|
||||||
|
// For Gaussian conditional, uses solve() method in GaussianConditional.
|
||||||
|
values->continuous.insert(
|
||||||
|
boost::static_pointer_cast<GaussianConditional>(inner_)->solve(
|
||||||
|
values->continuous));
|
||||||
|
} else if (isHybrid()) {
|
||||||
|
// For hybrid conditional, since children should not contain discrete
|
||||||
|
// variable, we can condition on the discrete variable in the parents and
|
||||||
|
// solve the resulting GaussianConditional.
|
||||||
|
auto conditional =
|
||||||
|
boost::static_pointer_cast<GaussianMixture>(inner_)->conditionals()(
|
||||||
|
values->discrete);
|
||||||
|
values->continuous.insert(conditional->solve(values->continuous));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************** */
|
||||||
|
HybridLookupDAG HybridLookupDAG::FromBayesNet(const HybridBayesNet& bayesNet) {
|
||||||
|
HybridLookupDAG dag;
|
||||||
|
for (auto&& conditional : bayesNet) {
|
||||||
|
HybridLookupTable hlt(*conditional);
|
||||||
|
dag.push_back(hlt);
|
||||||
|
}
|
||||||
|
return dag;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************** */
|
||||||
|
HybridValues HybridLookupDAG::argmax(HybridValues result) const {
|
||||||
|
// Argmax each node in turn in topological sort order (parents first).
|
||||||
|
for (auto lookupTable : boost::adaptors::reverse(*this))
|
||||||
|
lookupTable->argmaxInPlace(&result);
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -0,0 +1,119 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file HybridLookupDAG.h
|
||||||
|
* @date Aug, 2022
|
||||||
|
* @author Shangjie Xue
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <gtsam/discrete/DiscreteDistribution.h>
|
||||||
|
#include <gtsam/discrete/DiscreteLookupDAG.h>
|
||||||
|
#include <gtsam/hybrid/HybridConditional.h>
|
||||||
|
#include <gtsam/hybrid/HybridValues.h>
|
||||||
|
#include <gtsam/inference/BayesNet.h>
|
||||||
|
#include <gtsam/inference/FactorGraph.h>
|
||||||
|
|
||||||
|
#include <boost/shared_ptr.hpp>
|
||||||
|
#include <string>
|
||||||
|
#include <utility>
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief HybridLookupTable table for max-product
|
||||||
|
*
|
||||||
|
* Similar to DiscreteLookupTable, inherits from hybrid conditional for
|
||||||
|
* convenience. Is used in the max-product algorithm.
|
||||||
|
*/
|
||||||
|
class GTSAM_EXPORT HybridLookupTable : public HybridConditional {
|
||||||
|
public:
|
||||||
|
using Base = HybridConditional;
|
||||||
|
using This = HybridLookupTable;
|
||||||
|
using shared_ptr = boost::shared_ptr<This>;
|
||||||
|
using BaseConditional = Conditional<DecisionTreeFactor, This>;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Construct a new Hybrid Lookup Table object form a HybridConditional.
|
||||||
|
*
|
||||||
|
* @param conditional input hybrid conditional
|
||||||
|
*/
|
||||||
|
HybridLookupTable(HybridConditional& conditional) : Base(conditional){};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Calculate assignment for frontal variables that maximizes value.
|
||||||
|
* @param (in/out) parentsValues Known assignments for the parents.
|
||||||
|
*/
|
||||||
|
void argmaxInPlace(HybridValues* parentsValues) const;
|
||||||
|
};
|
||||||
|
|
||||||
|
/** A DAG made from hybrid lookup tables, as defined above. Similar to
|
||||||
|
* DiscreteLookupDAG */
|
||||||
|
class GTSAM_EXPORT HybridLookupDAG : public BayesNet<HybridLookupTable> {
|
||||||
|
public:
|
||||||
|
using Base = BayesNet<HybridLookupTable>;
|
||||||
|
using This = HybridLookupDAG;
|
||||||
|
using shared_ptr = boost::shared_ptr<This>;
|
||||||
|
|
||||||
|
/// @name Standard Constructors
|
||||||
|
/// @{
|
||||||
|
|
||||||
|
/// Construct empty DAG.
|
||||||
|
HybridLookupDAG() {}
|
||||||
|
|
||||||
|
/// Create from BayesNet with LookupTables
|
||||||
|
static HybridLookupDAG FromBayesNet(const HybridBayesNet& bayesNet);
|
||||||
|
|
||||||
|
/// Destructor
|
||||||
|
virtual ~HybridLookupDAG() {}
|
||||||
|
|
||||||
|
/// @}
|
||||||
|
|
||||||
|
/// @name Standard Interface
|
||||||
|
/// @{
|
||||||
|
|
||||||
|
/** Add a DiscreteLookupTable */
|
||||||
|
template <typename... Args>
|
||||||
|
void add(Args&&... args) {
|
||||||
|
emplace_shared<HybridLookupTable>(std::forward<Args>(args)...);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief argmax by back-substitution, optionally given certain variables.
|
||||||
|
*
|
||||||
|
* Assumes the DAG is reverse topologically sorted, i.e. last
|
||||||
|
* conditional will be optimized first *and* that the
|
||||||
|
* DAG does not contain any conditionals for the given variables. If the DAG
|
||||||
|
* resulted from eliminating a factor graph, this is true for the elimination
|
||||||
|
* ordering.
|
||||||
|
*
|
||||||
|
* @return given assignment extended w. optimal assignment for all variables.
|
||||||
|
*/
|
||||||
|
HybridValues argmax(HybridValues given = HybridValues()) const;
|
||||||
|
/// @}
|
||||||
|
|
||||||
|
private:
|
||||||
|
/** Serialization function */
|
||||||
|
friend class boost::serialization::access;
|
||||||
|
template <class ARCHIVE>
|
||||||
|
void serialize(ARCHIVE& ar, const unsigned int /*version*/) {
|
||||||
|
ar& BOOST_SERIALIZATION_BASE_OBJECT_NVP(Base);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
// traits
|
||||||
|
template <>
|
||||||
|
struct traits<HybridLookupDAG> : public Testable<HybridLookupDAG> {};
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -0,0 +1,41 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file HybridNonlinearFactor.cpp
|
||||||
|
* @date May 28, 2022
|
||||||
|
* @author Varun Agrawal
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <gtsam/hybrid/HybridNonlinearFactor.h>
|
||||||
|
|
||||||
|
#include <boost/make_shared.hpp>
|
||||||
|
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
HybridNonlinearFactor::HybridNonlinearFactor(
|
||||||
|
const NonlinearFactor::shared_ptr &other)
|
||||||
|
: Base(other->keys()), inner_(other) {}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
bool HybridNonlinearFactor::equals(const HybridFactor &lf, double tol) const {
|
||||||
|
return Base::equals(lf, tol);
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
void HybridNonlinearFactor::print(const std::string &s,
|
||||||
|
const KeyFormatter &formatter) const {
|
||||||
|
HybridFactor::print(s, formatter);
|
||||||
|
inner_->print("\n", formatter);
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -0,0 +1,62 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file HybridNonlinearFactor.h
|
||||||
|
* @date May 28, 2022
|
||||||
|
* @author Varun Agrawal
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <gtsam/hybrid/HybridFactor.h>
|
||||||
|
#include <gtsam/hybrid/HybridGaussianFactor.h>
|
||||||
|
#include <gtsam/nonlinear/NonlinearFactor.h>
|
||||||
|
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A HybridNonlinearFactor is a layer over NonlinearFactor so that we do not
|
||||||
|
* have a diamond inheritance.
|
||||||
|
*/
|
||||||
|
class HybridNonlinearFactor : public HybridFactor {
|
||||||
|
private:
|
||||||
|
NonlinearFactor::shared_ptr inner_;
|
||||||
|
|
||||||
|
public:
|
||||||
|
using Base = HybridFactor;
|
||||||
|
using This = HybridNonlinearFactor;
|
||||||
|
using shared_ptr = boost::shared_ptr<This>;
|
||||||
|
|
||||||
|
// Explicit conversion from a shared ptr of NonlinearFactor
|
||||||
|
explicit HybridNonlinearFactor(const NonlinearFactor::shared_ptr &other);
|
||||||
|
|
||||||
|
/// @name Testable
|
||||||
|
/// @{
|
||||||
|
|
||||||
|
/// Check equality.
|
||||||
|
virtual bool equals(const HybridFactor &lf, double tol) const override;
|
||||||
|
|
||||||
|
/// GTSAM print utility.
|
||||||
|
void print(
|
||||||
|
const std::string &s = "HybridNonlinearFactor\n",
|
||||||
|
const KeyFormatter &formatter = DefaultKeyFormatter) const override;
|
||||||
|
|
||||||
|
/// @}
|
||||||
|
|
||||||
|
NonlinearFactor::shared_ptr inner() const { return inner_; }
|
||||||
|
|
||||||
|
/// Linearize to a HybridGaussianFactor at the linearization point `c`.
|
||||||
|
boost::shared_ptr<HybridGaussianFactor> linearize(const Values &c) const {
|
||||||
|
return boost::make_shared<HybridGaussianFactor>(inner_->linearize(c));
|
||||||
|
}
|
||||||
|
};
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -0,0 +1,94 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file HybridNonlinearFactorGraph.cpp
|
||||||
|
* @brief Nonlinear hybrid factor graph that uses type erasure
|
||||||
|
* @author Varun Agrawal
|
||||||
|
* @date May 28, 2022
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <gtsam/hybrid/HybridNonlinearFactorGraph.h>
|
||||||
|
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
void HybridNonlinearFactorGraph::add(
|
||||||
|
boost::shared_ptr<NonlinearFactor> factor) {
|
||||||
|
FactorGraph::add(boost::make_shared<HybridNonlinearFactor>(factor));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
void HybridNonlinearFactorGraph::print(const std::string& s,
|
||||||
|
const KeyFormatter& keyFormatter) const {
|
||||||
|
// Base::print(str, keyFormatter);
|
||||||
|
std::cout << (s.empty() ? "" : s + " ") << std::endl;
|
||||||
|
std::cout << "size: " << size() << std::endl;
|
||||||
|
for (size_t i = 0; i < factors_.size(); i++) {
|
||||||
|
std::stringstream ss;
|
||||||
|
ss << "factor " << i << ": ";
|
||||||
|
if (factors_[i]) {
|
||||||
|
factors_[i]->print(ss.str(), keyFormatter);
|
||||||
|
std::cout << std::endl;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
HybridGaussianFactorGraph HybridNonlinearFactorGraph::linearize(
|
||||||
|
const Values& continuousValues) const {
|
||||||
|
// create an empty linear FG
|
||||||
|
HybridGaussianFactorGraph linearFG;
|
||||||
|
|
||||||
|
linearFG.reserve(size());
|
||||||
|
|
||||||
|
// linearize all hybrid factors
|
||||||
|
for (auto&& factor : factors_) {
|
||||||
|
// First check if it is a valid factor
|
||||||
|
if (factor) {
|
||||||
|
// Check if the factor is a hybrid factor.
|
||||||
|
// It can be either a nonlinear MixtureFactor or a linear
|
||||||
|
// GaussianMixtureFactor.
|
||||||
|
if (factor->isHybrid()) {
|
||||||
|
// Check if it is a nonlinear mixture factor
|
||||||
|
if (auto nlmf = boost::dynamic_pointer_cast<MixtureFactor>(factor)) {
|
||||||
|
linearFG.push_back(nlmf->linearize(continuousValues));
|
||||||
|
} else {
|
||||||
|
linearFG.push_back(factor);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now check if the factor is a continuous only factor.
|
||||||
|
} else if (factor->isContinuous()) {
|
||||||
|
// In this case, we check if factor->inner() is nonlinear since
|
||||||
|
// HybridFactors wrap over continuous factors.
|
||||||
|
auto nlhf = boost::dynamic_pointer_cast<HybridNonlinearFactor>(factor);
|
||||||
|
if (auto nlf =
|
||||||
|
boost::dynamic_pointer_cast<NonlinearFactor>(nlhf->inner())) {
|
||||||
|
auto hgf = boost::make_shared<HybridGaussianFactor>(
|
||||||
|
nlf->linearize(continuousValues));
|
||||||
|
linearFG.push_back(hgf);
|
||||||
|
} else {
|
||||||
|
linearFG.push_back(factor);
|
||||||
|
}
|
||||||
|
// Finally if nothing else, we are discrete-only which doesn't need
|
||||||
|
// lineariztion.
|
||||||
|
} else {
|
||||||
|
linearFG.push_back(factor);
|
||||||
|
}
|
||||||
|
|
||||||
|
} else {
|
||||||
|
linearFG.push_back(GaussianFactor::shared_ptr());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return linearFG;
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -0,0 +1,134 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file HybridNonlinearFactorGraph.h
|
||||||
|
* @brief Nonlinear hybrid factor graph that uses type erasure
|
||||||
|
* @author Varun Agrawal
|
||||||
|
* @date May 28, 2022
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <gtsam/hybrid/HybridFactor.h>
|
||||||
|
#include <gtsam/hybrid/HybridFactorGraph.h>
|
||||||
|
#include <gtsam/hybrid/HybridGaussianFactorGraph.h>
|
||||||
|
#include <gtsam/hybrid/HybridNonlinearFactor.h>
|
||||||
|
#include <gtsam/hybrid/MixtureFactor.h>
|
||||||
|
#include <gtsam/inference/Ordering.h>
|
||||||
|
#include <gtsam/nonlinear/NonlinearFactor.h>
|
||||||
|
|
||||||
|
#include <boost/format.hpp>
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Nonlinear Hybrid Factor Graph
|
||||||
|
* -----------------------
|
||||||
|
* This is the non-linear version of a hybrid factor graph.
|
||||||
|
* Everything inside needs to be hybrid factor or hybrid conditional.
|
||||||
|
*/
|
||||||
|
class GTSAM_EXPORT HybridNonlinearFactorGraph : public HybridFactorGraph {
|
||||||
|
protected:
|
||||||
|
/// Check if FACTOR type is derived from NonlinearFactor.
|
||||||
|
template <typename FACTOR>
|
||||||
|
using IsNonlinear = typename std::enable_if<
|
||||||
|
std::is_base_of<NonlinearFactor, FACTOR>::value>::type;
|
||||||
|
|
||||||
|
public:
|
||||||
|
using Base = HybridFactorGraph;
|
||||||
|
using This = HybridNonlinearFactorGraph; ///< this class
|
||||||
|
using shared_ptr = boost::shared_ptr<This>; ///< shared_ptr to This
|
||||||
|
|
||||||
|
using Values = gtsam::Values; ///< backwards compatibility
|
||||||
|
using Indices = KeyVector; ///> map from keys to values
|
||||||
|
|
||||||
|
/// @name Constructors
|
||||||
|
/// @{
|
||||||
|
|
||||||
|
HybridNonlinearFactorGraph() = default;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Implicit copy/downcast constructor to override explicit template container
|
||||||
|
* constructor. In BayesTree this is used for:
|
||||||
|
* `cachedSeparatorMarginal_.reset(*separatorMarginal)`
|
||||||
|
* */
|
||||||
|
template <class DERIVEDFACTOR>
|
||||||
|
HybridNonlinearFactorGraph(const FactorGraph<DERIVEDFACTOR>& graph)
|
||||||
|
: Base(graph) {}
|
||||||
|
|
||||||
|
/// @}
|
||||||
|
|
||||||
|
// Allow use of selected FactorGraph methods:
|
||||||
|
using Base::empty;
|
||||||
|
using Base::reserve;
|
||||||
|
using Base::size;
|
||||||
|
using Base::operator[];
|
||||||
|
using Base::add;
|
||||||
|
using Base::resize;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Add a nonlinear factor *pointer* to the internal nonlinear factor graph
|
||||||
|
* @param nonlinearFactor - boost::shared_ptr to the factor to add
|
||||||
|
*/
|
||||||
|
template <typename FACTOR>
|
||||||
|
IsNonlinear<FACTOR> push_nonlinear(
|
||||||
|
const boost::shared_ptr<FACTOR>& nonlinearFactor) {
|
||||||
|
Base::push_back(boost::make_shared<HybridNonlinearFactor>(nonlinearFactor));
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Construct a factor and add (shared pointer to it) to factor graph.
|
||||||
|
template <class FACTOR, class... Args>
|
||||||
|
IsNonlinear<FACTOR> emplace_nonlinear(Args&&... args) {
|
||||||
|
auto factor = boost::allocate_shared<FACTOR>(
|
||||||
|
Eigen::aligned_allocator<FACTOR>(), std::forward<Args>(args)...);
|
||||||
|
push_nonlinear(factor);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Add a single factor shared pointer to the hybrid factor graph.
|
||||||
|
* Dynamically handles the factor type and assigns it to the correct
|
||||||
|
* underlying container.
|
||||||
|
*
|
||||||
|
* @tparam FACTOR The factor type template
|
||||||
|
* @param sharedFactor The factor to add to this factor graph.
|
||||||
|
*/
|
||||||
|
template <typename FACTOR>
|
||||||
|
void push_back(const boost::shared_ptr<FACTOR>& sharedFactor) {
|
||||||
|
if (auto p = boost::dynamic_pointer_cast<NonlinearFactor>(sharedFactor)) {
|
||||||
|
push_nonlinear(p);
|
||||||
|
} else {
|
||||||
|
Base::push_back(sharedFactor);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Add a nonlinear factor as a shared ptr.
|
||||||
|
void add(boost::shared_ptr<NonlinearFactor> factor);
|
||||||
|
|
||||||
|
/// Print the factor graph.
|
||||||
|
void print(
|
||||||
|
const std::string& s = "HybridNonlinearFactorGraph",
|
||||||
|
const KeyFormatter& keyFormatter = DefaultKeyFormatter) const override;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Linearize all the continuous factors in the
|
||||||
|
* HybridNonlinearFactorGraph.
|
||||||
|
*
|
||||||
|
* @param continuousValues: Dictionary of continuous values.
|
||||||
|
* @return HybridGaussianFactorGraph::shared_ptr
|
||||||
|
*/
|
||||||
|
HybridGaussianFactorGraph linearize(const Values& continuousValues) const;
|
||||||
|
};
|
||||||
|
|
||||||
|
template <>
|
||||||
|
struct traits<HybridNonlinearFactorGraph>
|
||||||
|
: public Testable<HybridNonlinearFactorGraph> {};
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -0,0 +1,127 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file HybridValues.h
|
||||||
|
* @date Jul 28, 2022
|
||||||
|
* @author Shangjie Xue
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <gtsam/discrete/Assignment.h>
|
||||||
|
#include <gtsam/discrete/DiscreteKey.h>
|
||||||
|
#include <gtsam/discrete/DiscreteValues.h>
|
||||||
|
#include <gtsam/inference/Key.h>
|
||||||
|
#include <gtsam/linear/VectorValues.h>
|
||||||
|
#include <gtsam/nonlinear/Values.h>
|
||||||
|
|
||||||
|
#include <map>
|
||||||
|
#include <string>
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* HybridValues represents a collection of DiscreteValues and VectorValues. It
|
||||||
|
* is typically used to store the variables of a HybridGaussianFactorGraph.
|
||||||
|
* Optimizing a HybridGaussianBayesNet returns this class.
|
||||||
|
*/
|
||||||
|
class GTSAM_EXPORT HybridValues {
|
||||||
|
public:
|
||||||
|
// DiscreteValue stored the discrete components of the HybridValues.
|
||||||
|
DiscreteValues discrete;
|
||||||
|
|
||||||
|
// VectorValue stored the continuous components of the HybridValues.
|
||||||
|
VectorValues continuous;
|
||||||
|
|
||||||
|
// Default constructor creates an empty HybridValues.
|
||||||
|
HybridValues() : discrete(), continuous(){};
|
||||||
|
|
||||||
|
// Construct from DiscreteValues and VectorValues.
|
||||||
|
HybridValues(const DiscreteValues& dv, const VectorValues& cv)
|
||||||
|
: discrete(dv), continuous(cv){};
|
||||||
|
|
||||||
|
// print required by Testable for unit testing
|
||||||
|
void print(const std::string& s = "HybridValues",
|
||||||
|
const KeyFormatter& keyFormatter = DefaultKeyFormatter) const {
|
||||||
|
std::cout << s << ": \n";
|
||||||
|
discrete.print(" Discrete", keyFormatter); // print discrete components
|
||||||
|
continuous.print(" Continuous",
|
||||||
|
keyFormatter); // print continuous components
|
||||||
|
};
|
||||||
|
|
||||||
|
// equals required by Testable for unit testing
|
||||||
|
bool equals(const HybridValues& other, double tol = 1e-9) const {
|
||||||
|
return discrete.equals(other.discrete, tol) &&
|
||||||
|
continuous.equals(other.continuous, tol);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check whether a variable with key \c j exists in DiscreteValue.
|
||||||
|
bool existsDiscrete(Key j) { return (discrete.find(j) != discrete.end()); };
|
||||||
|
|
||||||
|
// Check whether a variable with key \c j exists in VectorValue.
|
||||||
|
bool existsVector(Key j) { return continuous.exists(j); };
|
||||||
|
|
||||||
|
// Check whether a variable with key \c j exists.
|
||||||
|
bool exists(Key j) { return existsDiscrete(j) || existsVector(j); };
|
||||||
|
|
||||||
|
/** Insert a discrete \c value with key \c j. Replaces the existing value if
|
||||||
|
* the key \c j is already used.
|
||||||
|
* @param value The vector to be inserted.
|
||||||
|
* @param j The index with which the value will be associated. */
|
||||||
|
void insert(Key j, int value) { discrete[j] = value; };
|
||||||
|
|
||||||
|
/** Insert a vector \c value with key \c j. Throws an invalid_argument
|
||||||
|
* exception if the key \c j is already used.
|
||||||
|
* @param value The vector to be inserted.
|
||||||
|
* @param j The index with which the value will be associated. */
|
||||||
|
void insert(Key j, const Vector& value) { continuous.insert(j, value); }
|
||||||
|
|
||||||
|
// TODO(Shangjie)- update() and insert_or_assign() , similar to Values.h
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Read/write access to the discrete value with key \c j, throws
|
||||||
|
* std::out_of_range if \c j does not exist.
|
||||||
|
*/
|
||||||
|
size_t& atDiscrete(Key j) { return discrete.at(j); };
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Read/write access to the vector value with key \c j, throws
|
||||||
|
* std::out_of_range if \c j does not exist.
|
||||||
|
*/
|
||||||
|
Vector& at(Key j) { return continuous.at(j); };
|
||||||
|
|
||||||
|
/// @name Wrapper support
|
||||||
|
/// @{
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Output as a html table.
|
||||||
|
*
|
||||||
|
* @param keyFormatter function that formats keys.
|
||||||
|
* @return string html output.
|
||||||
|
*/
|
||||||
|
std::string html(
|
||||||
|
const KeyFormatter& keyFormatter = DefaultKeyFormatter) const {
|
||||||
|
std::stringstream ss;
|
||||||
|
ss << this->discrete.html(keyFormatter);
|
||||||
|
ss << this->continuous.html(keyFormatter);
|
||||||
|
return ss.str();
|
||||||
|
};
|
||||||
|
|
||||||
|
/// @}
|
||||||
|
};
|
||||||
|
|
||||||
|
// traits
|
||||||
|
template <>
|
||||||
|
struct traits<HybridValues> : public Testable<HybridValues> {};
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -0,0 +1,244 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file MixtureFactor.h
|
||||||
|
* @brief Nonlinear Mixture factor of continuous and discrete.
|
||||||
|
* @author Kevin Doherty, kdoherty@mit.edu
|
||||||
|
* @author Varun Agrawal
|
||||||
|
* @date December 2021
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <gtsam/discrete/DiscreteValues.h>
|
||||||
|
#include <gtsam/hybrid/GaussianMixtureFactor.h>
|
||||||
|
#include <gtsam/hybrid/HybridNonlinearFactor.h>
|
||||||
|
#include <gtsam/nonlinear/NonlinearFactor.h>
|
||||||
|
#include <gtsam/nonlinear/Symbol.h>
|
||||||
|
|
||||||
|
#include <algorithm>
|
||||||
|
#include <boost/format.hpp>
|
||||||
|
#include <cmath>
|
||||||
|
#include <limits>
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
|
namespace gtsam {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Implementation of a discrete conditional mixture factor.
|
||||||
|
*
|
||||||
|
* Implements a joint discrete-continuous factor where the discrete variable
|
||||||
|
* serves to "select" a mixture component corresponding to a NonlinearFactor
|
||||||
|
* type of measurement.
|
||||||
|
*
|
||||||
|
* This class stores all factors as HybridFactors which can then be typecast to
|
||||||
|
* one of (NonlinearFactor, GaussianFactor) which can then be checked to perform
|
||||||
|
* the correct operation.
|
||||||
|
*/
|
||||||
|
class MixtureFactor : public HybridFactor {
|
||||||
|
public:
|
||||||
|
using Base = HybridFactor;
|
||||||
|
using This = MixtureFactor;
|
||||||
|
using shared_ptr = boost::shared_ptr<MixtureFactor>;
|
||||||
|
using sharedFactor = boost::shared_ptr<NonlinearFactor>;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief typedef for DecisionTree which has Keys as node labels and
|
||||||
|
* NonlinearFactor as leaf nodes.
|
||||||
|
*/
|
||||||
|
using Factors = DecisionTree<Key, sharedFactor>;
|
||||||
|
|
||||||
|
private:
|
||||||
|
/// Decision tree of Gaussian factors indexed by discrete keys.
|
||||||
|
Factors factors_;
|
||||||
|
bool normalized_;
|
||||||
|
|
||||||
|
public:
|
||||||
|
MixtureFactor() = default;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Construct from Decision tree.
|
||||||
|
*
|
||||||
|
* @param keys Vector of keys for continuous factors.
|
||||||
|
* @param discreteKeys Vector of discrete keys.
|
||||||
|
* @param factors Decision tree with of shared factors.
|
||||||
|
* @param normalized Flag indicating if the factor error is already
|
||||||
|
* normalized.
|
||||||
|
*/
|
||||||
|
MixtureFactor(const KeyVector& keys, const DiscreteKeys& discreteKeys,
|
||||||
|
const Factors& factors, bool normalized = false)
|
||||||
|
: Base(keys, discreteKeys), factors_(factors), normalized_(normalized) {}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Convenience constructor that generates the underlying factor
|
||||||
|
* decision tree for us.
|
||||||
|
*
|
||||||
|
* Here it is important that the vector of factors has the correct number of
|
||||||
|
* elements based on the number of discrete keys and the cardinality of the
|
||||||
|
* keys, so that the decision tree is constructed appropriately.
|
||||||
|
*
|
||||||
|
* @tparam FACTOR The type of the factor shared pointers being passed in. Will
|
||||||
|
* be typecast to NonlinearFactor shared pointers.
|
||||||
|
* @param keys Vector of keys for continuous factors.
|
||||||
|
* @param discreteKeys Vector of discrete keys.
|
||||||
|
* @param factors Vector of shared pointers to factors.
|
||||||
|
* @param normalized Flag indicating if the factor error is already
|
||||||
|
* normalized.
|
||||||
|
*/
|
||||||
|
template <typename FACTOR>
|
||||||
|
MixtureFactor(const KeyVector& keys, const DiscreteKeys& discreteKeys,
|
||||||
|
const std::vector<boost::shared_ptr<FACTOR>>& factors,
|
||||||
|
bool normalized = false)
|
||||||
|
: Base(keys, discreteKeys), normalized_(normalized) {
|
||||||
|
std::vector<NonlinearFactor::shared_ptr> nonlinear_factors;
|
||||||
|
for (auto&& f : factors) {
|
||||||
|
nonlinear_factors.push_back(
|
||||||
|
boost::dynamic_pointer_cast<NonlinearFactor>(f));
|
||||||
|
}
|
||||||
|
factors_ = Factors(discreteKeys, nonlinear_factors);
|
||||||
|
}
|
||||||
|
|
||||||
|
~MixtureFactor() = default;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Compute error of factor given both continuous and discrete values.
|
||||||
|
*
|
||||||
|
* @param continuousVals The continuous Values.
|
||||||
|
* @param discreteVals The discrete Values.
|
||||||
|
* @return double The error of this factor.
|
||||||
|
*/
|
||||||
|
double error(const Values& continuousVals,
|
||||||
|
const DiscreteValues& discreteVals) const {
|
||||||
|
// Retrieve the factor corresponding to the assignment in discreteVals.
|
||||||
|
auto factor = factors_(discreteVals);
|
||||||
|
// Compute the error for the selected factor
|
||||||
|
const double factorError = factor->error(continuousVals);
|
||||||
|
if (normalized_) return factorError;
|
||||||
|
return factorError +
|
||||||
|
this->nonlinearFactorLogNormalizingConstant(factor, continuousVals);
|
||||||
|
}
|
||||||
|
|
||||||
|
size_t dim() const {
|
||||||
|
// TODO(Varun)
|
||||||
|
throw std::runtime_error("MixtureFactor::dim not implemented.");
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Testable
|
||||||
|
/// @{
|
||||||
|
|
||||||
|
/// print to stdout
|
||||||
|
void print(
|
||||||
|
const std::string& s = "MixtureFactor",
|
||||||
|
const KeyFormatter& keyFormatter = DefaultKeyFormatter) const override {
|
||||||
|
std::cout << (s.empty() ? "" : s + " ");
|
||||||
|
Base::print("", keyFormatter);
|
||||||
|
std::cout << "\nMixtureFactor\n";
|
||||||
|
auto valueFormatter = [](const sharedFactor& v) {
|
||||||
|
if (v) {
|
||||||
|
return (boost::format("Nonlinear factor on %d keys") % v->size()).str();
|
||||||
|
} else {
|
||||||
|
return std::string("nullptr");
|
||||||
|
}
|
||||||
|
};
|
||||||
|
factors_.print("", keyFormatter, valueFormatter);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Check equality
|
||||||
|
bool equals(const HybridFactor& other, double tol = 1e-9) const override {
|
||||||
|
// We attempt a dynamic cast from HybridFactor to MixtureFactor. If it
|
||||||
|
// fails, return false.
|
||||||
|
if (!dynamic_cast<const MixtureFactor*>(&other)) return false;
|
||||||
|
|
||||||
|
// If the cast is successful, we'll properly construct a MixtureFactor
|
||||||
|
// object from `other`
|
||||||
|
const MixtureFactor& f(static_cast<const MixtureFactor&>(other));
|
||||||
|
|
||||||
|
// Ensure that this MixtureFactor and `f` have the same `factors_`.
|
||||||
|
auto compare = [tol](const sharedFactor& a, const sharedFactor& b) {
|
||||||
|
return traits<NonlinearFactor>::Equals(*a, *b, tol);
|
||||||
|
};
|
||||||
|
if (!factors_.equals(f.factors_, compare)) return false;
|
||||||
|
|
||||||
|
// If everything above passes, and the keys_, discreteKeys_ and normalized_
|
||||||
|
// member variables are identical, return true.
|
||||||
|
return (std::equal(keys_.begin(), keys_.end(), f.keys().begin()) &&
|
||||||
|
(discreteKeys_ == f.discreteKeys_) &&
|
||||||
|
(normalized_ == f.normalized_));
|
||||||
|
}
|
||||||
|
|
||||||
|
/// @}
|
||||||
|
|
||||||
|
/// Linearize specific nonlinear factors based on the assignment in
|
||||||
|
/// discreteValues.
|
||||||
|
GaussianFactor::shared_ptr linearize(
|
||||||
|
const Values& continuousVals, const DiscreteValues& discreteVals) const {
|
||||||
|
auto factor = factors_(discreteVals);
|
||||||
|
return factor->linearize(continuousVals);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Linearize all the continuous factors to get a GaussianMixtureFactor.
|
||||||
|
boost::shared_ptr<GaussianMixtureFactor> linearize(
|
||||||
|
const Values& continuousVals) const {
|
||||||
|
// functional to linearize each factor in the decision tree
|
||||||
|
auto linearizeDT = [continuousVals](const sharedFactor& factor) {
|
||||||
|
return factor->linearize(continuousVals);
|
||||||
|
};
|
||||||
|
|
||||||
|
DecisionTree<Key, GaussianFactor::shared_ptr> linearized_factors(
|
||||||
|
factors_, linearizeDT);
|
||||||
|
|
||||||
|
return boost::make_shared<GaussianMixtureFactor>(
|
||||||
|
continuousKeys_, discreteKeys_, linearized_factors);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* If the component factors are not already normalized, we want to compute
|
||||||
|
* their normalizing constants so that the resulting joint distribution is
|
||||||
|
* appropriately computed. Remember, this is the _negative_ normalizing
|
||||||
|
* constant for the measurement likelihood (since we are minimizing the
|
||||||
|
* _negative_ log-likelihood).
|
||||||
|
*/
|
||||||
|
double nonlinearFactorLogNormalizingConstant(const sharedFactor& factor,
|
||||||
|
const Values& values) const {
|
||||||
|
// Information matrix (inverse covariance matrix) for the factor.
|
||||||
|
Matrix infoMat;
|
||||||
|
|
||||||
|
// If this is a NoiseModelFactor, we'll use its noiseModel to
|
||||||
|
// otherwise noiseModelFactor will be nullptr
|
||||||
|
if (auto noiseModelFactor =
|
||||||
|
boost::dynamic_pointer_cast<NoiseModelFactor>(factor)) {
|
||||||
|
// If dynamic cast to NoiseModelFactor succeeded, see if the noise model
|
||||||
|
// is Gaussian
|
||||||
|
auto noiseModel = noiseModelFactor->noiseModel();
|
||||||
|
|
||||||
|
auto gaussianNoiseModel =
|
||||||
|
boost::dynamic_pointer_cast<noiseModel::Gaussian>(noiseModel);
|
||||||
|
if (gaussianNoiseModel) {
|
||||||
|
// If the noise model is Gaussian, retrieve the information matrix
|
||||||
|
infoMat = gaussianNoiseModel->information();
|
||||||
|
} else {
|
||||||
|
// If the factor is not a Gaussian factor, we'll linearize it to get
|
||||||
|
// something with a normalized noise model
|
||||||
|
// TODO(kevin): does this make sense to do? I think maybe not in
|
||||||
|
// general? Should we just yell at the user?
|
||||||
|
auto gaussianFactor = factor->linearize(values);
|
||||||
|
infoMat = gaussianFactor->information();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compute the (negative) log of the normalizing constant
|
||||||
|
return -(factor->dim() * log(2.0 * M_PI) / 2.0) -
|
||||||
|
(log(infoMat.determinant()) / 2.0);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace gtsam
|
||||||
|
|
@ -4,6 +4,22 @@
|
||||||
|
|
||||||
namespace gtsam {
|
namespace gtsam {
|
||||||
|
|
||||||
|
#include <gtsam/hybrid/HybridValues.h>
|
||||||
|
class HybridValues {
|
||||||
|
gtsam::DiscreteValues discrete;
|
||||||
|
gtsam::VectorValues continuous;
|
||||||
|
HybridValues();
|
||||||
|
HybridValues(const gtsam::DiscreteValues &dv, const gtsam::VectorValues &cv);
|
||||||
|
void print(string s = "HybridValues",
|
||||||
|
const gtsam::KeyFormatter& keyFormatter =
|
||||||
|
gtsam::DefaultKeyFormatter) const;
|
||||||
|
bool equals(const gtsam::HybridValues& other, double tol) const;
|
||||||
|
void insert(gtsam::Key j, int value);
|
||||||
|
void insert(gtsam::Key j, const gtsam::Vector& value);
|
||||||
|
size_t& atDiscrete(gtsam::Key j);
|
||||||
|
gtsam::Vector& at(gtsam::Key j);
|
||||||
|
};
|
||||||
|
|
||||||
#include <gtsam/hybrid/HybridFactor.h>
|
#include <gtsam/hybrid/HybridFactor.h>
|
||||||
virtual class HybridFactor {
|
virtual class HybridFactor {
|
||||||
void print(string s = "HybridFactor\n",
|
void print(string s = "HybridFactor\n",
|
||||||
|
|
@ -84,6 +100,7 @@ class HybridBayesNet {
|
||||||
size_t size() const;
|
size_t size() const;
|
||||||
gtsam::KeySet keys() const;
|
gtsam::KeySet keys() const;
|
||||||
const gtsam::HybridConditional* at(size_t i) const;
|
const gtsam::HybridConditional* at(size_t i) const;
|
||||||
|
gtsam::HybridValues optimize() const;
|
||||||
void print(string s = "HybridBayesNet\n",
|
void print(string s = "HybridBayesNet\n",
|
||||||
const gtsam::KeyFormatter& keyFormatter =
|
const gtsam::KeyFormatter& keyFormatter =
|
||||||
gtsam::DefaultKeyFormatter) const;
|
gtsam::DefaultKeyFormatter) const;
|
||||||
|
|
|
||||||
|
|
@ -18,20 +18,39 @@
|
||||||
|
|
||||||
#include <gtsam/base/Matrix.h>
|
#include <gtsam/base/Matrix.h>
|
||||||
#include <gtsam/discrete/DecisionTreeFactor.h>
|
#include <gtsam/discrete/DecisionTreeFactor.h>
|
||||||
|
#include <gtsam/discrete/DiscreteDistribution.h>
|
||||||
#include <gtsam/hybrid/GaussianMixtureFactor.h>
|
#include <gtsam/hybrid/GaussianMixtureFactor.h>
|
||||||
#include <gtsam/hybrid/HybridGaussianFactorGraph.h>
|
#include <gtsam/hybrid/HybridGaussianFactorGraph.h>
|
||||||
|
#include <gtsam/hybrid/HybridNonlinearFactorGraph.h>
|
||||||
|
#include <gtsam/hybrid/MixtureFactor.h>
|
||||||
#include <gtsam/inference/Symbol.h>
|
#include <gtsam/inference/Symbol.h>
|
||||||
#include <gtsam/linear/JacobianFactor.h>
|
#include <gtsam/linear/JacobianFactor.h>
|
||||||
|
#include <gtsam/linear/NoiseModel.h>
|
||||||
|
#include <gtsam/nonlinear/PriorFactor.h>
|
||||||
|
#include <gtsam/slam/BetweenFactor.h>
|
||||||
|
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
#pragma once
|
#pragma once
|
||||||
|
|
||||||
using gtsam::symbol_shorthand::C;
|
|
||||||
using gtsam::symbol_shorthand::X;
|
|
||||||
|
|
||||||
namespace gtsam {
|
namespace gtsam {
|
||||||
|
|
||||||
|
using symbol_shorthand::C;
|
||||||
|
using symbol_shorthand::M;
|
||||||
|
using symbol_shorthand::X;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Create a switching system chain. A switching system is a continuous
|
||||||
|
* system which depends on a discrete mode at each time step of the chain.
|
||||||
|
*
|
||||||
|
* @param n The number of chain elements.
|
||||||
|
* @param keyFunc The functional to help specify the continuous key.
|
||||||
|
* @param dKeyFunc The functional to help specify the discrete key.
|
||||||
|
* @return HybridGaussianFactorGraph::shared_ptr
|
||||||
|
*/
|
||||||
inline HybridGaussianFactorGraph::shared_ptr makeSwitchingChain(
|
inline HybridGaussianFactorGraph::shared_ptr makeSwitchingChain(
|
||||||
size_t n, std::function<Key(int)> keyFunc = X,
|
size_t n, std::function<Key(int)> keyFunc = X,
|
||||||
std::function<Key(int)> dKeyFunc = C) {
|
std::function<Key(int)> dKeyFunc = M) {
|
||||||
HybridGaussianFactorGraph hfg;
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
hfg.add(JacobianFactor(keyFunc(1), I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(keyFunc(1), I_3x3, Z_3x1));
|
||||||
|
|
@ -54,9 +73,19 @@ inline HybridGaussianFactorGraph::shared_ptr makeSwitchingChain(
|
||||||
return boost::make_shared<HybridGaussianFactorGraph>(std::move(hfg));
|
return boost::make_shared<HybridGaussianFactorGraph>(std::move(hfg));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Return the ordering as a binary tree such that all parent nodes are
|
||||||
|
* above their children.
|
||||||
|
*
|
||||||
|
* This will result in a nested dissection Bayes tree after elimination.
|
||||||
|
*
|
||||||
|
* @param input The original ordering.
|
||||||
|
* @return std::pair<KeyVector, std::vector<int>>
|
||||||
|
*/
|
||||||
inline std::pair<KeyVector, std::vector<int>> makeBinaryOrdering(
|
inline std::pair<KeyVector, std::vector<int>> makeBinaryOrdering(
|
||||||
std::vector<Key> &input) {
|
std::vector<Key> &input) {
|
||||||
KeyVector new_order;
|
KeyVector new_order;
|
||||||
|
|
||||||
std::vector<int> levels(input.size());
|
std::vector<int> levels(input.size());
|
||||||
std::function<void(std::vector<Key>::iterator, std::vector<Key>::iterator,
|
std::function<void(std::vector<Key>::iterator, std::vector<Key>::iterator,
|
||||||
int)>
|
int)>
|
||||||
|
|
@ -79,8 +108,107 @@ inline std::pair<KeyVector, std::vector<int>> makeBinaryOrdering(
|
||||||
|
|
||||||
bsg(input.begin(), input.end(), 0);
|
bsg(input.begin(), input.end(), 0);
|
||||||
std::reverse(new_order.begin(), new_order.end());
|
std::reverse(new_order.begin(), new_order.end());
|
||||||
// std::reverse(levels.begin(), levels.end());
|
|
||||||
return {new_order, levels};
|
return {new_order, levels};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* ***************************************************************************
|
||||||
|
*/
|
||||||
|
using MotionModel = BetweenFactor<double>;
|
||||||
|
// using MotionMixture = MixtureFactor<MotionModel>;
|
||||||
|
|
||||||
|
// Test fixture with switching network.
|
||||||
|
struct Switching {
|
||||||
|
size_t K;
|
||||||
|
DiscreteKeys modes;
|
||||||
|
HybridNonlinearFactorGraph nonlinearFactorGraph;
|
||||||
|
HybridGaussianFactorGraph linearizedFactorGraph;
|
||||||
|
Values linearizationPoint;
|
||||||
|
|
||||||
|
/// Create with given number of time steps.
|
||||||
|
Switching(size_t K, double between_sigma = 1.0, double prior_sigma = 0.1)
|
||||||
|
: K(K) {
|
||||||
|
using symbol_shorthand::M;
|
||||||
|
using symbol_shorthand::X;
|
||||||
|
|
||||||
|
// Create DiscreteKeys for binary K modes, modes[0] will not be used.
|
||||||
|
for (size_t k = 0; k <= K; k++) {
|
||||||
|
modes.emplace_back(M(k), 2);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create hybrid factor graph.
|
||||||
|
// Add a prior on X(1).
|
||||||
|
auto prior = boost::make_shared<PriorFactor<double>>(
|
||||||
|
X(1), 0, noiseModel::Isotropic::Sigma(1, prior_sigma));
|
||||||
|
nonlinearFactorGraph.push_nonlinear(prior);
|
||||||
|
|
||||||
|
// Add "motion models".
|
||||||
|
for (size_t k = 1; k < K; k++) {
|
||||||
|
KeyVector keys = {X(k), X(k + 1)};
|
||||||
|
auto motion_models = motionModels(k);
|
||||||
|
std::vector<NonlinearFactor::shared_ptr> components;
|
||||||
|
for (auto &&f : motion_models) {
|
||||||
|
components.push_back(boost::dynamic_pointer_cast<NonlinearFactor>(f));
|
||||||
|
}
|
||||||
|
nonlinearFactorGraph.emplace_hybrid<MixtureFactor>(
|
||||||
|
keys, DiscreteKeys{modes[k]}, components);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add measurement factors
|
||||||
|
auto measurement_noise = noiseModel::Isotropic::Sigma(1, 0.1);
|
||||||
|
for (size_t k = 2; k <= K; k++) {
|
||||||
|
nonlinearFactorGraph.emplace_nonlinear<PriorFactor<double>>(
|
||||||
|
X(k), 1.0 * (k - 1), measurement_noise);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add "mode chain"
|
||||||
|
addModeChain(&nonlinearFactorGraph);
|
||||||
|
|
||||||
|
// Create the linearization point.
|
||||||
|
for (size_t k = 1; k <= K; k++) {
|
||||||
|
linearizationPoint.insert<double>(X(k), static_cast<double>(k));
|
||||||
|
}
|
||||||
|
|
||||||
|
linearizedFactorGraph = nonlinearFactorGraph.linearize(linearizationPoint);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create motion models for a given time step
|
||||||
|
static std::vector<MotionModel::shared_ptr> motionModels(size_t k,
|
||||||
|
double sigma = 1.0) {
|
||||||
|
using symbol_shorthand::M;
|
||||||
|
using symbol_shorthand::X;
|
||||||
|
|
||||||
|
auto noise_model = noiseModel::Isotropic::Sigma(1, sigma);
|
||||||
|
auto still =
|
||||||
|
boost::make_shared<MotionModel>(X(k), X(k + 1), 0.0, noise_model),
|
||||||
|
moving =
|
||||||
|
boost::make_shared<MotionModel>(X(k), X(k + 1), 1.0, noise_model);
|
||||||
|
return {still, moving};
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add "mode chain" to HybridNonlinearFactorGraph
|
||||||
|
void addModeChain(HybridNonlinearFactorGraph *fg) {
|
||||||
|
auto prior = boost::make_shared<DiscreteDistribution>(modes[1], "1/1");
|
||||||
|
fg->push_discrete(prior);
|
||||||
|
for (size_t k = 1; k < K - 1; k++) {
|
||||||
|
auto parents = {modes[k]};
|
||||||
|
auto conditional = boost::make_shared<DiscreteConditional>(
|
||||||
|
modes[k + 1], parents, "1/2 3/2");
|
||||||
|
fg->push_discrete(conditional);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add "mode chain" to HybridGaussianFactorGraph
|
||||||
|
void addModeChain(HybridGaussianFactorGraph *fg) {
|
||||||
|
auto prior = boost::make_shared<DiscreteDistribution>(modes[1], "1/1");
|
||||||
|
fg->push_discrete(prior);
|
||||||
|
for (size_t k = 1; k < K - 1; k++) {
|
||||||
|
auto parents = {modes[k]};
|
||||||
|
auto conditional = boost::make_shared<DiscreteConditional>(
|
||||||
|
modes[k + 1], parents, "1/2 3/2");
|
||||||
|
fg->push_discrete(conditional);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
} // namespace gtsam
|
} // namespace gtsam
|
||||||
|
|
|
||||||
|
|
@ -17,6 +17,7 @@
|
||||||
|
|
||||||
#include <CppUnitLite/Test.h>
|
#include <CppUnitLite/Test.h>
|
||||||
#include <CppUnitLite/TestHarness.h>
|
#include <CppUnitLite/TestHarness.h>
|
||||||
|
#include <gtsam/base/TestableAssertions.h>
|
||||||
#include <gtsam/discrete/DecisionTreeFactor.h>
|
#include <gtsam/discrete/DecisionTreeFactor.h>
|
||||||
#include <gtsam/discrete/DiscreteKey.h>
|
#include <gtsam/discrete/DiscreteKey.h>
|
||||||
#include <gtsam/discrete/DiscreteValues.h>
|
#include <gtsam/discrete/DiscreteValues.h>
|
||||||
|
|
@ -30,6 +31,7 @@
|
||||||
#include <gtsam/hybrid/HybridGaussianFactor.h>
|
#include <gtsam/hybrid/HybridGaussianFactor.h>
|
||||||
#include <gtsam/hybrid/HybridGaussianFactorGraph.h>
|
#include <gtsam/hybrid/HybridGaussianFactorGraph.h>
|
||||||
#include <gtsam/hybrid/HybridGaussianISAM.h>
|
#include <gtsam/hybrid/HybridGaussianISAM.h>
|
||||||
|
#include <gtsam/hybrid/HybridValues.h>
|
||||||
#include <gtsam/inference/BayesNet.h>
|
#include <gtsam/inference/BayesNet.h>
|
||||||
#include <gtsam/inference/DotWriter.h>
|
#include <gtsam/inference/DotWriter.h>
|
||||||
#include <gtsam/inference/Key.h>
|
#include <gtsam/inference/Key.h>
|
||||||
|
|
@ -52,32 +54,36 @@ using namespace boost::assign;
|
||||||
using namespace std;
|
using namespace std;
|
||||||
using namespace gtsam;
|
using namespace gtsam;
|
||||||
|
|
||||||
using gtsam::symbol_shorthand::C;
|
|
||||||
using gtsam::symbol_shorthand::D;
|
using gtsam::symbol_shorthand::D;
|
||||||
|
using gtsam::symbol_shorthand::M;
|
||||||
using gtsam::symbol_shorthand::X;
|
using gtsam::symbol_shorthand::X;
|
||||||
using gtsam::symbol_shorthand::Y;
|
using gtsam::symbol_shorthand::Y;
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
TEST(HybridGaussianFactorGraph, creation) {
|
TEST(HybridGaussianFactorGraph, Creation) {
|
||||||
HybridConditional test;
|
HybridConditional conditional;
|
||||||
|
|
||||||
HybridGaussianFactorGraph hfg;
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
hfg.add(HybridGaussianFactor(JacobianFactor(0, I_3x3, Z_3x1)));
|
hfg.add(HybridGaussianFactor(JacobianFactor(X(0), I_3x3, Z_3x1)));
|
||||||
|
|
||||||
GaussianMixture clgc(
|
// Define a gaussian mixture conditional P(x0|x1, c0) and add it to the factor
|
||||||
{X(0)}, {X(1)}, DiscreteKeys(DiscreteKey{C(0), 2}),
|
// graph
|
||||||
GaussianMixture::Conditionals(
|
GaussianMixture gm({X(0)}, {X(1)}, DiscreteKeys(DiscreteKey{M(0), 2}),
|
||||||
C(0),
|
GaussianMixture::Conditionals(
|
||||||
boost::make_shared<GaussianConditional>(X(0), Z_3x1, I_3x3, X(1),
|
M(0),
|
||||||
I_3x3),
|
boost::make_shared<GaussianConditional>(
|
||||||
boost::make_shared<GaussianConditional>(X(0), Vector3::Ones(), I_3x3,
|
X(0), Z_3x1, I_3x3, X(1), I_3x3),
|
||||||
X(1), I_3x3)));
|
boost::make_shared<GaussianConditional>(
|
||||||
GTSAM_PRINT(clgc);
|
X(0), Vector3::Ones(), I_3x3, X(1), I_3x3)));
|
||||||
|
hfg.add(gm);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(2, hfg.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
TEST(HybridGaussianFactorGraph, eliminate) {
|
TEST(HybridGaussianFactorGraph, EliminateSequential) {
|
||||||
|
// Test elimination of a single variable.
|
||||||
HybridGaussianFactorGraph hfg;
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
hfg.add(HybridGaussianFactor(JacobianFactor(0, I_3x3, Z_3x1)));
|
hfg.add(HybridGaussianFactor(JacobianFactor(0, I_3x3, Z_3x1)));
|
||||||
|
|
@ -88,13 +94,15 @@ TEST(HybridGaussianFactorGraph, eliminate) {
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
TEST(HybridGaussianFactorGraph, eliminateMultifrontal) {
|
TEST(HybridGaussianFactorGraph, EliminateMultifrontal) {
|
||||||
|
// Test multifrontal elimination
|
||||||
HybridGaussianFactorGraph hfg;
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
DiscreteKey c(C(1), 2);
|
DiscreteKey m(M(1), 2);
|
||||||
|
|
||||||
|
// Add priors on x0 and c1
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
||||||
hfg.add(HybridDiscreteFactor(DecisionTreeFactor(c, {2, 8})));
|
hfg.add(HybridDiscreteFactor(DecisionTreeFactor(m, {2, 8})));
|
||||||
|
|
||||||
Ordering ordering;
|
Ordering ordering;
|
||||||
ordering.push_back(X(0));
|
ordering.push_back(X(0));
|
||||||
|
|
@ -108,117 +116,111 @@ TEST(HybridGaussianFactorGraph, eliminateMultifrontal) {
|
||||||
TEST(HybridGaussianFactorGraph, eliminateFullSequentialEqualChance) {
|
TEST(HybridGaussianFactorGraph, eliminateFullSequentialEqualChance) {
|
||||||
HybridGaussianFactorGraph hfg;
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
DiscreteKey c1(C(1), 2);
|
DiscreteKey m1(M(1), 2);
|
||||||
|
|
||||||
|
// Add prior on x0
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
||||||
|
// Add factor between x0 and x1
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
||||||
|
|
||||||
|
// Add a gaussian mixture factor ϕ(x1, c1)
|
||||||
DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
||||||
C(1), boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
M(1), boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
||||||
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones()));
|
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones()));
|
||||||
|
|
||||||
hfg.add(GaussianMixtureFactor({X(1)}, {c1}, dt));
|
hfg.add(GaussianMixtureFactor({X(1)}, {m1}, dt));
|
||||||
|
|
||||||
auto result =
|
auto result =
|
||||||
hfg.eliminateSequential(Ordering::ColamdConstrainedLast(hfg, {C(1)}));
|
hfg.eliminateSequential(Ordering::ColamdConstrainedLast(hfg, {M(1)}));
|
||||||
|
|
||||||
auto dc = result->at(2)->asDiscreteConditional();
|
auto dc = result->at(2)->asDiscreteConditional();
|
||||||
DiscreteValues dv;
|
DiscreteValues dv;
|
||||||
dv[C(1)] = 0;
|
dv[M(1)] = 0;
|
||||||
EXPECT_DOUBLES_EQUAL(0.6225, dc->operator()(dv), 1e-3);
|
EXPECT_DOUBLES_EQUAL(1, dc->operator()(dv), 1e-3);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
TEST(HybridGaussianFactorGraph, eliminateFullSequentialSimple) {
|
TEST(HybridGaussianFactorGraph, eliminateFullSequentialSimple) {
|
||||||
HybridGaussianFactorGraph hfg;
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
DiscreteKey c1(C(1), 2);
|
DiscreteKey m1(M(1), 2);
|
||||||
|
|
||||||
|
// Add prior on x0
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
||||||
|
// Add factor between x0 and x1
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
||||||
|
|
||||||
DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
std::vector<GaussianFactor::shared_ptr> factors = {
|
||||||
C(1), boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
||||||
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones()));
|
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones())};
|
||||||
|
hfg.add(GaussianMixtureFactor({X(1)}, {m1}, factors));
|
||||||
|
|
||||||
hfg.add(GaussianMixtureFactor({X(1)}, {c1}, dt));
|
// Discrete probability table for c1
|
||||||
// hfg.add(GaussianMixtureFactor({X(0)}, {c1}, dt));
|
hfg.add(DecisionTreeFactor(m1, {2, 8}));
|
||||||
hfg.add(HybridDiscreteFactor(DecisionTreeFactor(c1, {2, 8})));
|
// Joint discrete probability table for c1, c2
|
||||||
hfg.add(HybridDiscreteFactor(
|
hfg.add(DecisionTreeFactor({{M(1), 2}, {M(2), 2}}, "1 2 3 4"));
|
||||||
DecisionTreeFactor({{C(1), 2}, {C(2), 2}}, "1 2 3 4")));
|
|
||||||
// hfg.add(HybridDiscreteFactor(DecisionTreeFactor({{C(2), 2}, {C(3), 2}}, "1
|
|
||||||
// 2 3 4"))); hfg.add(HybridDiscreteFactor(DecisionTreeFactor({{C(3), 2},
|
|
||||||
// {C(1), 2}}, "1 2 2 1")));
|
|
||||||
|
|
||||||
auto result = hfg.eliminateSequential(
|
HybridBayesNet::shared_ptr result = hfg.eliminateSequential(
|
||||||
Ordering::ColamdConstrainedLast(hfg, {C(1), C(2)}));
|
Ordering::ColamdConstrainedLast(hfg, {M(1), M(2)}));
|
||||||
|
|
||||||
GTSAM_PRINT(*result);
|
// There are 4 variables (2 continuous + 2 discrete) in the bayes net.
|
||||||
|
EXPECT_LONGS_EQUAL(4, result->size());
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
TEST(HybridGaussianFactorGraph, eliminateFullMultifrontalSimple) {
|
TEST(HybridGaussianFactorGraph, eliminateFullMultifrontalSimple) {
|
||||||
HybridGaussianFactorGraph hfg;
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
DiscreteKey c1(C(1), 2);
|
DiscreteKey m1(M(1), 2);
|
||||||
|
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
||||||
|
|
||||||
// DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
|
||||||
// C(1), boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
|
||||||
// boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones()));
|
|
||||||
|
|
||||||
// hfg.add(GaussianMixtureFactor({X(1)}, {c1}, dt));
|
|
||||||
hfg.add(GaussianMixtureFactor::FromFactors(
|
hfg.add(GaussianMixtureFactor::FromFactors(
|
||||||
{X(1)}, {{C(1), 2}},
|
{X(1)}, {{M(1), 2}},
|
||||||
{boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
{boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
||||||
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones())}));
|
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones())}));
|
||||||
|
|
||||||
// hfg.add(GaussianMixtureFactor({X(0)}, {c1}, dt));
|
hfg.add(DecisionTreeFactor(m1, {2, 8}));
|
||||||
hfg.add(DecisionTreeFactor(c1, {2, 8}));
|
hfg.add(DecisionTreeFactor({{M(1), 2}, {M(2), 2}}, "1 2 3 4"));
|
||||||
hfg.add(DecisionTreeFactor({{C(1), 2}, {C(2), 2}}, "1 2 3 4"));
|
|
||||||
// hfg.add(HybridDiscreteFactor(DecisionTreeFactor({{C(2), 2}, {C(3), 2}}, "1
|
|
||||||
// 2 3 4"))); hfg.add(HybridDiscreteFactor(DecisionTreeFactor({{C(3), 2},
|
|
||||||
// {C(1), 2}}, "1 2 2 1")));
|
|
||||||
|
|
||||||
auto result = hfg.eliminateMultifrontal(
|
HybridBayesTree::shared_ptr result = hfg.eliminateMultifrontal(
|
||||||
Ordering::ColamdConstrainedLast(hfg, {C(1), C(2)}));
|
Ordering::ColamdConstrainedLast(hfg, {M(1), M(2)}));
|
||||||
|
|
||||||
GTSAM_PRINT(*result);
|
// The bayes tree should have 3 cliques
|
||||||
GTSAM_PRINT(*result->marginalFactor(C(2)));
|
EXPECT_LONGS_EQUAL(3, result->size());
|
||||||
|
// GTSAM_PRINT(*result);
|
||||||
|
// GTSAM_PRINT(*result->marginalFactor(M(2)));
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
TEST(HybridGaussianFactorGraph, eliminateFullMultifrontalCLG) {
|
TEST(HybridGaussianFactorGraph, eliminateFullMultifrontalCLG) {
|
||||||
HybridGaussianFactorGraph hfg;
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
DiscreteKey c(C(1), 2);
|
DiscreteKey m(M(1), 2);
|
||||||
|
|
||||||
|
// Prior on x0
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
||||||
|
// Factor between x0-x1
|
||||||
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
||||||
|
|
||||||
|
// Decision tree with different modes on x1
|
||||||
DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
||||||
C(1), boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
M(1), boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
||||||
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones()));
|
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones()));
|
||||||
|
|
||||||
hfg.add(GaussianMixtureFactor({X(1)}, {c}, dt));
|
// Hybrid factor P(x1|c1)
|
||||||
hfg.add(HybridDiscreteFactor(DecisionTreeFactor(c, {2, 8})));
|
hfg.add(GaussianMixtureFactor({X(1)}, {m}, dt));
|
||||||
// hfg.add(HybridDiscreteFactor(DecisionTreeFactor({{C(1), 2}, {C(2), 2}}, "1
|
// Prior factor on c1
|
||||||
// 2 3 4")));
|
hfg.add(HybridDiscreteFactor(DecisionTreeFactor(m, {2, 8})));
|
||||||
|
|
||||||
auto ordering_full = Ordering::ColamdConstrainedLast(hfg, {C(1)});
|
// Get a constrained ordering keeping c1 last
|
||||||
|
auto ordering_full = Ordering::ColamdConstrainedLast(hfg, {M(1)});
|
||||||
|
|
||||||
|
// Returns a Hybrid Bayes Tree with distribution P(x0|x1)P(x1|c1)P(c1)
|
||||||
HybridBayesTree::shared_ptr hbt = hfg.eliminateMultifrontal(ordering_full);
|
HybridBayesTree::shared_ptr hbt = hfg.eliminateMultifrontal(ordering_full);
|
||||||
|
|
||||||
GTSAM_PRINT(*hbt);
|
EXPECT_LONGS_EQUAL(3, hbt->size());
|
||||||
/*
|
|
||||||
Explanation: the Junction tree will need to reeliminate to get to the marginal
|
|
||||||
on X(1), which is not possible because it involves eliminating discrete before
|
|
||||||
continuous. The solution to this, however, is in Murphy02. TLDR is that this
|
|
||||||
is 1. expensive and 2. inexact. neverless it is doable. And I believe that we
|
|
||||||
should do this.
|
|
||||||
*/
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
|
|
@ -233,66 +235,72 @@ TEST(HybridGaussianFactorGraph, eliminateFullMultifrontalTwoClique) {
|
||||||
hfg.add(JacobianFactor(X(1), I_3x3, X(2), -I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(1), I_3x3, X(2), -I_3x3, Z_3x1));
|
||||||
|
|
||||||
{
|
{
|
||||||
// DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
|
||||||
// C(0), boost::make_shared<JacobianFactor>(X(0), I_3x3, Z_3x1),
|
|
||||||
// boost::make_shared<JacobianFactor>(X(0), I_3x3, Vector3::Ones()));
|
|
||||||
|
|
||||||
hfg.add(GaussianMixtureFactor::FromFactors(
|
hfg.add(GaussianMixtureFactor::FromFactors(
|
||||||
{X(0)}, {{C(0), 2}},
|
{X(0)}, {{M(0), 2}},
|
||||||
{boost::make_shared<JacobianFactor>(X(0), I_3x3, Z_3x1),
|
{boost::make_shared<JacobianFactor>(X(0), I_3x3, Z_3x1),
|
||||||
boost::make_shared<JacobianFactor>(X(0), I_3x3, Vector3::Ones())}));
|
boost::make_shared<JacobianFactor>(X(0), I_3x3, Vector3::Ones())}));
|
||||||
|
|
||||||
DecisionTree<Key, GaussianFactor::shared_ptr> dt1(
|
DecisionTree<Key, GaussianFactor::shared_ptr> dt1(
|
||||||
C(1), boost::make_shared<JacobianFactor>(X(2), I_3x3, Z_3x1),
|
M(1), boost::make_shared<JacobianFactor>(X(2), I_3x3, Z_3x1),
|
||||||
boost::make_shared<JacobianFactor>(X(2), I_3x3, Vector3::Ones()));
|
boost::make_shared<JacobianFactor>(X(2), I_3x3, Vector3::Ones()));
|
||||||
|
|
||||||
hfg.add(GaussianMixtureFactor({X(2)}, {{C(1), 2}}, dt1));
|
hfg.add(GaussianMixtureFactor({X(2)}, {{M(1), 2}}, dt1));
|
||||||
}
|
}
|
||||||
|
|
||||||
// hfg.add(HybridDiscreteFactor(DecisionTreeFactor(c, {2, 8})));
|
|
||||||
hfg.add(HybridDiscreteFactor(
|
hfg.add(HybridDiscreteFactor(
|
||||||
DecisionTreeFactor({{C(1), 2}, {C(2), 2}}, "1 2 3 4")));
|
DecisionTreeFactor({{M(1), 2}, {M(2), 2}}, "1 2 3 4")));
|
||||||
|
|
||||||
hfg.add(JacobianFactor(X(3), I_3x3, X(4), -I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(3), I_3x3, X(4), -I_3x3, Z_3x1));
|
||||||
hfg.add(JacobianFactor(X(4), I_3x3, X(5), -I_3x3, Z_3x1));
|
hfg.add(JacobianFactor(X(4), I_3x3, X(5), -I_3x3, Z_3x1));
|
||||||
|
|
||||||
{
|
{
|
||||||
DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
||||||
C(3), boost::make_shared<JacobianFactor>(X(3), I_3x3, Z_3x1),
|
M(3), boost::make_shared<JacobianFactor>(X(3), I_3x3, Z_3x1),
|
||||||
boost::make_shared<JacobianFactor>(X(3), I_3x3, Vector3::Ones()));
|
boost::make_shared<JacobianFactor>(X(3), I_3x3, Vector3::Ones()));
|
||||||
|
|
||||||
hfg.add(GaussianMixtureFactor({X(3)}, {{C(3), 2}}, dt));
|
hfg.add(GaussianMixtureFactor({X(3)}, {{M(3), 2}}, dt));
|
||||||
|
|
||||||
DecisionTree<Key, GaussianFactor::shared_ptr> dt1(
|
DecisionTree<Key, GaussianFactor::shared_ptr> dt1(
|
||||||
C(2), boost::make_shared<JacobianFactor>(X(5), I_3x3, Z_3x1),
|
M(2), boost::make_shared<JacobianFactor>(X(5), I_3x3, Z_3x1),
|
||||||
boost::make_shared<JacobianFactor>(X(5), I_3x3, Vector3::Ones()));
|
boost::make_shared<JacobianFactor>(X(5), I_3x3, Vector3::Ones()));
|
||||||
|
|
||||||
hfg.add(GaussianMixtureFactor({X(5)}, {{C(2), 2}}, dt1));
|
hfg.add(GaussianMixtureFactor({X(5)}, {{M(2), 2}}, dt1));
|
||||||
}
|
}
|
||||||
|
|
||||||
auto ordering_full =
|
auto ordering_full =
|
||||||
Ordering::ColamdConstrainedLast(hfg, {C(0), C(1), C(2), C(3)});
|
Ordering::ColamdConstrainedLast(hfg, {M(0), M(1), M(2), M(3)});
|
||||||
|
|
||||||
GTSAM_PRINT(ordering_full);
|
|
||||||
|
|
||||||
HybridBayesTree::shared_ptr hbt;
|
HybridBayesTree::shared_ptr hbt;
|
||||||
HybridGaussianFactorGraph::shared_ptr remaining;
|
HybridGaussianFactorGraph::shared_ptr remaining;
|
||||||
std::tie(hbt, remaining) = hfg.eliminatePartialMultifrontal(ordering_full);
|
std::tie(hbt, remaining) = hfg.eliminatePartialMultifrontal(ordering_full);
|
||||||
|
|
||||||
GTSAM_PRINT(*hbt);
|
// 9 cliques in the bayes tree and 0 remaining variables to eliminate.
|
||||||
|
EXPECT_LONGS_EQUAL(9, hbt->size());
|
||||||
|
EXPECT_LONGS_EQUAL(0, remaining->size());
|
||||||
|
|
||||||
GTSAM_PRINT(*remaining);
|
|
||||||
|
|
||||||
hbt->dot(std::cout);
|
|
||||||
/*
|
/*
|
||||||
Explanation: the Junction tree will need to reeliminate to get to the marginal
|
(Fan) Explanation: the Junction tree will need to reeliminate to get to the
|
||||||
on X(1), which is not possible because it involves eliminating discrete before
|
marginal on X(1), which is not possible because it involves eliminating
|
||||||
continuous. The solution to this, however, is in Murphy02. TLDR is that this
|
discrete before continuous. The solution to this, however, is in Murphy02.
|
||||||
is 1. expensive and 2. inexact. neverless it is doable. And I believe that we
|
TLDR is that this is 1. expensive and 2. inexact. nevertheless it is doable.
|
||||||
should do this.
|
And I believe that we should do this.
|
||||||
*/
|
*/
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void dotPrint(const HybridGaussianFactorGraph::shared_ptr &hfg,
|
||||||
|
const HybridBayesTree::shared_ptr &hbt,
|
||||||
|
const Ordering &ordering) {
|
||||||
|
DotWriter dw;
|
||||||
|
dw.positionHints['c'] = 2;
|
||||||
|
dw.positionHints['x'] = 1;
|
||||||
|
std::cout << hfg->dot(DefaultKeyFormatter, dw);
|
||||||
|
std::cout << "\n";
|
||||||
|
hbt->dot(std::cout);
|
||||||
|
|
||||||
|
std::cout << "\n";
|
||||||
|
std::cout << hfg->eliminateSequential(ordering)->dot(DefaultKeyFormatter, dw);
|
||||||
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
// TODO(fan): make a graph like Varun's paper one
|
// TODO(fan): make a graph like Varun's paper one
|
||||||
TEST(HybridGaussianFactorGraph, Switching) {
|
TEST(HybridGaussianFactorGraph, Switching) {
|
||||||
|
|
@ -303,13 +311,13 @@ TEST(HybridGaussianFactorGraph, Switching) {
|
||||||
// X(3), X(7)
|
// X(3), X(7)
|
||||||
// X(2), X(8)
|
// X(2), X(8)
|
||||||
// X(1), X(4), X(6), X(9)
|
// X(1), X(4), X(6), X(9)
|
||||||
// C(5) will be the center, C(1-4), C(6-8)
|
// M(5) will be the center, M(1-4), M(6-8)
|
||||||
// C(3), C(7)
|
// M(3), M(7)
|
||||||
// C(1), C(4), C(2), C(6), C(8)
|
// M(1), M(4), M(2), M(6), M(8)
|
||||||
// auto ordering_full =
|
// auto ordering_full =
|
||||||
// Ordering(KeyVector{X(1), X(4), X(2), X(6), X(9), X(8), X(3), X(7),
|
// Ordering(KeyVector{X(1), X(4), X(2), X(6), X(9), X(8), X(3), X(7),
|
||||||
// X(5),
|
// X(5),
|
||||||
// C(1), C(4), C(2), C(6), C(8), C(3), C(7), C(5)});
|
// M(1), M(4), M(2), M(6), M(8), M(3), M(7), M(5)});
|
||||||
KeyVector ordering;
|
KeyVector ordering;
|
||||||
|
|
||||||
{
|
{
|
||||||
|
|
@ -326,79 +334,32 @@ TEST(HybridGaussianFactorGraph, Switching) {
|
||||||
for (auto &l : lvls) {
|
for (auto &l : lvls) {
|
||||||
l = -l;
|
l = -l;
|
||||||
}
|
}
|
||||||
std::copy(lvls.begin(), lvls.end(),
|
|
||||||
std::ostream_iterator<int>(std::cout, ","));
|
|
||||||
std::cout << "\n";
|
|
||||||
}
|
}
|
||||||
{
|
{
|
||||||
std::vector<int> naturalC(N - 1);
|
std::vector<int> naturalC(N - 1);
|
||||||
std::iota(naturalC.begin(), naturalC.end(), 1);
|
std::iota(naturalC.begin(), naturalC.end(), 1);
|
||||||
std::vector<Key> ordC;
|
std::vector<Key> ordC;
|
||||||
std::transform(naturalC.begin(), naturalC.end(), std::back_inserter(ordC),
|
std::transform(naturalC.begin(), naturalC.end(), std::back_inserter(ordC),
|
||||||
[](int x) { return C(x); });
|
[](int x) { return M(x); });
|
||||||
KeyVector ndC;
|
KeyVector ndC;
|
||||||
std::vector<int> lvls;
|
std::vector<int> lvls;
|
||||||
|
|
||||||
// std::copy(ordC.begin(), ordC.end(), std::back_inserter(ordering));
|
// std::copy(ordC.begin(), ordC.end(), std::back_inserter(ordering));
|
||||||
std::tie(ndC, lvls) = makeBinaryOrdering(ordC);
|
std::tie(ndC, lvls) = makeBinaryOrdering(ordC);
|
||||||
std::copy(ndC.begin(), ndC.end(), std::back_inserter(ordering));
|
std::copy(ndC.begin(), ndC.end(), std::back_inserter(ordering));
|
||||||
std::copy(lvls.begin(), lvls.end(),
|
|
||||||
std::ostream_iterator<int>(std::cout, ","));
|
|
||||||
}
|
}
|
||||||
auto ordering_full = Ordering(ordering);
|
auto ordering_full = Ordering(ordering);
|
||||||
|
|
||||||
// auto ordering_full =
|
|
||||||
// Ordering();
|
|
||||||
|
|
||||||
// for (int i = 1; i <= 9; i++) {
|
|
||||||
// ordering_full.push_back(X(i));
|
|
||||||
// }
|
|
||||||
|
|
||||||
// for (int i = 1; i < 9; i++) {
|
|
||||||
// ordering_full.push_back(C(i));
|
|
||||||
// }
|
|
||||||
|
|
||||||
// auto ordering_full =
|
|
||||||
// Ordering(KeyVector{X(1), X(4), X(2), X(6), X(9), X(8), X(3), X(7),
|
|
||||||
// X(5),
|
|
||||||
// C(1), C(2), C(3), C(4), C(5), C(6), C(7), C(8)});
|
|
||||||
|
|
||||||
// GTSAM_PRINT(*hfg);
|
// GTSAM_PRINT(*hfg);
|
||||||
GTSAM_PRINT(ordering_full);
|
// GTSAM_PRINT(ordering_full);
|
||||||
|
|
||||||
HybridBayesTree::shared_ptr hbt;
|
HybridBayesTree::shared_ptr hbt;
|
||||||
HybridGaussianFactorGraph::shared_ptr remaining;
|
HybridGaussianFactorGraph::shared_ptr remaining;
|
||||||
std::tie(hbt, remaining) = hfg->eliminatePartialMultifrontal(ordering_full);
|
std::tie(hbt, remaining) = hfg->eliminatePartialMultifrontal(ordering_full);
|
||||||
|
|
||||||
// GTSAM_PRINT(*hbt);
|
// 12 cliques in the bayes tree and 0 remaining variables to eliminate.
|
||||||
|
EXPECT_LONGS_EQUAL(12, hbt->size());
|
||||||
// GTSAM_PRINT(*remaining);
|
EXPECT_LONGS_EQUAL(0, remaining->size());
|
||||||
|
|
||||||
{
|
|
||||||
DotWriter dw;
|
|
||||||
dw.positionHints['c'] = 2;
|
|
||||||
dw.positionHints['x'] = 1;
|
|
||||||
std::cout << hfg->dot(DefaultKeyFormatter, dw);
|
|
||||||
std::cout << "\n";
|
|
||||||
hbt->dot(std::cout);
|
|
||||||
}
|
|
||||||
|
|
||||||
{
|
|
||||||
DotWriter dw;
|
|
||||||
// dw.positionHints['c'] = 2;
|
|
||||||
// dw.positionHints['x'] = 1;
|
|
||||||
std::cout << "\n";
|
|
||||||
std::cout << hfg->eliminateSequential(ordering_full)
|
|
||||||
->dot(DefaultKeyFormatter, dw);
|
|
||||||
}
|
|
||||||
/*
|
|
||||||
Explanation: the Junction tree will need to reeliminate to get to the marginal
|
|
||||||
on X(1), which is not possible because it involves eliminating discrete before
|
|
||||||
continuous. The solution to this, however, is in Murphy02. TLDR is that this
|
|
||||||
is 1. expensive and 2. inexact. neverless it is doable. And I believe that we
|
|
||||||
should do this.
|
|
||||||
*/
|
|
||||||
hbt->marginalFactor(C(11))->print("HBT: ");
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
|
|
@ -411,13 +372,13 @@ TEST(HybridGaussianFactorGraph, SwitchingISAM) {
|
||||||
// X(3), X(7)
|
// X(3), X(7)
|
||||||
// X(2), X(8)
|
// X(2), X(8)
|
||||||
// X(1), X(4), X(6), X(9)
|
// X(1), X(4), X(6), X(9)
|
||||||
// C(5) will be the center, C(1-4), C(6-8)
|
// M(5) will be the center, M(1-4), M(6-8)
|
||||||
// C(3), C(7)
|
// M(3), M(7)
|
||||||
// C(1), C(4), C(2), C(6), C(8)
|
// M(1), M(4), M(2), M(6), M(8)
|
||||||
// auto ordering_full =
|
// auto ordering_full =
|
||||||
// Ordering(KeyVector{X(1), X(4), X(2), X(6), X(9), X(8), X(3), X(7),
|
// Ordering(KeyVector{X(1), X(4), X(2), X(6), X(9), X(8), X(3), X(7),
|
||||||
// X(5),
|
// X(5),
|
||||||
// C(1), C(4), C(2), C(6), C(8), C(3), C(7), C(5)});
|
// M(1), M(4), M(2), M(6), M(8), M(3), M(7), M(5)});
|
||||||
KeyVector ordering;
|
KeyVector ordering;
|
||||||
|
|
||||||
{
|
{
|
||||||
|
|
@ -434,67 +395,37 @@ TEST(HybridGaussianFactorGraph, SwitchingISAM) {
|
||||||
for (auto &l : lvls) {
|
for (auto &l : lvls) {
|
||||||
l = -l;
|
l = -l;
|
||||||
}
|
}
|
||||||
std::copy(lvls.begin(), lvls.end(),
|
|
||||||
std::ostream_iterator<int>(std::cout, ","));
|
|
||||||
std::cout << "\n";
|
|
||||||
}
|
}
|
||||||
{
|
{
|
||||||
std::vector<int> naturalC(N - 1);
|
std::vector<int> naturalC(N - 1);
|
||||||
std::iota(naturalC.begin(), naturalC.end(), 1);
|
std::iota(naturalC.begin(), naturalC.end(), 1);
|
||||||
std::vector<Key> ordC;
|
std::vector<Key> ordC;
|
||||||
std::transform(naturalC.begin(), naturalC.end(), std::back_inserter(ordC),
|
std::transform(naturalC.begin(), naturalC.end(), std::back_inserter(ordC),
|
||||||
[](int x) { return C(x); });
|
[](int x) { return M(x); });
|
||||||
KeyVector ndC;
|
KeyVector ndC;
|
||||||
std::vector<int> lvls;
|
std::vector<int> lvls;
|
||||||
|
|
||||||
// std::copy(ordC.begin(), ordC.end(), std::back_inserter(ordering));
|
// std::copy(ordC.begin(), ordC.end(), std::back_inserter(ordering));
|
||||||
std::tie(ndC, lvls) = makeBinaryOrdering(ordC);
|
std::tie(ndC, lvls) = makeBinaryOrdering(ordC);
|
||||||
std::copy(ndC.begin(), ndC.end(), std::back_inserter(ordering));
|
std::copy(ndC.begin(), ndC.end(), std::back_inserter(ordering));
|
||||||
std::copy(lvls.begin(), lvls.end(),
|
|
||||||
std::ostream_iterator<int>(std::cout, ","));
|
|
||||||
}
|
}
|
||||||
auto ordering_full = Ordering(ordering);
|
auto ordering_full = Ordering(ordering);
|
||||||
|
|
||||||
// GTSAM_PRINT(*hfg);
|
|
||||||
GTSAM_PRINT(ordering_full);
|
|
||||||
|
|
||||||
HybridBayesTree::shared_ptr hbt;
|
HybridBayesTree::shared_ptr hbt;
|
||||||
HybridGaussianFactorGraph::shared_ptr remaining;
|
HybridGaussianFactorGraph::shared_ptr remaining;
|
||||||
std::tie(hbt, remaining) = hfg->eliminatePartialMultifrontal(ordering_full);
|
std::tie(hbt, remaining) = hfg->eliminatePartialMultifrontal(ordering_full);
|
||||||
|
|
||||||
// GTSAM_PRINT(*hbt);
|
|
||||||
|
|
||||||
// GTSAM_PRINT(*remaining);
|
|
||||||
|
|
||||||
{
|
|
||||||
DotWriter dw;
|
|
||||||
dw.positionHints['c'] = 2;
|
|
||||||
dw.positionHints['x'] = 1;
|
|
||||||
std::cout << hfg->dot(DefaultKeyFormatter, dw);
|
|
||||||
std::cout << "\n";
|
|
||||||
hbt->dot(std::cout);
|
|
||||||
}
|
|
||||||
|
|
||||||
{
|
|
||||||
DotWriter dw;
|
|
||||||
// dw.positionHints['c'] = 2;
|
|
||||||
// dw.positionHints['x'] = 1;
|
|
||||||
std::cout << "\n";
|
|
||||||
std::cout << hfg->eliminateSequential(ordering_full)
|
|
||||||
->dot(DefaultKeyFormatter, dw);
|
|
||||||
}
|
|
||||||
|
|
||||||
auto new_fg = makeSwitchingChain(12);
|
auto new_fg = makeSwitchingChain(12);
|
||||||
auto isam = HybridGaussianISAM(*hbt);
|
auto isam = HybridGaussianISAM(*hbt);
|
||||||
|
|
||||||
{
|
// Run an ISAM update.
|
||||||
HybridGaussianFactorGraph factorGraph;
|
HybridGaussianFactorGraph factorGraph;
|
||||||
factorGraph.push_back(new_fg->at(new_fg->size() - 2));
|
factorGraph.push_back(new_fg->at(new_fg->size() - 2));
|
||||||
factorGraph.push_back(new_fg->at(new_fg->size() - 1));
|
factorGraph.push_back(new_fg->at(new_fg->size() - 1));
|
||||||
isam.update(factorGraph);
|
isam.update(factorGraph);
|
||||||
std::cout << isam.dot();
|
|
||||||
isam.marginalFactor(C(11))->print();
|
// ISAM should have 12 factors after the last update
|
||||||
}
|
EXPECT_LONGS_EQUAL(12, isam.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
|
|
@ -517,23 +448,8 @@ TEST(HybridGaussianFactorGraph, SwitchingTwoVar) {
|
||||||
ordX.emplace_back(Y(i));
|
ordX.emplace_back(Y(i));
|
||||||
}
|
}
|
||||||
|
|
||||||
// {
|
|
||||||
// KeyVector ndX;
|
|
||||||
// std::vector<int> lvls;
|
|
||||||
// std::tie(ndX, lvls) = makeBinaryOrdering(naturalX);
|
|
||||||
// std::copy(ndX.begin(), ndX.end(), std::back_inserter(ordering));
|
|
||||||
// std::copy(lvls.begin(), lvls.end(),
|
|
||||||
// std::ostream_iterator<int>(std::cout, ","));
|
|
||||||
// std::cout << "\n";
|
|
||||||
|
|
||||||
// for (size_t i = 0; i < N; i++) {
|
|
||||||
// ordX.emplace_back(X(ndX[i]));
|
|
||||||
// ordX.emplace_back(Y(ndX[i]));
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|
||||||
for (size_t i = 1; i <= N - 1; i++) {
|
for (size_t i = 1; i <= N - 1; i++) {
|
||||||
ordX.emplace_back(C(i));
|
ordX.emplace_back(M(i));
|
||||||
}
|
}
|
||||||
for (size_t i = 1; i <= N - 1; i++) {
|
for (size_t i = 1; i <= N - 1; i++) {
|
||||||
ordX.emplace_back(D(i));
|
ordX.emplace_back(D(i));
|
||||||
|
|
@ -545,8 +461,8 @@ TEST(HybridGaussianFactorGraph, SwitchingTwoVar) {
|
||||||
dw.positionHints['c'] = 0;
|
dw.positionHints['c'] = 0;
|
||||||
dw.positionHints['d'] = 3;
|
dw.positionHints['d'] = 3;
|
||||||
dw.positionHints['y'] = 2;
|
dw.positionHints['y'] = 2;
|
||||||
std::cout << hfg->dot(DefaultKeyFormatter, dw);
|
// std::cout << hfg->dot(DefaultKeyFormatter, dw);
|
||||||
std::cout << "\n";
|
// std::cout << "\n";
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
|
|
@ -555,10 +471,10 @@ TEST(HybridGaussianFactorGraph, SwitchingTwoVar) {
|
||||||
// dw.positionHints['c'] = 0;
|
// dw.positionHints['c'] = 0;
|
||||||
// dw.positionHints['d'] = 3;
|
// dw.positionHints['d'] = 3;
|
||||||
dw.positionHints['x'] = 1;
|
dw.positionHints['x'] = 1;
|
||||||
std::cout << "\n";
|
// std::cout << "\n";
|
||||||
// std::cout << hfg->eliminateSequential(Ordering(ordX))
|
// std::cout << hfg->eliminateSequential(Ordering(ordX))
|
||||||
// ->dot(DefaultKeyFormatter, dw);
|
// ->dot(DefaultKeyFormatter, dw);
|
||||||
hfg->eliminateMultifrontal(Ordering(ordX))->dot(std::cout);
|
// hfg->eliminateMultifrontal(Ordering(ordX))->dot(std::cout);
|
||||||
}
|
}
|
||||||
|
|
||||||
Ordering ordering_partial;
|
Ordering ordering_partial;
|
||||||
|
|
@ -566,25 +482,46 @@ TEST(HybridGaussianFactorGraph, SwitchingTwoVar) {
|
||||||
ordering_partial.emplace_back(X(i));
|
ordering_partial.emplace_back(X(i));
|
||||||
ordering_partial.emplace_back(Y(i));
|
ordering_partial.emplace_back(Y(i));
|
||||||
}
|
}
|
||||||
{
|
HybridBayesNet::shared_ptr hbn;
|
||||||
HybridBayesNet::shared_ptr hbn;
|
HybridGaussianFactorGraph::shared_ptr remaining;
|
||||||
HybridGaussianFactorGraph::shared_ptr remaining;
|
std::tie(hbn, remaining) =
|
||||||
std::tie(hbn, remaining) =
|
hfg->eliminatePartialSequential(ordering_partial);
|
||||||
hfg->eliminatePartialSequential(ordering_partial);
|
|
||||||
|
|
||||||
// remaining->print();
|
EXPECT_LONGS_EQUAL(14, hbn->size());
|
||||||
{
|
EXPECT_LONGS_EQUAL(11, remaining->size());
|
||||||
DotWriter dw;
|
|
||||||
dw.positionHints['x'] = 1;
|
{
|
||||||
dw.positionHints['c'] = 0;
|
DotWriter dw;
|
||||||
dw.positionHints['d'] = 3;
|
dw.positionHints['x'] = 1;
|
||||||
dw.positionHints['y'] = 2;
|
dw.positionHints['c'] = 0;
|
||||||
std::cout << remaining->dot(DefaultKeyFormatter, dw);
|
dw.positionHints['d'] = 3;
|
||||||
std::cout << "\n";
|
dw.positionHints['y'] = 2;
|
||||||
}
|
// std::cout << remaining->dot(DefaultKeyFormatter, dw);
|
||||||
|
// std::cout << "\n";
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
TEST(HybridGaussianFactorGraph, optimize) {
|
||||||
|
HybridGaussianFactorGraph hfg;
|
||||||
|
|
||||||
|
DiscreteKey c1(C(1), 2);
|
||||||
|
|
||||||
|
hfg.add(JacobianFactor(X(0), I_3x3, Z_3x1));
|
||||||
|
hfg.add(JacobianFactor(X(0), I_3x3, X(1), -I_3x3, Z_3x1));
|
||||||
|
|
||||||
|
DecisionTree<Key, GaussianFactor::shared_ptr> dt(
|
||||||
|
C(1), boost::make_shared<JacobianFactor>(X(1), I_3x3, Z_3x1),
|
||||||
|
boost::make_shared<JacobianFactor>(X(1), I_3x3, Vector3::Ones()));
|
||||||
|
|
||||||
|
hfg.add(GaussianMixtureFactor({X(1)}, {c1}, dt));
|
||||||
|
|
||||||
|
auto result =
|
||||||
|
hfg.eliminateSequential(Ordering::ColamdConstrainedLast(hfg, {C(1)}));
|
||||||
|
|
||||||
|
HybridValues hv = result->optimize();
|
||||||
|
|
||||||
|
EXPECT(assert_equal(hv.atDiscrete(C(1)), int(0)));
|
||||||
|
}
|
||||||
/* ************************************************************************* */
|
/* ************************************************************************* */
|
||||||
int main() {
|
int main() {
|
||||||
TestResult tr;
|
TestResult tr;
|
||||||
|
|
|
||||||
|
|
@ -74,7 +74,7 @@ TEST(GaussianMixtureFactor, Sum) {
|
||||||
// Check that number of keys is 3
|
// Check that number of keys is 3
|
||||||
EXPECT_LONGS_EQUAL(3, mixtureFactorA.keys().size());
|
EXPECT_LONGS_EQUAL(3, mixtureFactorA.keys().size());
|
||||||
|
|
||||||
// Check that number of discrete keys is 1 // TODO(Frank): should not exist?
|
// Check that number of discrete keys is 1
|
||||||
EXPECT_LONGS_EQUAL(1, mixtureFactorA.discreteKeys().size());
|
EXPECT_LONGS_EQUAL(1, mixtureFactorA.discreteKeys().size());
|
||||||
|
|
||||||
// Create sum of two mixture factors: it will be a decision tree now on both
|
// Create sum of two mixture factors: it will be a decision tree now on both
|
||||||
|
|
@ -104,7 +104,7 @@ TEST(GaussianMixtureFactor, Printing) {
|
||||||
GaussianMixtureFactor mixtureFactor({X(1), X(2)}, {m1}, factors);
|
GaussianMixtureFactor mixtureFactor({X(1), X(2)}, {m1}, factors);
|
||||||
|
|
||||||
std::string expected =
|
std::string expected =
|
||||||
R"(Hybrid x1 x2; 1 ]{
|
R"(Hybrid [x1 x2; 1]{
|
||||||
Choice(1)
|
Choice(1)
|
||||||
0 Leaf :
|
0 Leaf :
|
||||||
A[x1] = [
|
A[x1] = [
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,93 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file testHybridBayesNet.cpp
|
||||||
|
* @brief Unit tests for HybridBayesNet
|
||||||
|
* @author Varun Agrawal
|
||||||
|
* @author Fan Jiang
|
||||||
|
* @author Frank Dellaert
|
||||||
|
* @date December 2021
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <gtsam/hybrid/HybridBayesNet.h>
|
||||||
|
|
||||||
|
#include "Switching.h"
|
||||||
|
|
||||||
|
// Include for test suite
|
||||||
|
#include <CppUnitLite/TestHarness.h>
|
||||||
|
|
||||||
|
using namespace std;
|
||||||
|
using namespace gtsam;
|
||||||
|
using noiseModel::Isotropic;
|
||||||
|
using symbol_shorthand::M;
|
||||||
|
using symbol_shorthand::X;
|
||||||
|
|
||||||
|
static const DiscreteKey Asia(0, 2);
|
||||||
|
|
||||||
|
/* ****************************************************************************/
|
||||||
|
// Test creation
|
||||||
|
TEST(HybridBayesNet, Creation) {
|
||||||
|
HybridBayesNet bayesNet;
|
||||||
|
|
||||||
|
bayesNet.add(Asia, "99/1");
|
||||||
|
|
||||||
|
DiscreteConditional expected(Asia, "99/1");
|
||||||
|
|
||||||
|
CHECK(bayesNet.atDiscrete(0));
|
||||||
|
auto& df = *bayesNet.atDiscrete(0);
|
||||||
|
EXPECT(df.equals(expected));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ****************************************************************************/
|
||||||
|
// Test choosing an assignment of conditionals
|
||||||
|
TEST(HybridBayesNet, Choose) {
|
||||||
|
Switching s(4);
|
||||||
|
|
||||||
|
Ordering ordering;
|
||||||
|
for (auto&& kvp : s.linearizationPoint) {
|
||||||
|
ordering += kvp.key;
|
||||||
|
}
|
||||||
|
|
||||||
|
HybridBayesNet::shared_ptr hybridBayesNet;
|
||||||
|
HybridGaussianFactorGraph::shared_ptr remainingFactorGraph;
|
||||||
|
std::tie(hybridBayesNet, remainingFactorGraph) =
|
||||||
|
s.linearizedFactorGraph.eliminatePartialSequential(ordering);
|
||||||
|
|
||||||
|
DiscreteValues assignment;
|
||||||
|
assignment[M(1)] = 1;
|
||||||
|
assignment[M(2)] = 1;
|
||||||
|
assignment[M(3)] = 0;
|
||||||
|
|
||||||
|
GaussianBayesNet gbn = hybridBayesNet->choose(assignment);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(4, gbn.size());
|
||||||
|
|
||||||
|
EXPECT(assert_equal(*(*boost::dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
hybridBayesNet->atGaussian(0)))(assignment),
|
||||||
|
*gbn.at(0)));
|
||||||
|
EXPECT(assert_equal(*(*boost::dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
hybridBayesNet->atGaussian(1)))(assignment),
|
||||||
|
*gbn.at(1)));
|
||||||
|
EXPECT(assert_equal(*(*boost::dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
hybridBayesNet->atGaussian(2)))(assignment),
|
||||||
|
*gbn.at(2)));
|
||||||
|
EXPECT(assert_equal(*(*boost::dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
hybridBayesNet->atGaussian(3)))(assignment),
|
||||||
|
*gbn.at(3)));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
int main() {
|
||||||
|
TestResult tr;
|
||||||
|
return TestRegistry::runAllTests(tr);
|
||||||
|
}
|
||||||
|
/* ************************************************************************* */
|
||||||
|
|
@ -0,0 +1,45 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
* See LICENSE for the license information
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file testHybridFactorGraph.cpp
|
||||||
|
* @brief Unit tests for HybridFactorGraph
|
||||||
|
* @author Varun Agrawal
|
||||||
|
* @date May 2021
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <CppUnitLite/Test.h>
|
||||||
|
#include <CppUnitLite/TestHarness.h>
|
||||||
|
#include <gtsam/base/TestableAssertions.h>
|
||||||
|
#include <gtsam/base/utilities.h>
|
||||||
|
#include <gtsam/hybrid/HybridFactorGraph.h>
|
||||||
|
#include <gtsam/inference/Symbol.h>
|
||||||
|
#include <gtsam/nonlinear/PriorFactor.h>
|
||||||
|
|
||||||
|
using namespace std;
|
||||||
|
using namespace gtsam;
|
||||||
|
using noiseModel::Isotropic;
|
||||||
|
using symbol_shorthand::L;
|
||||||
|
using symbol_shorthand::M;
|
||||||
|
using symbol_shorthand::X;
|
||||||
|
|
||||||
|
/* ****************************************************************************
|
||||||
|
* Test that any linearizedFactorGraph gaussian factors are appended to the
|
||||||
|
* existing gaussian factor graph in the hybrid factor graph.
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, Constructor) {
|
||||||
|
// Initialize the hybrid factor graph
|
||||||
|
HybridFactorGraph fg;
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
int main() {
|
||||||
|
TestResult tr;
|
||||||
|
return TestRegistry::runAllTests(tr);
|
||||||
|
}
|
||||||
|
/* ************************************************************************* */
|
||||||
|
|
@ -0,0 +1,563 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file testHybridIncremental.cpp
|
||||||
|
* @brief Unit tests for incremental inference
|
||||||
|
* @author Fan Jiang, Varun Agrawal, Frank Dellaert
|
||||||
|
* @date Jan 2021
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <gtsam/discrete/DiscreteBayesNet.h>
|
||||||
|
#include <gtsam/discrete/DiscreteDistribution.h>
|
||||||
|
#include <gtsam/discrete/DiscreteFactorGraph.h>
|
||||||
|
#include <gtsam/geometry/Pose2.h>
|
||||||
|
#include <gtsam/hybrid/HybridConditional.h>
|
||||||
|
#include <gtsam/hybrid/HybridGaussianISAM.h>
|
||||||
|
#include <gtsam/linear/GaussianBayesNet.h>
|
||||||
|
#include <gtsam/linear/GaussianFactorGraph.h>
|
||||||
|
#include <gtsam/nonlinear/PriorFactor.h>
|
||||||
|
#include <gtsam/sam/BearingRangeFactor.h>
|
||||||
|
|
||||||
|
#include <numeric>
|
||||||
|
|
||||||
|
#include "Switching.h"
|
||||||
|
|
||||||
|
// Include for test suite
|
||||||
|
#include <CppUnitLite/TestHarness.h>
|
||||||
|
|
||||||
|
using namespace std;
|
||||||
|
using namespace gtsam;
|
||||||
|
using noiseModel::Isotropic;
|
||||||
|
using symbol_shorthand::L;
|
||||||
|
using symbol_shorthand::M;
|
||||||
|
using symbol_shorthand::W;
|
||||||
|
using symbol_shorthand::X;
|
||||||
|
using symbol_shorthand::Y;
|
||||||
|
using symbol_shorthand::Z;
|
||||||
|
|
||||||
|
/* ****************************************************************************/
|
||||||
|
// Test if we can perform elimination incrementally.
|
||||||
|
TEST(HybridGaussianElimination, IncrementalElimination) {
|
||||||
|
Switching switching(3);
|
||||||
|
HybridGaussianISAM isam;
|
||||||
|
HybridGaussianFactorGraph graph1;
|
||||||
|
|
||||||
|
// Create initial factor graph
|
||||||
|
// * * *
|
||||||
|
// | | |
|
||||||
|
// X1 -*- X2 -*- X3
|
||||||
|
// \*-M1-*/
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(0)); // P(X1)
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(1)); // P(X1, X2 | M1)
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(2)); // P(X2, X3 | M2)
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(5)); // P(M1)
|
||||||
|
|
||||||
|
// Run update step
|
||||||
|
isam.update(graph1);
|
||||||
|
|
||||||
|
// Check that after update we have 3 hybrid Bayes net nodes:
|
||||||
|
// P(X1 | X2, M1) and P(X2, X3 | M1, M2), P(M1, M2)
|
||||||
|
EXPECT_LONGS_EQUAL(3, isam.size());
|
||||||
|
EXPECT(isam[X(1)]->conditional()->frontals() == KeyVector{X(1)});
|
||||||
|
EXPECT(isam[X(1)]->conditional()->parents() == KeyVector({X(2), M(1)}));
|
||||||
|
EXPECT(isam[X(2)]->conditional()->frontals() == KeyVector({X(2), X(3)}));
|
||||||
|
EXPECT(isam[X(2)]->conditional()->parents() == KeyVector({M(1), M(2)}));
|
||||||
|
|
||||||
|
/********************************************************/
|
||||||
|
// New factor graph for incremental update.
|
||||||
|
HybridGaussianFactorGraph graph2;
|
||||||
|
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(3)); // P(X2)
|
||||||
|
graph2.push_back(switching.linearizedFactorGraph.at(4)); // P(X3)
|
||||||
|
graph2.push_back(switching.linearizedFactorGraph.at(6)); // P(M1, M2)
|
||||||
|
|
||||||
|
isam.update(graph2);
|
||||||
|
|
||||||
|
// Check that after the second update we have
|
||||||
|
// 1 additional hybrid Bayes net node:
|
||||||
|
// P(X2, X3 | M1, M2)
|
||||||
|
EXPECT_LONGS_EQUAL(3, isam.size());
|
||||||
|
EXPECT(isam[X(3)]->conditional()->frontals() == KeyVector({X(2), X(3)}));
|
||||||
|
EXPECT(isam[X(3)]->conditional()->parents() == KeyVector({M(1), M(2)}));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ****************************************************************************/
|
||||||
|
// Test if we can incrementally do the inference
|
||||||
|
TEST(HybridGaussianElimination, IncrementalInference) {
|
||||||
|
Switching switching(3);
|
||||||
|
HybridGaussianISAM isam;
|
||||||
|
HybridGaussianFactorGraph graph1;
|
||||||
|
|
||||||
|
// Create initial factor graph
|
||||||
|
// * * *
|
||||||
|
// | | |
|
||||||
|
// X1 -*- X2 -*- X3
|
||||||
|
// | |
|
||||||
|
// *-M1 - * - M2
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(0)); // P(X1)
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(1)); // P(X1, X2 | M1)
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(3)); // P(X2)
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(5)); // P(M1)
|
||||||
|
|
||||||
|
// Run update step
|
||||||
|
isam.update(graph1);
|
||||||
|
|
||||||
|
auto discreteConditional_m1 =
|
||||||
|
isam[M(1)]->conditional()->asDiscreteConditional();
|
||||||
|
EXPECT(discreteConditional_m1->keys() == KeyVector({M(1)}));
|
||||||
|
|
||||||
|
/********************************************************/
|
||||||
|
// New factor graph for incremental update.
|
||||||
|
HybridGaussianFactorGraph graph2;
|
||||||
|
|
||||||
|
graph2.push_back(switching.linearizedFactorGraph.at(2)); // P(X2, X3 | M2)
|
||||||
|
graph2.push_back(switching.linearizedFactorGraph.at(4)); // P(X3)
|
||||||
|
graph2.push_back(switching.linearizedFactorGraph.at(6)); // P(M1, M2)
|
||||||
|
|
||||||
|
isam.update(graph2);
|
||||||
|
|
||||||
|
/********************************************************/
|
||||||
|
// Run batch elimination so we can compare results.
|
||||||
|
Ordering ordering;
|
||||||
|
ordering += X(1);
|
||||||
|
ordering += X(2);
|
||||||
|
ordering += X(3);
|
||||||
|
|
||||||
|
// Now we calculate the actual factors using full elimination
|
||||||
|
HybridBayesTree::shared_ptr expectedHybridBayesTree;
|
||||||
|
HybridGaussianFactorGraph::shared_ptr expectedRemainingGraph;
|
||||||
|
std::tie(expectedHybridBayesTree, expectedRemainingGraph) =
|
||||||
|
switching.linearizedFactorGraph.eliminatePartialMultifrontal(ordering);
|
||||||
|
|
||||||
|
// The densities on X(1) should be the same
|
||||||
|
auto x1_conditional =
|
||||||
|
dynamic_pointer_cast<GaussianMixture>(isam[X(1)]->conditional()->inner());
|
||||||
|
auto actual_x1_conditional = dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
(*expectedHybridBayesTree)[X(1)]->conditional()->inner());
|
||||||
|
EXPECT(assert_equal(*x1_conditional, *actual_x1_conditional));
|
||||||
|
|
||||||
|
// The densities on X(2) should be the same
|
||||||
|
auto x2_conditional =
|
||||||
|
dynamic_pointer_cast<GaussianMixture>(isam[X(2)]->conditional()->inner());
|
||||||
|
auto actual_x2_conditional = dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
(*expectedHybridBayesTree)[X(2)]->conditional()->inner());
|
||||||
|
EXPECT(assert_equal(*x2_conditional, *actual_x2_conditional));
|
||||||
|
|
||||||
|
// The densities on X(3) should be the same
|
||||||
|
auto x3_conditional =
|
||||||
|
dynamic_pointer_cast<GaussianMixture>(isam[X(3)]->conditional()->inner());
|
||||||
|
auto actual_x3_conditional = dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
(*expectedHybridBayesTree)[X(2)]->conditional()->inner());
|
||||||
|
EXPECT(assert_equal(*x3_conditional, *actual_x3_conditional));
|
||||||
|
|
||||||
|
// We only perform manual continuous elimination for 0,0.
|
||||||
|
// The other discrete probabilities on M(2) are calculated the same way
|
||||||
|
Ordering discrete_ordering;
|
||||||
|
discrete_ordering += M(1);
|
||||||
|
discrete_ordering += M(2);
|
||||||
|
HybridBayesTree::shared_ptr discreteBayesTree =
|
||||||
|
expectedRemainingGraph->eliminateMultifrontal(discrete_ordering);
|
||||||
|
|
||||||
|
DiscreteValues m00;
|
||||||
|
m00[M(1)] = 0, m00[M(2)] = 0;
|
||||||
|
DiscreteConditional decisionTree =
|
||||||
|
*(*discreteBayesTree)[M(2)]->conditional()->asDiscreteConditional();
|
||||||
|
double m00_prob = decisionTree(m00);
|
||||||
|
|
||||||
|
auto discreteConditional = isam[M(2)]->conditional()->asDiscreteConditional();
|
||||||
|
|
||||||
|
// Test if the probability values are as expected with regression tests.
|
||||||
|
DiscreteValues assignment;
|
||||||
|
EXPECT(assert_equal(m00_prob, 0.0619233, 1e-5));
|
||||||
|
assignment[M(1)] = 0;
|
||||||
|
assignment[M(2)] = 0;
|
||||||
|
EXPECT(assert_equal(m00_prob, (*discreteConditional)(assignment), 1e-5));
|
||||||
|
assignment[M(1)] = 1;
|
||||||
|
assignment[M(2)] = 0;
|
||||||
|
EXPECT(assert_equal(0.183743, (*discreteConditional)(assignment), 1e-5));
|
||||||
|
assignment[M(1)] = 0;
|
||||||
|
assignment[M(2)] = 1;
|
||||||
|
EXPECT(assert_equal(0.204159, (*discreteConditional)(assignment), 1e-5));
|
||||||
|
assignment[M(1)] = 1;
|
||||||
|
assignment[M(2)] = 1;
|
||||||
|
EXPECT(assert_equal(0.2, (*discreteConditional)(assignment), 1e-5));
|
||||||
|
|
||||||
|
// Check if the clique conditional generated from incremental elimination
|
||||||
|
// matches that of batch elimination.
|
||||||
|
auto expectedChordal = expectedRemainingGraph->eliminateMultifrontal();
|
||||||
|
auto expectedConditional = dynamic_pointer_cast<DecisionTreeFactor>(
|
||||||
|
(*expectedChordal)[M(2)]->conditional()->inner());
|
||||||
|
auto actualConditional = dynamic_pointer_cast<DecisionTreeFactor>(
|
||||||
|
isam[M(2)]->conditional()->inner());
|
||||||
|
EXPECT(assert_equal(*actualConditional, *expectedConditional, 1e-6));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ****************************************************************************/
|
||||||
|
// Test if we can approximately do the inference
|
||||||
|
TEST(HybridGaussianElimination, Approx_inference) {
|
||||||
|
Switching switching(4);
|
||||||
|
HybridGaussianISAM incrementalHybrid;
|
||||||
|
HybridGaussianFactorGraph graph1;
|
||||||
|
|
||||||
|
// Add the 3 hybrid factors, x1-x2, x2-x3, x3-x4
|
||||||
|
for (size_t i = 1; i < 4; i++) {
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(i));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add the Gaussian factors, 1 prior on X(1),
|
||||||
|
// 3 measurements on X(2), X(3), X(4)
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(0));
|
||||||
|
for (size_t i = 4; i <= 7; i++) {
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(i));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create ordering.
|
||||||
|
Ordering ordering;
|
||||||
|
for (size_t j = 1; j <= 4; j++) {
|
||||||
|
ordering += X(j);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now we calculate the actual factors using full elimination
|
||||||
|
HybridBayesTree::shared_ptr unprunedHybridBayesTree;
|
||||||
|
HybridGaussianFactorGraph::shared_ptr unprunedRemainingGraph;
|
||||||
|
std::tie(unprunedHybridBayesTree, unprunedRemainingGraph) =
|
||||||
|
switching.linearizedFactorGraph.eliminatePartialMultifrontal(ordering);
|
||||||
|
|
||||||
|
size_t maxNrLeaves = 5;
|
||||||
|
incrementalHybrid.update(graph1);
|
||||||
|
|
||||||
|
incrementalHybrid.prune(M(3), maxNrLeaves);
|
||||||
|
|
||||||
|
/*
|
||||||
|
unpruned factor is:
|
||||||
|
Choice(m3)
|
||||||
|
0 Choice(m2)
|
||||||
|
0 0 Choice(m1)
|
||||||
|
0 0 0 Leaf 0.11267528
|
||||||
|
0 0 1 Leaf 0.18576102
|
||||||
|
0 1 Choice(m1)
|
||||||
|
0 1 0 Leaf 0.18754662
|
||||||
|
0 1 1 Leaf 0.30623871
|
||||||
|
1 Choice(m2)
|
||||||
|
1 0 Choice(m1)
|
||||||
|
1 0 0 Leaf 0.18576102
|
||||||
|
1 0 1 Leaf 0.30622428
|
||||||
|
1 1 Choice(m1)
|
||||||
|
1 1 0 Leaf 0.30623871
|
||||||
|
1 1 1 Leaf 0.5
|
||||||
|
|
||||||
|
pruned factors is:
|
||||||
|
Choice(m3)
|
||||||
|
0 Choice(m2)
|
||||||
|
0 0 Leaf 0
|
||||||
|
0 1 Choice(m1)
|
||||||
|
0 1 0 Leaf 0.18754662
|
||||||
|
0 1 1 Leaf 0.30623871
|
||||||
|
1 Choice(m2)
|
||||||
|
1 0 Choice(m1)
|
||||||
|
1 0 0 Leaf 0
|
||||||
|
1 0 1 Leaf 0.30622428
|
||||||
|
1 1 Choice(m1)
|
||||||
|
1 1 0 Leaf 0.30623871
|
||||||
|
1 1 1 Leaf 0.5
|
||||||
|
*/
|
||||||
|
|
||||||
|
auto discreteConditional_m1 = *dynamic_pointer_cast<DiscreteConditional>(
|
||||||
|
incrementalHybrid[M(1)]->conditional()->inner());
|
||||||
|
EXPECT(discreteConditional_m1.keys() == KeyVector({M(1), M(2), M(3)}));
|
||||||
|
|
||||||
|
// Get the number of elements which are greater than 0.
|
||||||
|
auto count = [](const double &value, int count) {
|
||||||
|
return value > 0 ? count + 1 : count;
|
||||||
|
};
|
||||||
|
// Check that the number of leaves after pruning is 5.
|
||||||
|
EXPECT_LONGS_EQUAL(5, discreteConditional_m1.fold(count, 0));
|
||||||
|
|
||||||
|
// Check that the hybrid nodes of the bayes net match those of the pre-pruning
|
||||||
|
// bayes net, at the same positions.
|
||||||
|
auto &unprunedLastDensity = *dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
unprunedHybridBayesTree->clique(X(4))->conditional()->inner());
|
||||||
|
auto &lastDensity = *dynamic_pointer_cast<GaussianMixture>(
|
||||||
|
incrementalHybrid[X(4)]->conditional()->inner());
|
||||||
|
|
||||||
|
std::vector<std::pair<DiscreteValues, double>> assignments =
|
||||||
|
discreteConditional_m1.enumerate();
|
||||||
|
// Loop over all assignments and check the pruned components
|
||||||
|
for (auto &&av : assignments) {
|
||||||
|
const DiscreteValues &assignment = av.first;
|
||||||
|
const double value = av.second;
|
||||||
|
|
||||||
|
if (value == 0.0) {
|
||||||
|
EXPECT(lastDensity(assignment) == nullptr);
|
||||||
|
} else {
|
||||||
|
CHECK(lastDensity(assignment));
|
||||||
|
EXPECT(assert_equal(*unprunedLastDensity(assignment),
|
||||||
|
*lastDensity(assignment)));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ****************************************************************************/
|
||||||
|
// Test approximate inference with an additional pruning step.
|
||||||
|
TEST(HybridGaussianElimination, Incremental_approximate) {
|
||||||
|
Switching switching(5);
|
||||||
|
HybridGaussianISAM incrementalHybrid;
|
||||||
|
HybridGaussianFactorGraph graph1;
|
||||||
|
|
||||||
|
/***** Run Round 1 *****/
|
||||||
|
// Add the 3 hybrid factors, x1-x2, x2-x3, x3-x4
|
||||||
|
for (size_t i = 1; i < 4; i++) {
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(i));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add the Gaussian factors, 1 prior on X(1),
|
||||||
|
// 3 measurements on X(2), X(3), X(4)
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(0));
|
||||||
|
for (size_t i = 5; i <= 7; i++) {
|
||||||
|
graph1.push_back(switching.linearizedFactorGraph.at(i));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Run update with pruning
|
||||||
|
size_t maxComponents = 5;
|
||||||
|
incrementalHybrid.update(graph1);
|
||||||
|
incrementalHybrid.prune(M(3), maxComponents);
|
||||||
|
|
||||||
|
// Check if we have a bayes tree with 4 hybrid nodes,
|
||||||
|
// each with 2, 4, 8, and 5 (pruned) leaves respetively.
|
||||||
|
EXPECT_LONGS_EQUAL(4, incrementalHybrid.size());
|
||||||
|
EXPECT_LONGS_EQUAL(
|
||||||
|
2, incrementalHybrid[X(1)]->conditional()->asMixture()->nrComponents());
|
||||||
|
EXPECT_LONGS_EQUAL(
|
||||||
|
4, incrementalHybrid[X(2)]->conditional()->asMixture()->nrComponents());
|
||||||
|
EXPECT_LONGS_EQUAL(
|
||||||
|
5, incrementalHybrid[X(3)]->conditional()->asMixture()->nrComponents());
|
||||||
|
EXPECT_LONGS_EQUAL(
|
||||||
|
5, incrementalHybrid[X(4)]->conditional()->asMixture()->nrComponents());
|
||||||
|
|
||||||
|
/***** Run Round 2 *****/
|
||||||
|
HybridGaussianFactorGraph graph2;
|
||||||
|
graph2.push_back(switching.linearizedFactorGraph.at(4));
|
||||||
|
graph2.push_back(switching.linearizedFactorGraph.at(8));
|
||||||
|
|
||||||
|
// Run update with pruning a second time.
|
||||||
|
incrementalHybrid.update(graph2);
|
||||||
|
incrementalHybrid.prune(M(4), maxComponents);
|
||||||
|
|
||||||
|
// Check if we have a bayes tree with pruned hybrid nodes,
|
||||||
|
// with 5 (pruned) leaves.
|
||||||
|
CHECK_EQUAL(5, incrementalHybrid.size());
|
||||||
|
EXPECT_LONGS_EQUAL(
|
||||||
|
5, incrementalHybrid[X(4)]->conditional()->asMixture()->nrComponents());
|
||||||
|
EXPECT_LONGS_EQUAL(
|
||||||
|
5, incrementalHybrid[X(5)]->conditional()->asMixture()->nrComponents());
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************/
|
||||||
|
// A GTSAM-only test for running inference on a single-legged robot.
|
||||||
|
// The leg links are represented by the chain X-Y-Z-W, where X is the base and
|
||||||
|
// W is the foot.
|
||||||
|
// We use BetweenFactor<Pose2> as constraints between each of the poses.
|
||||||
|
TEST(HybridGaussianISAM, NonTrivial) {
|
||||||
|
/*************** Run Round 1 ***************/
|
||||||
|
HybridNonlinearFactorGraph fg;
|
||||||
|
|
||||||
|
// Add a prior on pose x1 at the origin.
|
||||||
|
// A prior factor consists of a mean and
|
||||||
|
// a noise model (covariance matrix)
|
||||||
|
Pose2 prior(0.0, 0.0, 0.0); // prior mean is at origin
|
||||||
|
auto priorNoise = noiseModel::Diagonal::Sigmas(
|
||||||
|
Vector3(0.3, 0.3, 0.1)); // 30cm std on x,y, 0.1 rad on theta
|
||||||
|
fg.emplace_nonlinear<PriorFactor<Pose2>>(X(0), prior, priorNoise);
|
||||||
|
|
||||||
|
// create a noise model for the landmark measurements
|
||||||
|
auto poseNoise = noiseModel::Isotropic::Sigma(3, 0.1);
|
||||||
|
|
||||||
|
// We model a robot's single leg as X - Y - Z - W
|
||||||
|
// where X is the base link and W is the foot link.
|
||||||
|
|
||||||
|
// Add connecting poses similar to PoseFactors in GTD
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(X(0), Y(0), Pose2(0, 1.0, 0),
|
||||||
|
poseNoise);
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(Y(0), Z(0), Pose2(0, 1.0, 0),
|
||||||
|
poseNoise);
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(Z(0), W(0), Pose2(0, 1.0, 0),
|
||||||
|
poseNoise);
|
||||||
|
|
||||||
|
// Create initial estimate
|
||||||
|
Values initial;
|
||||||
|
initial.insert(X(0), Pose2(0.0, 0.0, 0.0));
|
||||||
|
initial.insert(Y(0), Pose2(0.0, 1.0, 0.0));
|
||||||
|
initial.insert(Z(0), Pose2(0.0, 2.0, 0.0));
|
||||||
|
initial.insert(W(0), Pose2(0.0, 3.0, 0.0));
|
||||||
|
|
||||||
|
HybridGaussianFactorGraph gfg = fg.linearize(initial);
|
||||||
|
fg = HybridNonlinearFactorGraph();
|
||||||
|
|
||||||
|
HybridGaussianISAM inc;
|
||||||
|
|
||||||
|
// Update without pruning
|
||||||
|
// The result is a HybridBayesNet with no discrete variables
|
||||||
|
// (equivalent to a GaussianBayesNet).
|
||||||
|
// Factorization is:
|
||||||
|
// `P(X | measurements) = P(W0|Z0) P(Z0|Y0) P(Y0|X0) P(X0)`
|
||||||
|
inc.update(gfg);
|
||||||
|
|
||||||
|
/*************** Run Round 2 ***************/
|
||||||
|
using PlanarMotionModel = BetweenFactor<Pose2>;
|
||||||
|
|
||||||
|
// Add odometry factor with discrete modes.
|
||||||
|
Pose2 odometry(1.0, 0.0, 0.0);
|
||||||
|
KeyVector contKeys = {W(0), W(1)};
|
||||||
|
auto noise_model = noiseModel::Isotropic::Sigma(3, 1.0);
|
||||||
|
auto still = boost::make_shared<PlanarMotionModel>(W(0), W(1), Pose2(0, 0, 0),
|
||||||
|
noise_model),
|
||||||
|
moving = boost::make_shared<PlanarMotionModel>(W(0), W(1), odometry,
|
||||||
|
noise_model);
|
||||||
|
std::vector<PlanarMotionModel::shared_ptr> components = {moving, still};
|
||||||
|
auto mixtureFactor = boost::make_shared<MixtureFactor>(
|
||||||
|
contKeys, DiscreteKeys{gtsam::DiscreteKey(M(1), 2)}, components);
|
||||||
|
fg.push_back(mixtureFactor);
|
||||||
|
|
||||||
|
// Add equivalent of ImuFactor
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(X(0), X(1), Pose2(1.0, 0.0, 0),
|
||||||
|
poseNoise);
|
||||||
|
// PoseFactors-like at k=1
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(X(1), Y(1), Pose2(0, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(Y(1), Z(1), Pose2(0, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(Z(1), W(1), Pose2(-1, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
|
||||||
|
initial.insert(X(1), Pose2(1.0, 0.0, 0.0));
|
||||||
|
initial.insert(Y(1), Pose2(1.0, 1.0, 0.0));
|
||||||
|
initial.insert(Z(1), Pose2(1.0, 2.0, 0.0));
|
||||||
|
// The leg link did not move so we set the expected pose accordingly.
|
||||||
|
initial.insert(W(1), Pose2(0.0, 3.0, 0.0));
|
||||||
|
|
||||||
|
gfg = fg.linearize(initial);
|
||||||
|
fg = HybridNonlinearFactorGraph();
|
||||||
|
|
||||||
|
// Update without pruning
|
||||||
|
// The result is a HybridBayesNet with 1 discrete variable M(1).
|
||||||
|
// P(X | measurements) = P(W0|Z0, W1, M1) P(Z0|Y0, W1, M1) P(Y0|X0, W1, M1)
|
||||||
|
// P(X0 | X1, W1, M1) P(W1|Z1, X1, M1) P(Z1|Y1, X1, M1)
|
||||||
|
// P(Y1 | X1, M1)P(X1 | M1)P(M1)
|
||||||
|
// The MHS tree is a 1 level tree for time indices (1,) with 2 leaves.
|
||||||
|
inc.update(gfg);
|
||||||
|
|
||||||
|
/*************** Run Round 3 ***************/
|
||||||
|
// Add odometry factor with discrete modes.
|
||||||
|
contKeys = {W(1), W(2)};
|
||||||
|
still = boost::make_shared<PlanarMotionModel>(W(1), W(2), Pose2(0, 0, 0),
|
||||||
|
noise_model);
|
||||||
|
moving =
|
||||||
|
boost::make_shared<PlanarMotionModel>(W(1), W(2), odometry, noise_model);
|
||||||
|
components = {moving, still};
|
||||||
|
mixtureFactor = boost::make_shared<MixtureFactor>(
|
||||||
|
contKeys, DiscreteKeys{gtsam::DiscreteKey(M(2), 2)}, components);
|
||||||
|
fg.push_back(mixtureFactor);
|
||||||
|
|
||||||
|
// Add equivalent of ImuFactor
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(X(1), X(2), Pose2(1.0, 0.0, 0),
|
||||||
|
poseNoise);
|
||||||
|
// PoseFactors-like at k=1
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(X(2), Y(2), Pose2(0, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(Y(2), Z(2), Pose2(0, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(Z(2), W(2), Pose2(-2, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
|
||||||
|
initial.insert(X(2), Pose2(2.0, 0.0, 0.0));
|
||||||
|
initial.insert(Y(2), Pose2(2.0, 1.0, 0.0));
|
||||||
|
initial.insert(Z(2), Pose2(2.0, 2.0, 0.0));
|
||||||
|
initial.insert(W(2), Pose2(0.0, 3.0, 0.0));
|
||||||
|
|
||||||
|
gfg = fg.linearize(initial);
|
||||||
|
fg = HybridNonlinearFactorGraph();
|
||||||
|
|
||||||
|
// Now we prune!
|
||||||
|
// P(X | measurements) = P(W0|Z0, W1, M1) P(Z0|Y0, W1, M1) P(Y0|X0, W1, M1)
|
||||||
|
// P(X0 | X1, W1, M1) P(W1|W2, Z1, X1, M1, M2)
|
||||||
|
// P(Z1| W2, Y1, X1, M1, M2) P(Y1 | W2, X1, M1, M2)
|
||||||
|
// P(X1 | W2, X2, M1, M2) P(W2|Z2, X2, M1, M2)
|
||||||
|
// P(Z2|Y2, X2, M1, M2) P(Y2 | X2, M1, M2)
|
||||||
|
// P(X2 | M1, M2) P(M1, M2)
|
||||||
|
// The MHS at this point should be a 2 level tree on (1, 2).
|
||||||
|
// 1 has 2 choices, and 2 has 4 choices.
|
||||||
|
inc.update(gfg);
|
||||||
|
inc.prune(M(2), 2);
|
||||||
|
|
||||||
|
/*************** Run Round 4 ***************/
|
||||||
|
// Add odometry factor with discrete modes.
|
||||||
|
contKeys = {W(2), W(3)};
|
||||||
|
still = boost::make_shared<PlanarMotionModel>(W(2), W(3), Pose2(0, 0, 0),
|
||||||
|
noise_model);
|
||||||
|
moving =
|
||||||
|
boost::make_shared<PlanarMotionModel>(W(2), W(3), odometry, noise_model);
|
||||||
|
components = {moving, still};
|
||||||
|
mixtureFactor = boost::make_shared<MixtureFactor>(
|
||||||
|
contKeys, DiscreteKeys{gtsam::DiscreteKey(M(3), 2)}, components);
|
||||||
|
fg.push_back(mixtureFactor);
|
||||||
|
|
||||||
|
// Add equivalent of ImuFactor
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(X(2), X(3), Pose2(1.0, 0.0, 0),
|
||||||
|
poseNoise);
|
||||||
|
// PoseFactors-like at k=3
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(X(3), Y(3), Pose2(0, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(Y(3), Z(3), Pose2(0, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
fg.emplace_nonlinear<BetweenFactor<Pose2>>(Z(3), W(3), Pose2(-3, 1, 0),
|
||||||
|
poseNoise);
|
||||||
|
|
||||||
|
initial.insert(X(3), Pose2(3.0, 0.0, 0.0));
|
||||||
|
initial.insert(Y(3), Pose2(3.0, 1.0, 0.0));
|
||||||
|
initial.insert(Z(3), Pose2(3.0, 2.0, 0.0));
|
||||||
|
initial.insert(W(3), Pose2(0.0, 3.0, 0.0));
|
||||||
|
|
||||||
|
gfg = fg.linearize(initial);
|
||||||
|
fg = HybridNonlinearFactorGraph();
|
||||||
|
|
||||||
|
// Keep pruning!
|
||||||
|
inc.update(gfg);
|
||||||
|
inc.prune(M(3), 3);
|
||||||
|
|
||||||
|
// The final discrete graph should not be empty since we have eliminated
|
||||||
|
// all continuous variables.
|
||||||
|
auto discreteTree = inc[M(3)]->conditional()->asDiscreteConditional();
|
||||||
|
EXPECT_LONGS_EQUAL(3, discreteTree->size());
|
||||||
|
|
||||||
|
// Test if the optimal discrete mode assignment is (1, 1, 1).
|
||||||
|
DiscreteFactorGraph discreteGraph;
|
||||||
|
discreteGraph.push_back(discreteTree);
|
||||||
|
DiscreteValues optimal_assignment = discreteGraph.optimize();
|
||||||
|
|
||||||
|
DiscreteValues expected_assignment;
|
||||||
|
expected_assignment[M(1)] = 1;
|
||||||
|
expected_assignment[M(2)] = 1;
|
||||||
|
expected_assignment[M(3)] = 1;
|
||||||
|
|
||||||
|
EXPECT(assert_equal(expected_assignment, optimal_assignment));
|
||||||
|
|
||||||
|
// Test if pruning worked correctly by checking that we only have 3 leaves in
|
||||||
|
// the last node.
|
||||||
|
auto lastConditional = inc[X(3)]->conditional()->asMixture();
|
||||||
|
EXPECT_LONGS_EQUAL(3, lastConditional->nrComponents());
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
int main() {
|
||||||
|
TestResult tr;
|
||||||
|
return TestRegistry::runAllTests(tr);
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,272 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file testHybridLookupDAG.cpp
|
||||||
|
* @date Aug, 2022
|
||||||
|
* @author Shangjie Xue
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <gtsam/base/Testable.h>
|
||||||
|
#include <gtsam/base/TestableAssertions.h>
|
||||||
|
#include <gtsam/discrete/Assignment.h>
|
||||||
|
#include <gtsam/discrete/DecisionTreeFactor.h>
|
||||||
|
#include <gtsam/discrete/DiscreteKey.h>
|
||||||
|
#include <gtsam/discrete/DiscreteValues.h>
|
||||||
|
#include <gtsam/hybrid/GaussianMixture.h>
|
||||||
|
#include <gtsam/hybrid/HybridBayesNet.h>
|
||||||
|
#include <gtsam/hybrid/HybridLookupDAG.h>
|
||||||
|
#include <gtsam/hybrid/HybridValues.h>
|
||||||
|
#include <gtsam/inference/Key.h>
|
||||||
|
#include <gtsam/inference/Symbol.h>
|
||||||
|
#include <gtsam/linear/GaussianConditional.h>
|
||||||
|
#include <gtsam/linear/VectorValues.h>
|
||||||
|
#include <gtsam/nonlinear/Values.h>
|
||||||
|
|
||||||
|
// Include for test suite
|
||||||
|
#include <CppUnitLite/TestHarness.h>
|
||||||
|
|
||||||
|
#include <iostream>
|
||||||
|
|
||||||
|
using namespace std;
|
||||||
|
using namespace gtsam;
|
||||||
|
using noiseModel::Isotropic;
|
||||||
|
using symbol_shorthand::M;
|
||||||
|
using symbol_shorthand::X;
|
||||||
|
|
||||||
|
TEST(HybridLookupTable, basics) {
|
||||||
|
// create a conditional gaussian node
|
||||||
|
Matrix S1(2, 2);
|
||||||
|
S1(0, 0) = 1;
|
||||||
|
S1(1, 0) = 2;
|
||||||
|
S1(0, 1) = 3;
|
||||||
|
S1(1, 1) = 4;
|
||||||
|
|
||||||
|
Matrix S2(2, 2);
|
||||||
|
S2(0, 0) = 6;
|
||||||
|
S2(1, 0) = 0.2;
|
||||||
|
S2(0, 1) = 8;
|
||||||
|
S2(1, 1) = 0.4;
|
||||||
|
|
||||||
|
Matrix R1(2, 2);
|
||||||
|
R1(0, 0) = 0.1;
|
||||||
|
R1(1, 0) = 0.3;
|
||||||
|
R1(0, 1) = 0.0;
|
||||||
|
R1(1, 1) = 0.34;
|
||||||
|
|
||||||
|
Matrix R2(2, 2);
|
||||||
|
R2(0, 0) = 0.1;
|
||||||
|
R2(1, 0) = 0.3;
|
||||||
|
R2(0, 1) = 0.0;
|
||||||
|
R2(1, 1) = 0.34;
|
||||||
|
|
||||||
|
SharedDiagonal model = noiseModel::Diagonal::Sigmas(Vector2(1.0, 0.34));
|
||||||
|
|
||||||
|
Vector2 d1(0.2, 0.5), d2(0.5, 0.2);
|
||||||
|
|
||||||
|
auto conditional0 = boost::make_shared<GaussianConditional>(X(1), d1, R1,
|
||||||
|
X(2), S1, model),
|
||||||
|
conditional1 = boost::make_shared<GaussianConditional>(X(1), d2, R2,
|
||||||
|
X(2), S2, model);
|
||||||
|
|
||||||
|
// Create decision tree
|
||||||
|
DiscreteKey m1(1, 2);
|
||||||
|
GaussianMixture::Conditionals conditionals(
|
||||||
|
{m1},
|
||||||
|
vector<GaussianConditional::shared_ptr>{conditional0, conditional1});
|
||||||
|
// GaussianMixture mixtureFactor2({X(1)}, {X(2)}, {m1}, conditionals);
|
||||||
|
|
||||||
|
boost::shared_ptr<GaussianMixture> mixtureFactor(
|
||||||
|
new GaussianMixture({X(1)}, {X(2)}, {m1}, conditionals));
|
||||||
|
|
||||||
|
HybridConditional hc(mixtureFactor);
|
||||||
|
|
||||||
|
GaussianMixture::Conditionals conditional2 =
|
||||||
|
boost::static_pointer_cast<GaussianMixture>(hc.inner())->conditionals();
|
||||||
|
|
||||||
|
DiscreteValues dv;
|
||||||
|
dv[1] = 1;
|
||||||
|
|
||||||
|
VectorValues cv;
|
||||||
|
cv.insert(X(2), Vector2(0.0, 0.0));
|
||||||
|
|
||||||
|
HybridValues hv(dv, cv);
|
||||||
|
|
||||||
|
// std::cout << conditional2(values).markdown();
|
||||||
|
EXPECT(assert_equal(*conditional2(dv), *conditionals(dv), 1e-6));
|
||||||
|
EXPECT(conditional2(dv) == conditionals(dv));
|
||||||
|
HybridLookupTable hlt(hc);
|
||||||
|
|
||||||
|
// hlt.argmaxInPlace(&hv);
|
||||||
|
|
||||||
|
HybridLookupDAG dag;
|
||||||
|
dag.push_back(hlt);
|
||||||
|
dag.argmax(hv);
|
||||||
|
|
||||||
|
// HybridBayesNet hbn;
|
||||||
|
// hbn.push_back(hc);
|
||||||
|
// hbn.optimize();
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST(HybridLookupTable, hybrid_argmax) {
|
||||||
|
Matrix S1(2, 2);
|
||||||
|
S1(0, 0) = 1;
|
||||||
|
S1(1, 0) = 0;
|
||||||
|
S1(0, 1) = 0;
|
||||||
|
S1(1, 1) = 1;
|
||||||
|
|
||||||
|
Vector2 d1(0.2, 0.5), d2(-0.5, 0.6);
|
||||||
|
|
||||||
|
SharedDiagonal model = noiseModel::Diagonal::Sigmas(Vector2(1.0, 0.34));
|
||||||
|
|
||||||
|
auto conditional0 =
|
||||||
|
boost::make_shared<GaussianConditional>(X(1), d1, S1, model),
|
||||||
|
conditional1 =
|
||||||
|
boost::make_shared<GaussianConditional>(X(1), d2, S1, model);
|
||||||
|
|
||||||
|
DiscreteKey m1(1, 2);
|
||||||
|
GaussianMixture::Conditionals conditionals(
|
||||||
|
{m1},
|
||||||
|
vector<GaussianConditional::shared_ptr>{conditional0, conditional1});
|
||||||
|
boost::shared_ptr<GaussianMixture> mixtureFactor(
|
||||||
|
new GaussianMixture({X(1)}, {}, {m1}, conditionals));
|
||||||
|
|
||||||
|
HybridConditional hc(mixtureFactor);
|
||||||
|
|
||||||
|
DiscreteValues dv;
|
||||||
|
dv[1] = 1;
|
||||||
|
VectorValues cv;
|
||||||
|
// cv.insert(X(2),Vector2(0.0, 0.0));
|
||||||
|
HybridValues hv(dv, cv);
|
||||||
|
|
||||||
|
HybridLookupTable hlt(hc);
|
||||||
|
|
||||||
|
hlt.argmaxInPlace(&hv);
|
||||||
|
|
||||||
|
EXPECT(assert_equal(hv.at(X(1)), d2));
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST(HybridLookupTable, discrete_argmax) {
|
||||||
|
DiscreteKey X(0, 2), Y(1, 2);
|
||||||
|
|
||||||
|
auto conditional = boost::make_shared<DiscreteConditional>(X | Y = "0/1 3/2");
|
||||||
|
|
||||||
|
HybridConditional hc(conditional);
|
||||||
|
|
||||||
|
HybridLookupTable hlt(hc);
|
||||||
|
|
||||||
|
DiscreteValues dv;
|
||||||
|
dv[1] = 0;
|
||||||
|
VectorValues cv;
|
||||||
|
// cv.insert(X(2),Vector2(0.0, 0.0));
|
||||||
|
HybridValues hv(dv, cv);
|
||||||
|
|
||||||
|
hlt.argmaxInPlace(&hv);
|
||||||
|
|
||||||
|
EXPECT(assert_equal(hv.atDiscrete(0), 1));
|
||||||
|
|
||||||
|
DecisionTreeFactor f1(X, "2 3");
|
||||||
|
auto conditional2 = boost::make_shared<DiscreteConditional>(1, f1);
|
||||||
|
|
||||||
|
HybridConditional hc2(conditional2);
|
||||||
|
|
||||||
|
HybridLookupTable hlt2(hc2);
|
||||||
|
|
||||||
|
HybridValues hv2;
|
||||||
|
|
||||||
|
hlt2.argmaxInPlace(&hv2);
|
||||||
|
|
||||||
|
EXPECT(assert_equal(hv2.atDiscrete(0), 1));
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST(HybridLookupTable, gaussian_argmax) {
|
||||||
|
Matrix S1(2, 2);
|
||||||
|
S1(0, 0) = 1;
|
||||||
|
S1(1, 0) = 0;
|
||||||
|
S1(0, 1) = 0;
|
||||||
|
S1(1, 1) = 1;
|
||||||
|
|
||||||
|
Vector2 d1(0.2, 0.5), d2(-0.5, 0.6);
|
||||||
|
|
||||||
|
SharedDiagonal model = noiseModel::Diagonal::Sigmas(Vector2(1.0, 0.34));
|
||||||
|
|
||||||
|
auto conditional =
|
||||||
|
boost::make_shared<GaussianConditional>(X(1), d1, S1, X(2), -S1, model);
|
||||||
|
|
||||||
|
HybridConditional hc(conditional);
|
||||||
|
|
||||||
|
HybridLookupTable hlt(hc);
|
||||||
|
|
||||||
|
DiscreteValues dv;
|
||||||
|
// dv[1]=0;
|
||||||
|
VectorValues cv;
|
||||||
|
cv.insert(X(2), d2);
|
||||||
|
HybridValues hv(dv, cv);
|
||||||
|
|
||||||
|
hlt.argmaxInPlace(&hv);
|
||||||
|
|
||||||
|
EXPECT(assert_equal(hv.at(X(1)), d1 + d2));
|
||||||
|
}
|
||||||
|
|
||||||
|
TEST(HybridLookupDAG, argmax) {
|
||||||
|
Matrix S1(2, 2);
|
||||||
|
S1(0, 0) = 1;
|
||||||
|
S1(1, 0) = 0;
|
||||||
|
S1(0, 1) = 0;
|
||||||
|
S1(1, 1) = 1;
|
||||||
|
|
||||||
|
Vector2 d1(0.2, 0.5), d2(-0.5, 0.6);
|
||||||
|
|
||||||
|
SharedDiagonal model = noiseModel::Diagonal::Sigmas(Vector2(1.0, 0.34));
|
||||||
|
|
||||||
|
auto conditional0 =
|
||||||
|
boost::make_shared<GaussianConditional>(X(2), d1, S1, model),
|
||||||
|
conditional1 =
|
||||||
|
boost::make_shared<GaussianConditional>(X(2), d2, S1, model);
|
||||||
|
|
||||||
|
DiscreteKey m1(1, 2);
|
||||||
|
GaussianMixture::Conditionals conditionals(
|
||||||
|
{m1},
|
||||||
|
vector<GaussianConditional::shared_ptr>{conditional0, conditional1});
|
||||||
|
boost::shared_ptr<GaussianMixture> mixtureFactor(
|
||||||
|
new GaussianMixture({X(2)}, {}, {m1}, conditionals));
|
||||||
|
HybridConditional hc2(mixtureFactor);
|
||||||
|
HybridLookupTable hlt2(hc2);
|
||||||
|
|
||||||
|
auto conditional2 =
|
||||||
|
boost::make_shared<GaussianConditional>(X(1), d1, S1, X(2), -S1, model);
|
||||||
|
|
||||||
|
HybridConditional hc1(conditional2);
|
||||||
|
HybridLookupTable hlt1(hc1);
|
||||||
|
|
||||||
|
DecisionTreeFactor f1(m1, "2 3");
|
||||||
|
auto discrete_conditional = boost::make_shared<DiscreteConditional>(1, f1);
|
||||||
|
|
||||||
|
HybridConditional hc3(discrete_conditional);
|
||||||
|
HybridLookupTable hlt3(hc3);
|
||||||
|
|
||||||
|
HybridLookupDAG dag;
|
||||||
|
dag.push_back(hlt1);
|
||||||
|
dag.push_back(hlt2);
|
||||||
|
dag.push_back(hlt3);
|
||||||
|
auto hv = dag.argmax();
|
||||||
|
|
||||||
|
EXPECT(assert_equal(hv.atDiscrete(1), 1));
|
||||||
|
EXPECT(assert_equal(hv.at(X(2)), d2));
|
||||||
|
EXPECT(assert_equal(hv.at(X(1)), d2 + d1));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
int main() {
|
||||||
|
TestResult tr;
|
||||||
|
return TestRegistry::runAllTests(tr);
|
||||||
|
}
|
||||||
|
/* ************************************************************************* */
|
||||||
|
|
@ -0,0 +1,776 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
* See LICENSE for the license information
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file testHybridNonlinearFactorGraph.cpp
|
||||||
|
* @brief Unit tests for HybridNonlinearFactorGraph
|
||||||
|
* @author Varun Agrawal
|
||||||
|
* @author Fan Jiang
|
||||||
|
* @author Frank Dellaert
|
||||||
|
* @date December 2021
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <gtsam/base/TestableAssertions.h>
|
||||||
|
#include <gtsam/base/utilities.h>
|
||||||
|
#include <gtsam/discrete/DiscreteBayesNet.h>
|
||||||
|
#include <gtsam/discrete/DiscreteDistribution.h>
|
||||||
|
#include <gtsam/discrete/DiscreteFactorGraph.h>
|
||||||
|
#include <gtsam/geometry/Pose2.h>
|
||||||
|
#include <gtsam/hybrid/HybridEliminationTree.h>
|
||||||
|
#include <gtsam/hybrid/HybridFactor.h>
|
||||||
|
#include <gtsam/hybrid/HybridNonlinearFactorGraph.h>
|
||||||
|
#include <gtsam/hybrid/MixtureFactor.h>
|
||||||
|
#include <gtsam/linear/GaussianBayesNet.h>
|
||||||
|
#include <gtsam/linear/GaussianFactorGraph.h>
|
||||||
|
#include <gtsam/nonlinear/NonlinearFactorGraph.h>
|
||||||
|
#include <gtsam/nonlinear/PriorFactor.h>
|
||||||
|
#include <gtsam/sam/BearingRangeFactor.h>
|
||||||
|
#include <gtsam/slam/BetweenFactor.h>
|
||||||
|
|
||||||
|
#include <numeric>
|
||||||
|
|
||||||
|
#include "Switching.h"
|
||||||
|
|
||||||
|
// Include for test suite
|
||||||
|
#include <CppUnitLite/TestHarness.h>
|
||||||
|
|
||||||
|
using namespace std;
|
||||||
|
using namespace gtsam;
|
||||||
|
using noiseModel::Isotropic;
|
||||||
|
using symbol_shorthand::L;
|
||||||
|
using symbol_shorthand::M;
|
||||||
|
using symbol_shorthand::X;
|
||||||
|
|
||||||
|
/* ****************************************************************************
|
||||||
|
* Test that any linearizedFactorGraph gaussian factors are appended to the
|
||||||
|
* existing gaussian factor graph in the hybrid factor graph.
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, GaussianFactorGraph) {
|
||||||
|
HybridNonlinearFactorGraph fg;
|
||||||
|
|
||||||
|
// Add a simple prior factor to the nonlinear factor graph
|
||||||
|
fg.emplace_nonlinear<PriorFactor<double>>(X(0), 0, Isotropic::Sigma(1, 0.1));
|
||||||
|
|
||||||
|
// Linearization point
|
||||||
|
Values linearizationPoint;
|
||||||
|
linearizationPoint.insert<double>(X(0), 0);
|
||||||
|
|
||||||
|
HybridGaussianFactorGraph ghfg = fg.linearize(linearizationPoint);
|
||||||
|
|
||||||
|
// Add a factor to the GaussianFactorGraph
|
||||||
|
ghfg.add(JacobianFactor(X(0), I_1x1, Vector1(5)));
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(2, ghfg.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
/***************************************************************************
|
||||||
|
* Test equality for Hybrid Nonlinear Factor Graph.
|
||||||
|
*/
|
||||||
|
TEST(HybridNonlinearFactorGraph, Equals) {
|
||||||
|
HybridNonlinearFactorGraph graph1;
|
||||||
|
HybridNonlinearFactorGraph graph2;
|
||||||
|
|
||||||
|
// Test empty factor graphs
|
||||||
|
EXPECT(assert_equal(graph1, graph2));
|
||||||
|
|
||||||
|
auto f0 = boost::make_shared<PriorFactor<Pose2>>(
|
||||||
|
1, Pose2(), noiseModel::Isotropic::Sigma(3, 0.001));
|
||||||
|
graph1.push_back(f0);
|
||||||
|
graph2.push_back(f0);
|
||||||
|
|
||||||
|
auto f1 = boost::make_shared<BetweenFactor<Pose2>>(
|
||||||
|
1, 2, Pose2(), noiseModel::Isotropic::Sigma(3, 0.1));
|
||||||
|
graph1.push_back(f1);
|
||||||
|
graph2.push_back(f1);
|
||||||
|
|
||||||
|
// Test non-empty graphs
|
||||||
|
EXPECT(assert_equal(graph1, graph2));
|
||||||
|
}
|
||||||
|
|
||||||
|
/***************************************************************************
|
||||||
|
* Test that the resize method works correctly for a HybridNonlinearFactorGraph.
|
||||||
|
*/
|
||||||
|
TEST(HybridNonlinearFactorGraph, Resize) {
|
||||||
|
HybridNonlinearFactorGraph fg;
|
||||||
|
auto nonlinearFactor = boost::make_shared<BetweenFactor<double>>();
|
||||||
|
fg.push_back(nonlinearFactor);
|
||||||
|
|
||||||
|
auto discreteFactor = boost::make_shared<DecisionTreeFactor>();
|
||||||
|
fg.push_back(discreteFactor);
|
||||||
|
|
||||||
|
auto dcFactor = boost::make_shared<MixtureFactor>();
|
||||||
|
fg.push_back(dcFactor);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(fg.size(), 3);
|
||||||
|
|
||||||
|
fg.resize(0);
|
||||||
|
EXPECT_LONGS_EQUAL(fg.size(), 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
/***************************************************************************
|
||||||
|
* Test that the resize method works correctly for a
|
||||||
|
* HybridGaussianFactorGraph.
|
||||||
|
*/
|
||||||
|
TEST(HybridGaussianFactorGraph, Resize) {
|
||||||
|
HybridNonlinearFactorGraph nhfg;
|
||||||
|
auto nonlinearFactor = boost::make_shared<BetweenFactor<double>>(
|
||||||
|
X(0), X(1), 0.0, Isotropic::Sigma(1, 0.1));
|
||||||
|
nhfg.push_back(nonlinearFactor);
|
||||||
|
auto discreteFactor = boost::make_shared<DecisionTreeFactor>();
|
||||||
|
nhfg.push_back(discreteFactor);
|
||||||
|
|
||||||
|
KeyVector contKeys = {X(0), X(1)};
|
||||||
|
auto noise_model = noiseModel::Isotropic::Sigma(1, 1.0);
|
||||||
|
auto still = boost::make_shared<MotionModel>(X(0), X(1), 0.0, noise_model),
|
||||||
|
moving = boost::make_shared<MotionModel>(X(0), X(1), 1.0, noise_model);
|
||||||
|
|
||||||
|
std::vector<MotionModel::shared_ptr> components = {still, moving};
|
||||||
|
auto dcFactor = boost::make_shared<MixtureFactor>(
|
||||||
|
contKeys, DiscreteKeys{gtsam::DiscreteKey(M(1), 2)}, components);
|
||||||
|
nhfg.push_back(dcFactor);
|
||||||
|
|
||||||
|
Values linearizationPoint;
|
||||||
|
linearizationPoint.insert<double>(X(0), 0);
|
||||||
|
linearizationPoint.insert<double>(X(1), 1);
|
||||||
|
|
||||||
|
// Generate `HybridGaussianFactorGraph` by linearizing
|
||||||
|
HybridGaussianFactorGraph gfg = nhfg.linearize(linearizationPoint);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(gfg.size(), 3);
|
||||||
|
|
||||||
|
gfg.resize(0);
|
||||||
|
EXPECT_LONGS_EQUAL(gfg.size(), 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*****************************************************************************
|
||||||
|
* Test push_back on HFG makes the correct distinction.
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, PushBack) {
|
||||||
|
HybridNonlinearFactorGraph fg;
|
||||||
|
|
||||||
|
auto nonlinearFactor = boost::make_shared<BetweenFactor<double>>();
|
||||||
|
fg.push_back(nonlinearFactor);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(fg.size(), 1);
|
||||||
|
|
||||||
|
fg = HybridNonlinearFactorGraph();
|
||||||
|
|
||||||
|
auto discreteFactor = boost::make_shared<DecisionTreeFactor>();
|
||||||
|
fg.push_back(discreteFactor);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(fg.size(), 1);
|
||||||
|
|
||||||
|
fg = HybridNonlinearFactorGraph();
|
||||||
|
|
||||||
|
auto dcFactor = boost::make_shared<MixtureFactor>();
|
||||||
|
fg.push_back(dcFactor);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(fg.size(), 1);
|
||||||
|
|
||||||
|
// Now do the same with HybridGaussianFactorGraph
|
||||||
|
HybridGaussianFactorGraph ghfg;
|
||||||
|
|
||||||
|
auto gaussianFactor = boost::make_shared<JacobianFactor>();
|
||||||
|
ghfg.push_back(gaussianFactor);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(ghfg.size(), 1);
|
||||||
|
|
||||||
|
ghfg = HybridGaussianFactorGraph();
|
||||||
|
ghfg.push_back(discreteFactor);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(ghfg.size(), 1);
|
||||||
|
|
||||||
|
ghfg = HybridGaussianFactorGraph();
|
||||||
|
ghfg.push_back(dcFactor);
|
||||||
|
|
||||||
|
HybridGaussianFactorGraph hgfg2;
|
||||||
|
hgfg2.push_back(ghfg.begin(), ghfg.end());
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(ghfg.size(), 1);
|
||||||
|
|
||||||
|
HybridNonlinearFactorGraph hnfg;
|
||||||
|
NonlinearFactorGraph factors;
|
||||||
|
auto noise = noiseModel::Isotropic::Sigma(3, 1.0);
|
||||||
|
factors.emplace_shared<PriorFactor<Pose2>>(0, Pose2(0, 0, 0), noise);
|
||||||
|
factors.emplace_shared<PriorFactor<Pose2>>(1, Pose2(1, 0, 0), noise);
|
||||||
|
factors.emplace_shared<PriorFactor<Pose2>>(2, Pose2(2, 0, 0), noise);
|
||||||
|
// TODO(Varun) This does not currently work. It should work once HybridFactor
|
||||||
|
// becomes a base class of NonlinearFactor.
|
||||||
|
// hnfg.push_back(factors.begin(), factors.end());
|
||||||
|
|
||||||
|
// EXPECT_LONGS_EQUAL(3, hnfg.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Test construction of switching-like hybrid factor graph.
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, Switching) {
|
||||||
|
Switching self(3);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(7, self.nonlinearFactorGraph.size());
|
||||||
|
EXPECT_LONGS_EQUAL(7, self.linearizedFactorGraph.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Test linearization on a switching-like hybrid factor graph.
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, Linearization) {
|
||||||
|
Switching self(3);
|
||||||
|
|
||||||
|
// Linearize here:
|
||||||
|
HybridGaussianFactorGraph actualLinearized =
|
||||||
|
self.nonlinearFactorGraph.linearize(self.linearizationPoint);
|
||||||
|
|
||||||
|
EXPECT_LONGS_EQUAL(7, actualLinearized.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Test elimination tree construction
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, EliminationTree) {
|
||||||
|
Switching self(3);
|
||||||
|
|
||||||
|
// Create ordering.
|
||||||
|
Ordering ordering;
|
||||||
|
for (size_t k = 1; k <= self.K; k++) ordering += X(k);
|
||||||
|
|
||||||
|
// Create elimination tree.
|
||||||
|
HybridEliminationTree etree(self.linearizedFactorGraph, ordering);
|
||||||
|
EXPECT_LONGS_EQUAL(1, etree.roots().size())
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
*Test elimination function by eliminating x1 in *-x1-*-x2 graph.
|
||||||
|
*/
|
||||||
|
TEST(GaussianElimination, Eliminate_x1) {
|
||||||
|
Switching self(3);
|
||||||
|
|
||||||
|
// Gather factors on x1, has a simple Gaussian and a mixture factor.
|
||||||
|
HybridGaussianFactorGraph factors;
|
||||||
|
// Add gaussian prior
|
||||||
|
factors.push_back(self.linearizedFactorGraph[0]);
|
||||||
|
// Add first hybrid factor
|
||||||
|
factors.push_back(self.linearizedFactorGraph[1]);
|
||||||
|
|
||||||
|
// TODO(Varun) remove this block since sum is no longer exposed.
|
||||||
|
// // Check that sum works:
|
||||||
|
// auto sum = factors.sum();
|
||||||
|
// Assignment<Key> mode;
|
||||||
|
// mode[M(1)] = 1;
|
||||||
|
// auto actual = sum(mode); // Selects one of 2 modes.
|
||||||
|
// EXPECT_LONGS_EQUAL(2, actual.size()); // Prior and motion model.
|
||||||
|
|
||||||
|
// Eliminate x1
|
||||||
|
Ordering ordering;
|
||||||
|
ordering += X(1);
|
||||||
|
|
||||||
|
auto result = EliminateHybrid(factors, ordering);
|
||||||
|
CHECK(result.first);
|
||||||
|
EXPECT_LONGS_EQUAL(1, result.first->nrFrontals());
|
||||||
|
CHECK(result.second);
|
||||||
|
// Has two keys, x2 and m1
|
||||||
|
EXPECT_LONGS_EQUAL(2, result.second->size());
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Test elimination function by eliminating x2 in x1-*-x2-*-x3 chain.
|
||||||
|
* m1/ \m2
|
||||||
|
*/
|
||||||
|
TEST(HybridsGaussianElimination, Eliminate_x2) {
|
||||||
|
Switching self(3);
|
||||||
|
|
||||||
|
// Gather factors on x2, will be two mixture factors (with x1 and x3, resp.).
|
||||||
|
HybridGaussianFactorGraph factors;
|
||||||
|
factors.push_back(self.linearizedFactorGraph[1]); // involves m1
|
||||||
|
factors.push_back(self.linearizedFactorGraph[2]); // involves m2
|
||||||
|
|
||||||
|
// TODO(Varun) remove this block since sum is no longer exposed.
|
||||||
|
// // Check that sum works:
|
||||||
|
// auto sum = factors.sum();
|
||||||
|
// Assignment<Key> mode;
|
||||||
|
// mode[M(1)] = 0;
|
||||||
|
// mode[M(2)] = 1;
|
||||||
|
// auto actual = sum(mode); // Selects one of 4 mode
|
||||||
|
// combinations. EXPECT_LONGS_EQUAL(2, actual.size()); // 2 motion models.
|
||||||
|
|
||||||
|
// Eliminate x2
|
||||||
|
Ordering ordering;
|
||||||
|
ordering += X(2);
|
||||||
|
|
||||||
|
std::pair<HybridConditional::shared_ptr, HybridFactor::shared_ptr> result =
|
||||||
|
EliminateHybrid(factors, ordering);
|
||||||
|
CHECK(result.first);
|
||||||
|
EXPECT_LONGS_EQUAL(1, result.first->nrFrontals());
|
||||||
|
CHECK(result.second);
|
||||||
|
// Note: separator keys should include m1, m2.
|
||||||
|
EXPECT_LONGS_EQUAL(4, result.second->size());
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Helper method to generate gaussian factor graphs with a specific mode.
|
||||||
|
*/
|
||||||
|
GaussianFactorGraph::shared_ptr batchGFG(double between,
|
||||||
|
Values linearizationPoint) {
|
||||||
|
NonlinearFactorGraph graph;
|
||||||
|
graph.addPrior<double>(X(1), 0, Isotropic::Sigma(1, 0.1));
|
||||||
|
|
||||||
|
auto between_x1_x2 = boost::make_shared<MotionModel>(
|
||||||
|
X(1), X(2), between, Isotropic::Sigma(1, 1.0));
|
||||||
|
|
||||||
|
graph.push_back(between_x1_x2);
|
||||||
|
|
||||||
|
return graph.linearize(linearizationPoint);
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Test elimination function by eliminating x1 and x2 in graph.
|
||||||
|
*/
|
||||||
|
TEST(HybridGaussianElimination, EliminateHybrid_2_Variable) {
|
||||||
|
Switching self(2, 1.0, 0.1);
|
||||||
|
|
||||||
|
auto factors = self.linearizedFactorGraph;
|
||||||
|
|
||||||
|
// Eliminate x1
|
||||||
|
Ordering ordering;
|
||||||
|
ordering += X(1);
|
||||||
|
ordering += X(2);
|
||||||
|
|
||||||
|
HybridConditional::shared_ptr hybridConditionalMixture;
|
||||||
|
HybridFactor::shared_ptr factorOnModes;
|
||||||
|
|
||||||
|
std::tie(hybridConditionalMixture, factorOnModes) =
|
||||||
|
EliminateHybrid(factors, ordering);
|
||||||
|
|
||||||
|
auto gaussianConditionalMixture =
|
||||||
|
dynamic_pointer_cast<GaussianMixture>(hybridConditionalMixture->inner());
|
||||||
|
|
||||||
|
CHECK(gaussianConditionalMixture);
|
||||||
|
// Frontals = [x1, x2]
|
||||||
|
EXPECT_LONGS_EQUAL(2, gaussianConditionalMixture->nrFrontals());
|
||||||
|
// 1 parent, which is the mode
|
||||||
|
EXPECT_LONGS_EQUAL(1, gaussianConditionalMixture->nrParents());
|
||||||
|
|
||||||
|
// This is now a HybridDiscreteFactor
|
||||||
|
auto hybridDiscreteFactor =
|
||||||
|
dynamic_pointer_cast<HybridDiscreteFactor>(factorOnModes);
|
||||||
|
// Access the type-erased inner object and convert to DecisionTreeFactor
|
||||||
|
auto discreteFactor =
|
||||||
|
dynamic_pointer_cast<DecisionTreeFactor>(hybridDiscreteFactor->inner());
|
||||||
|
CHECK(discreteFactor);
|
||||||
|
EXPECT_LONGS_EQUAL(1, discreteFactor->discreteKeys().size());
|
||||||
|
EXPECT(discreteFactor->root_->isLeaf() == false);
|
||||||
|
}
|
||||||
|
|
||||||
|
// /*
|
||||||
|
// ****************************************************************************/
|
||||||
|
// /// Test the toDecisionTreeFactor method
|
||||||
|
// TEST(HybridFactorGraph, ToDecisionTreeFactor) {
|
||||||
|
// size_t K = 3;
|
||||||
|
|
||||||
|
// // Provide tight sigma values so that the errors are visibly different.
|
||||||
|
// double between_sigma = 5e-8, prior_sigma = 1e-7;
|
||||||
|
|
||||||
|
// Switching self(K, between_sigma, prior_sigma);
|
||||||
|
|
||||||
|
// // Clear out discrete factors since sum() cannot hanldle those
|
||||||
|
// HybridGaussianFactorGraph linearizedFactorGraph(
|
||||||
|
// self.linearizedFactorGraph.gaussianGraph(), DiscreteFactorGraph(),
|
||||||
|
// self.linearizedFactorGraph.dcGraph());
|
||||||
|
|
||||||
|
// auto decisionTreeFactor = linearizedFactorGraph.toDecisionTreeFactor();
|
||||||
|
|
||||||
|
// auto allAssignments =
|
||||||
|
// DiscreteValues::CartesianProduct(linearizedFactorGraph.discreteKeys());
|
||||||
|
|
||||||
|
// // Get the error of the discrete assignment m1=0, m2=1.
|
||||||
|
// double actual = (*decisionTreeFactor)(allAssignments[1]);
|
||||||
|
|
||||||
|
// /********************************************/
|
||||||
|
// // Create equivalent factor graph for m1=0, m2=1
|
||||||
|
// GaussianFactorGraph graph = linearizedFactorGraph.gaussianGraph();
|
||||||
|
|
||||||
|
// for (auto &p : linearizedFactorGraph.dcGraph()) {
|
||||||
|
// if (auto mixture =
|
||||||
|
// boost::dynamic_pointer_cast<DCGaussianMixtureFactor>(p)) {
|
||||||
|
// graph.add((*mixture)(allAssignments[1]));
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
|
||||||
|
// VectorValues values = graph.optimize();
|
||||||
|
// double expected = graph.probPrime(values);
|
||||||
|
// /********************************************/
|
||||||
|
// EXPECT_DOUBLES_EQUAL(expected, actual, 1e-12);
|
||||||
|
// // REGRESSION:
|
||||||
|
// EXPECT_DOUBLES_EQUAL(0.6125, actual, 1e-4);
|
||||||
|
// }
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Test partial elimination
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, Partial_Elimination) {
|
||||||
|
Switching self(3);
|
||||||
|
|
||||||
|
auto linearizedFactorGraph = self.linearizedFactorGraph;
|
||||||
|
|
||||||
|
// Create ordering.
|
||||||
|
Ordering ordering;
|
||||||
|
for (size_t k = 1; k <= self.K; k++) ordering += X(k);
|
||||||
|
|
||||||
|
// Eliminate partially.
|
||||||
|
HybridBayesNet::shared_ptr hybridBayesNet;
|
||||||
|
HybridGaussianFactorGraph::shared_ptr remainingFactorGraph;
|
||||||
|
std::tie(hybridBayesNet, remainingFactorGraph) =
|
||||||
|
linearizedFactorGraph.eliminatePartialSequential(ordering);
|
||||||
|
|
||||||
|
CHECK(hybridBayesNet);
|
||||||
|
// GTSAM_PRINT(*hybridBayesNet); // HybridBayesNet
|
||||||
|
EXPECT_LONGS_EQUAL(3, hybridBayesNet->size());
|
||||||
|
EXPECT(hybridBayesNet->at(0)->frontals() == KeyVector{X(1)});
|
||||||
|
EXPECT(hybridBayesNet->at(0)->parents() == KeyVector({X(2), M(1)}));
|
||||||
|
EXPECT(hybridBayesNet->at(1)->frontals() == KeyVector{X(2)});
|
||||||
|
EXPECT(hybridBayesNet->at(1)->parents() == KeyVector({X(3), M(1), M(2)}));
|
||||||
|
EXPECT(hybridBayesNet->at(2)->frontals() == KeyVector{X(3)});
|
||||||
|
EXPECT(hybridBayesNet->at(2)->parents() == KeyVector({M(1), M(2)}));
|
||||||
|
|
||||||
|
CHECK(remainingFactorGraph);
|
||||||
|
// GTSAM_PRINT(*remainingFactorGraph); // HybridFactorGraph
|
||||||
|
EXPECT_LONGS_EQUAL(3, remainingFactorGraph->size());
|
||||||
|
EXPECT(remainingFactorGraph->at(0)->keys() == KeyVector({M(1)}));
|
||||||
|
EXPECT(remainingFactorGraph->at(1)->keys() == KeyVector({M(2), M(1)}));
|
||||||
|
EXPECT(remainingFactorGraph->at(2)->keys() == KeyVector({M(1), M(2)}));
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Test full elimination
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, Full_Elimination) {
|
||||||
|
Switching self(3);
|
||||||
|
|
||||||
|
auto linearizedFactorGraph = self.linearizedFactorGraph;
|
||||||
|
|
||||||
|
// We first do a partial elimination
|
||||||
|
HybridBayesNet::shared_ptr hybridBayesNet_partial;
|
||||||
|
HybridGaussianFactorGraph::shared_ptr remainingFactorGraph_partial;
|
||||||
|
DiscreteBayesNet discreteBayesNet;
|
||||||
|
|
||||||
|
{
|
||||||
|
// Create ordering.
|
||||||
|
Ordering ordering;
|
||||||
|
for (size_t k = 1; k <= self.K; k++) ordering += X(k);
|
||||||
|
|
||||||
|
// Eliminate partially.
|
||||||
|
std::tie(hybridBayesNet_partial, remainingFactorGraph_partial) =
|
||||||
|
linearizedFactorGraph.eliminatePartialSequential(ordering);
|
||||||
|
|
||||||
|
DiscreteFactorGraph discrete_fg;
|
||||||
|
// TODO(Varun) Make this a function of HybridGaussianFactorGraph?
|
||||||
|
for (HybridFactor::shared_ptr& factor : (*remainingFactorGraph_partial)) {
|
||||||
|
auto df = dynamic_pointer_cast<HybridDiscreteFactor>(factor);
|
||||||
|
discrete_fg.push_back(df->inner());
|
||||||
|
}
|
||||||
|
ordering.clear();
|
||||||
|
for (size_t k = 1; k < self.K; k++) ordering += M(k);
|
||||||
|
discreteBayesNet =
|
||||||
|
*discrete_fg.eliminateSequential(ordering, EliminateForMPE);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create ordering.
|
||||||
|
Ordering ordering;
|
||||||
|
for (size_t k = 1; k <= self.K; k++) ordering += X(k);
|
||||||
|
for (size_t k = 1; k < self.K; k++) ordering += M(k);
|
||||||
|
|
||||||
|
// Eliminate partially.
|
||||||
|
HybridBayesNet::shared_ptr hybridBayesNet =
|
||||||
|
linearizedFactorGraph.eliminateSequential(ordering);
|
||||||
|
|
||||||
|
CHECK(hybridBayesNet);
|
||||||
|
EXPECT_LONGS_EQUAL(5, hybridBayesNet->size());
|
||||||
|
// p(x1 | x2, m1)
|
||||||
|
EXPECT(hybridBayesNet->at(0)->frontals() == KeyVector{X(1)});
|
||||||
|
EXPECT(hybridBayesNet->at(0)->parents() == KeyVector({X(2), M(1)}));
|
||||||
|
// p(x2 | x3, m1, m2)
|
||||||
|
EXPECT(hybridBayesNet->at(1)->frontals() == KeyVector{X(2)});
|
||||||
|
EXPECT(hybridBayesNet->at(1)->parents() == KeyVector({X(3), M(1), M(2)}));
|
||||||
|
// p(x3 | m1, m2)
|
||||||
|
EXPECT(hybridBayesNet->at(2)->frontals() == KeyVector{X(3)});
|
||||||
|
EXPECT(hybridBayesNet->at(2)->parents() == KeyVector({M(1), M(2)}));
|
||||||
|
// P(m1 | m2)
|
||||||
|
EXPECT(hybridBayesNet->at(3)->frontals() == KeyVector{M(1)});
|
||||||
|
EXPECT(hybridBayesNet->at(3)->parents() == KeyVector({M(2)}));
|
||||||
|
EXPECT(
|
||||||
|
dynamic_pointer_cast<DiscreteConditional>(hybridBayesNet->at(3)->inner())
|
||||||
|
->equals(*discreteBayesNet.at(0)));
|
||||||
|
// P(m2)
|
||||||
|
EXPECT(hybridBayesNet->at(4)->frontals() == KeyVector{M(2)});
|
||||||
|
EXPECT_LONGS_EQUAL(0, hybridBayesNet->at(4)->nrParents());
|
||||||
|
EXPECT(
|
||||||
|
dynamic_pointer_cast<DiscreteConditional>(hybridBayesNet->at(4)->inner())
|
||||||
|
->equals(*discreteBayesNet.at(1)));
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Test printing
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, Printing) {
|
||||||
|
Switching self(3);
|
||||||
|
|
||||||
|
auto linearizedFactorGraph = self.linearizedFactorGraph;
|
||||||
|
|
||||||
|
// Create ordering.
|
||||||
|
Ordering ordering;
|
||||||
|
for (size_t k = 1; k <= self.K; k++) ordering += X(k);
|
||||||
|
|
||||||
|
// Eliminate partially.
|
||||||
|
HybridBayesNet::shared_ptr hybridBayesNet;
|
||||||
|
HybridGaussianFactorGraph::shared_ptr remainingFactorGraph;
|
||||||
|
std::tie(hybridBayesNet, remainingFactorGraph) =
|
||||||
|
linearizedFactorGraph.eliminatePartialSequential(ordering);
|
||||||
|
|
||||||
|
string expected_hybridFactorGraph = R"(
|
||||||
|
size: 7
|
||||||
|
factor 0: Continuous [x1]
|
||||||
|
|
||||||
|
A[x1] = [
|
||||||
|
10
|
||||||
|
]
|
||||||
|
b = [ -10 ]
|
||||||
|
No noise model
|
||||||
|
factor 1: Hybrid [x1 x2; m1]{
|
||||||
|
Choice(m1)
|
||||||
|
0 Leaf :
|
||||||
|
A[x1] = [
|
||||||
|
-1
|
||||||
|
]
|
||||||
|
A[x2] = [
|
||||||
|
1
|
||||||
|
]
|
||||||
|
b = [ -1 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
1 Leaf :
|
||||||
|
A[x1] = [
|
||||||
|
-1
|
||||||
|
]
|
||||||
|
A[x2] = [
|
||||||
|
1
|
||||||
|
]
|
||||||
|
b = [ -0 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
}
|
||||||
|
factor 2: Hybrid [x2 x3; m2]{
|
||||||
|
Choice(m2)
|
||||||
|
0 Leaf :
|
||||||
|
A[x2] = [
|
||||||
|
-1
|
||||||
|
]
|
||||||
|
A[x3] = [
|
||||||
|
1
|
||||||
|
]
|
||||||
|
b = [ -1 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
1 Leaf :
|
||||||
|
A[x2] = [
|
||||||
|
-1
|
||||||
|
]
|
||||||
|
A[x3] = [
|
||||||
|
1
|
||||||
|
]
|
||||||
|
b = [ -0 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
}
|
||||||
|
factor 3: Continuous [x2]
|
||||||
|
|
||||||
|
A[x2] = [
|
||||||
|
10
|
||||||
|
]
|
||||||
|
b = [ -10 ]
|
||||||
|
No noise model
|
||||||
|
factor 4: Continuous [x3]
|
||||||
|
|
||||||
|
A[x3] = [
|
||||||
|
10
|
||||||
|
]
|
||||||
|
b = [ -10 ]
|
||||||
|
No noise model
|
||||||
|
factor 5: Discrete [m1]
|
||||||
|
P( m1 ):
|
||||||
|
Leaf 0.5
|
||||||
|
|
||||||
|
factor 6: Discrete [m2 m1]
|
||||||
|
P( m2 | m1 ):
|
||||||
|
Choice(m2)
|
||||||
|
0 Choice(m1)
|
||||||
|
0 0 Leaf 0.33333333
|
||||||
|
0 1 Leaf 0.6
|
||||||
|
1 Choice(m1)
|
||||||
|
1 0 Leaf 0.66666667
|
||||||
|
1 1 Leaf 0.4
|
||||||
|
|
||||||
|
)";
|
||||||
|
EXPECT(assert_print_equal(expected_hybridFactorGraph, linearizedFactorGraph));
|
||||||
|
|
||||||
|
// Expected output for hybridBayesNet.
|
||||||
|
string expected_hybridBayesNet = R"(
|
||||||
|
size: 3
|
||||||
|
factor 0: Hybrid P( x1 | x2 m1)
|
||||||
|
Discrete Keys = (m1, 2),
|
||||||
|
Choice(m1)
|
||||||
|
0 Leaf p(x1 | x2)
|
||||||
|
R = [ 10.0499 ]
|
||||||
|
S[x2] = [ -0.0995037 ]
|
||||||
|
d = [ -9.85087 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
1 Leaf p(x1 | x2)
|
||||||
|
R = [ 10.0499 ]
|
||||||
|
S[x2] = [ -0.0995037 ]
|
||||||
|
d = [ -9.95037 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
factor 1: Hybrid P( x2 | x3 m1 m2)
|
||||||
|
Discrete Keys = (m1, 2), (m2, 2),
|
||||||
|
Choice(m2)
|
||||||
|
0 Choice(m1)
|
||||||
|
0 0 Leaf p(x2 | x3)
|
||||||
|
R = [ 10.099 ]
|
||||||
|
S[x3] = [ -0.0990196 ]
|
||||||
|
d = [ -9.99901 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
0 1 Leaf p(x2 | x3)
|
||||||
|
R = [ 10.099 ]
|
||||||
|
S[x3] = [ -0.0990196 ]
|
||||||
|
d = [ -9.90098 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
1 Choice(m1)
|
||||||
|
1 0 Leaf p(x2 | x3)
|
||||||
|
R = [ 10.099 ]
|
||||||
|
S[x3] = [ -0.0990196 ]
|
||||||
|
d = [ -10.098 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
1 1 Leaf p(x2 | x3)
|
||||||
|
R = [ 10.099 ]
|
||||||
|
S[x3] = [ -0.0990196 ]
|
||||||
|
d = [ -10 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
factor 2: Hybrid P( x3 | m1 m2)
|
||||||
|
Discrete Keys = (m1, 2), (m2, 2),
|
||||||
|
Choice(m2)
|
||||||
|
0 Choice(m1)
|
||||||
|
0 0 Leaf p(x3)
|
||||||
|
R = [ 10.0494 ]
|
||||||
|
d = [ -10.1489 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
0 1 Leaf p(x3)
|
||||||
|
R = [ 10.0494 ]
|
||||||
|
d = [ -10.1479 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
1 Choice(m1)
|
||||||
|
1 0 Leaf p(x3)
|
||||||
|
R = [ 10.0494 ]
|
||||||
|
d = [ -10.0504 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
1 1 Leaf p(x3)
|
||||||
|
R = [ 10.0494 ]
|
||||||
|
d = [ -10.0494 ]
|
||||||
|
No noise model
|
||||||
|
|
||||||
|
)";
|
||||||
|
EXPECT(assert_print_equal(expected_hybridBayesNet, *hybridBayesNet));
|
||||||
|
}
|
||||||
|
|
||||||
|
/****************************************************************************
|
||||||
|
* Simple PlanarSLAM example test with 2 poses and 2 landmarks (each pose
|
||||||
|
* connects to 1 landmark) to expose issue with default decision tree creation
|
||||||
|
* in hybrid elimination. The hybrid factor is between the poses X0 and X1.
|
||||||
|
* The issue arises if we eliminate a landmark variable first since it is not
|
||||||
|
* connected to a HybridFactor.
|
||||||
|
*/
|
||||||
|
TEST(HybridFactorGraph, DefaultDecisionTree) {
|
||||||
|
HybridNonlinearFactorGraph fg;
|
||||||
|
|
||||||
|
// Add a prior on pose x1 at the origin.
|
||||||
|
// A prior factor consists of a mean and a noise model (covariance matrix)
|
||||||
|
Pose2 prior(0.0, 0.0, 0.0); // prior mean is at origin
|
||||||
|
auto priorNoise = noiseModel::Diagonal::Sigmas(
|
||||||
|
Vector3(0.3, 0.3, 0.1)); // 30cm std on x,y, 0.1 rad on theta
|
||||||
|
fg.emplace_nonlinear<PriorFactor<Pose2>>(X(0), prior, priorNoise);
|
||||||
|
|
||||||
|
using PlanarMotionModel = BetweenFactor<Pose2>;
|
||||||
|
|
||||||
|
// Add odometry factor
|
||||||
|
Pose2 odometry(2.0, 0.0, 0.0);
|
||||||
|
KeyVector contKeys = {X(0), X(1)};
|
||||||
|
auto noise_model = noiseModel::Isotropic::Sigma(3, 1.0);
|
||||||
|
auto still = boost::make_shared<PlanarMotionModel>(X(0), X(1), Pose2(0, 0, 0),
|
||||||
|
noise_model),
|
||||||
|
moving = boost::make_shared<PlanarMotionModel>(X(0), X(1), odometry,
|
||||||
|
noise_model);
|
||||||
|
std::vector<PlanarMotionModel::shared_ptr> motion_models = {still, moving};
|
||||||
|
// TODO(Varun) Make a templated constructor for MixtureFactor which does this?
|
||||||
|
std::vector<NonlinearFactor::shared_ptr> components;
|
||||||
|
for (auto&& f : motion_models) {
|
||||||
|
components.push_back(boost::dynamic_pointer_cast<NonlinearFactor>(f));
|
||||||
|
}
|
||||||
|
fg.emplace_hybrid<MixtureFactor>(
|
||||||
|
contKeys, DiscreteKeys{gtsam::DiscreteKey(M(1), 2)}, components);
|
||||||
|
|
||||||
|
// Add Range-Bearing measurements to from X0 to L0 and X1 to L1.
|
||||||
|
// create a noise model for the landmark measurements
|
||||||
|
auto measurementNoise = noiseModel::Diagonal::Sigmas(
|
||||||
|
Vector2(0.1, 0.2)); // 0.1 rad std on bearing, 20cm on range
|
||||||
|
// create the measurement values - indices are (pose id, landmark id)
|
||||||
|
Rot2 bearing11 = Rot2::fromDegrees(45), bearing22 = Rot2::fromDegrees(90);
|
||||||
|
double range11 = std::sqrt(4.0 + 4.0), range22 = 2.0;
|
||||||
|
|
||||||
|
// Add Bearing-Range factors
|
||||||
|
fg.emplace_nonlinear<BearingRangeFactor<Pose2, Point2>>(
|
||||||
|
X(0), L(0), bearing11, range11, measurementNoise);
|
||||||
|
fg.emplace_nonlinear<BearingRangeFactor<Pose2, Point2>>(
|
||||||
|
X(1), L(1), bearing22, range22, measurementNoise);
|
||||||
|
|
||||||
|
// Create (deliberately inaccurate) initial estimate
|
||||||
|
Values initialEstimate;
|
||||||
|
initialEstimate.insert(X(0), Pose2(0.5, 0.0, 0.2));
|
||||||
|
initialEstimate.insert(X(1), Pose2(2.3, 0.1, -0.2));
|
||||||
|
initialEstimate.insert(L(0), Point2(1.8, 2.1));
|
||||||
|
initialEstimate.insert(L(1), Point2(4.1, 1.8));
|
||||||
|
|
||||||
|
// We want to eliminate variables not connected to DCFactors first.
|
||||||
|
Ordering ordering;
|
||||||
|
ordering += L(0);
|
||||||
|
ordering += L(1);
|
||||||
|
ordering += X(0);
|
||||||
|
ordering += X(1);
|
||||||
|
|
||||||
|
HybridGaussianFactorGraph linearized = fg.linearize(initialEstimate);
|
||||||
|
gtsam::HybridBayesNet::shared_ptr hybridBayesNet;
|
||||||
|
gtsam::HybridGaussianFactorGraph::shared_ptr remainingFactorGraph;
|
||||||
|
|
||||||
|
// This should NOT fail
|
||||||
|
std::tie(hybridBayesNet, remainingFactorGraph) =
|
||||||
|
linearized.eliminatePartialSequential(ordering);
|
||||||
|
EXPECT_LONGS_EQUAL(4, hybridBayesNet->size());
|
||||||
|
EXPECT_LONGS_EQUAL(1, remainingFactorGraph->size());
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
int main() {
|
||||||
|
TestResult tr;
|
||||||
|
return TestRegistry::runAllTests(tr);
|
||||||
|
}
|
||||||
|
/* ************************************************************************* */
|
||||||
|
|
@ -0,0 +1,58 @@
|
||||||
|
/* ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
* GTSAM Copyright 2010, Georgia Tech Research Corporation,
|
||||||
|
* Atlanta, Georgia 30332-0415
|
||||||
|
* All Rights Reserved
|
||||||
|
* Authors: Frank Dellaert, et al. (see THANKS for the full author list)
|
||||||
|
|
||||||
|
* See LICENSE for the license information
|
||||||
|
|
||||||
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @file testHybridValues.cpp
|
||||||
|
* @date Jul 28, 2022
|
||||||
|
* @author Shangjie Xue
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include <gtsam/base/Testable.h>
|
||||||
|
#include <gtsam/base/TestableAssertions.h>
|
||||||
|
#include <gtsam/discrete/Assignment.h>
|
||||||
|
#include <gtsam/discrete/DiscreteKey.h>
|
||||||
|
#include <gtsam/discrete/DiscreteValues.h>
|
||||||
|
#include <gtsam/hybrid/HybridValues.h>
|
||||||
|
#include <gtsam/inference/Key.h>
|
||||||
|
#include <gtsam/inference/Symbol.h>
|
||||||
|
#include <gtsam/linear/VectorValues.h>
|
||||||
|
#include <gtsam/nonlinear/Values.h>
|
||||||
|
|
||||||
|
// Include for test suite
|
||||||
|
#include <CppUnitLite/TestHarness.h>
|
||||||
|
|
||||||
|
using namespace std;
|
||||||
|
using namespace gtsam;
|
||||||
|
|
||||||
|
TEST(HybridValues, basics) {
|
||||||
|
HybridValues values;
|
||||||
|
values.insert(99, Vector2(2, 3));
|
||||||
|
values.insert(100, 3);
|
||||||
|
EXPECT(assert_equal(values.at(99), Vector2(2, 3)));
|
||||||
|
EXPECT(assert_equal(values.atDiscrete(100), int(3)));
|
||||||
|
|
||||||
|
values.print();
|
||||||
|
|
||||||
|
HybridValues values2;
|
||||||
|
values2.insert(100, 3);
|
||||||
|
values2.insert(99, Vector2(2, 3));
|
||||||
|
EXPECT(assert_equal(values2, values));
|
||||||
|
|
||||||
|
values2.insert(98, Vector2(2, 3));
|
||||||
|
EXPECT(!assert_equal(values2, values));
|
||||||
|
}
|
||||||
|
|
||||||
|
/* ************************************************************************* */
|
||||||
|
int main() {
|
||||||
|
TestResult tr;
|
||||||
|
return TestRegistry::runAllTests(tr);
|
||||||
|
}
|
||||||
|
/* ************************************************************************* */
|
||||||
|
|
@ -116,7 +116,7 @@ class FactorGraph {
|
||||||
using HasDerivedValueType = typename std::enable_if<
|
using HasDerivedValueType = typename std::enable_if<
|
||||||
std::is_base_of<FactorType, typename T::value_type>::value>::type;
|
std::is_base_of<FactorType, typename T::value_type>::value>::type;
|
||||||
|
|
||||||
/// Check if T has a value_type derived from FactorType.
|
/// Check if T has a pointer type derived from FactorType.
|
||||||
template <typename T>
|
template <typename T>
|
||||||
using HasDerivedElementType = typename std::enable_if<std::is_base_of<
|
using HasDerivedElementType = typename std::enable_if<std::is_base_of<
|
||||||
FactorType, typename T::value_type::element_type>::value>::type;
|
FactorType, typename T::value_type::element_type>::value>::type;
|
||||||
|
|
|
||||||
|
|
@ -10,7 +10,7 @@
|
||||||
* -------------------------------------------------------------------------- */
|
* -------------------------------------------------------------------------- */
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @file SymbolicISAM.h
|
* @file GaussianISAM.h
|
||||||
* @date July 29, 2013
|
* @date July 29, 2013
|
||||||
* @author Frank Dellaert
|
* @author Frank Dellaert
|
||||||
* @author Richard Roberts
|
* @author Richard Roberts
|
||||||
|
|
|
||||||
|
|
@ -55,6 +55,34 @@ class TestHybridGaussianFactorGraph(GtsamTestCase):
|
||||||
discrete_conditional = hbn.at(hbn.size() - 1).inner()
|
discrete_conditional = hbn.at(hbn.size() - 1).inner()
|
||||||
self.assertIsInstance(discrete_conditional, gtsam.DiscreteConditional)
|
self.assertIsInstance(discrete_conditional, gtsam.DiscreteConditional)
|
||||||
|
|
||||||
|
def test_optimize(self):
|
||||||
|
"""Test contruction of hybrid factor graph."""
|
||||||
|
noiseModel = gtsam.noiseModel.Unit.Create(3)
|
||||||
|
dk = gtsam.DiscreteKeys()
|
||||||
|
dk.push_back((C(0), 2))
|
||||||
|
|
||||||
|
jf1 = gtsam.JacobianFactor(X(0), np.eye(3), np.zeros((3, 1)),
|
||||||
|
noiseModel)
|
||||||
|
jf2 = gtsam.JacobianFactor(X(0), np.eye(3), np.ones((3, 1)),
|
||||||
|
noiseModel)
|
||||||
|
|
||||||
|
gmf = gtsam.GaussianMixtureFactor.FromFactors([X(0)], dk, [jf1, jf2])
|
||||||
|
|
||||||
|
hfg = gtsam.HybridGaussianFactorGraph()
|
||||||
|
hfg.add(jf1)
|
||||||
|
hfg.add(jf2)
|
||||||
|
hfg.push_back(gmf)
|
||||||
|
|
||||||
|
dtf = gtsam.DecisionTreeFactor([(C(0), 2)],"0 1")
|
||||||
|
hfg.add(dtf)
|
||||||
|
|
||||||
|
hbn = hfg.eliminateSequential(
|
||||||
|
gtsam.Ordering.ColamdConstrainedLastHybridGaussianFactorGraph(
|
||||||
|
hfg, [C(0)]))
|
||||||
|
|
||||||
|
# print("hbn = ", hbn)
|
||||||
|
hv = hbn.optimize()
|
||||||
|
self.assertEqual(hv.atDiscrete(C(0)), 1)
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,41 @@
|
||||||
|
"""
|
||||||
|
GTSAM Copyright 2010-2019, Georgia Tech Research Corporation,
|
||||||
|
Atlanta, Georgia 30332-0415
|
||||||
|
All Rights Reserved
|
||||||
|
|
||||||
|
See LICENSE for the license information
|
||||||
|
|
||||||
|
Unit tests for Hybrid Values.
|
||||||
|
Author: Shangjie Xue
|
||||||
|
"""
|
||||||
|
# pylint: disable=invalid-name, no-name-in-module, no-member
|
||||||
|
|
||||||
|
from __future__ import print_function
|
||||||
|
|
||||||
|
import unittest
|
||||||
|
|
||||||
|
import gtsam
|
||||||
|
import numpy as np
|
||||||
|
from gtsam.symbol_shorthand import C, X
|
||||||
|
from gtsam.utils.test_case import GtsamTestCase
|
||||||
|
|
||||||
|
|
||||||
|
class TestHybridGaussianFactorGraph(GtsamTestCase):
|
||||||
|
"""Unit tests for HybridValues."""
|
||||||
|
|
||||||
|
def test_basic(self):
|
||||||
|
"""Test contruction and basic methods of hybrid values."""
|
||||||
|
|
||||||
|
hv1 = gtsam.HybridValues()
|
||||||
|
hv1.insert(X(0), np.ones((3,1)))
|
||||||
|
hv1.insert(C(0), 2)
|
||||||
|
|
||||||
|
hv2 = gtsam.HybridValues()
|
||||||
|
hv2.insert(C(0), 2)
|
||||||
|
hv2.insert(X(0), np.ones((3,1)))
|
||||||
|
|
||||||
|
self.assertEqual(hv1.atDiscrete(C(0)), 2)
|
||||||
|
self.assertEqual(hv1.at(X(0))[0], np.ones((3,1))[0])
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
unittest.main()
|
||||||
Loading…
Reference in New Issue