Mercurial > hg > camir-aes2014
diff toolboxes/FullBNT-1.0.7/docs/majorFeatures.html @ 0:e9a9cd732c1e tip
first hg version after svn
author | wolffd |
---|---|
date | Tue, 10 Feb 2015 15:05:51 +0000 |
parents | |
children |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/toolboxes/FullBNT-1.0.7/docs/majorFeatures.html Tue Feb 10 15:05:51 2015 +0000 @@ -0,0 +1,113 @@ + +<h2><a name="features">Major features</h2> +<ul> + +<li> BNT supports many types of +<b>conditional probability distributions</b> (nodes), +and it is easy to add more. +<ul> +<li>Tabular (multinomial) +<li>Gaussian +<li>Softmax (logistic/ sigmoid) +<li>Multi-layer perceptron (neural network) +<li>Noisy-or +<li>Deterministic +</ul> +<p> + +<li> BNT supports <b>decision and utility nodes</b>, as well as chance +nodes, +i.e., influence diagrams as well as Bayes nets. +<p> + +<li> BNT supports static and dynamic BNs (useful for modelling dynamical systems +and sequence data). +<p> + +<li> BNT supports many different <b>inference algorithms</b>, +and it is easy to add more. + +<ul> +<li> Exact inference for static BNs: +<ul> +<li>junction tree +<li>variable elimination +<li>brute force enumeration (for discrete nets) +<li>linear algebra (for Gaussian nets) +<li>Pearl's algorithm (for polytrees) +<li>quickscore (for QMR) +</ul> + +<p> +<li> Approximate inference for static BNs: +<ul> +<li>likelihood weighting +<li> Gibbs sampling +<li>loopy belief propagation +</ul> + +<p> +<li> Exact inference for DBNs: +<ul> +<li>junction tree +<li>frontier algorithm +<li>forwards-backwards (for HMMs) +<li>Kalman-RTS (for LDSs) +</ul> + +<p> +<li> Approximate inference for DBNs: +<ul> +<li>Boyen-Koller +<li>factored-frontier/loopy belief propagation +</ul> + +</ul> +<p> + +<li> +BNT supports several methods for <b>parameter learning</b>, +and it is easy to add more. +<ul> + +<li> Batch MLE/MAP parameter learning using EM. +(Each node type has its own M method, e.g. softmax nodes use IRLS,<br> +and each inference engine has its own E method, so the code is fully modular.) + +<li> Sequential/batch Bayesian parameter learning (for fully observed tabular nodes only). +</ul> + + +<p> +<li> +BNT supports several methods for <b>regularization</b>, +and it is easy to add more. +<ul> +<li> Any node can have its parameters clamped (made non-adjustable). +<li> Any set of compatible nodes can have their parameters tied (c.f., +weight sharing in a neural net). +<li> Some node types (e.g., tabular) supports priors for MAP estimation. +<li> Gaussian covariance matrices can be declared full or diagonal, and can +be tied across states of their discrete parents (if any). +</ul> + +<p> +<li> +BNT supports several methods for <b>structure learning</b>, +and it is easy to add more. +<ul> + +<li> Bayesian structure learning, +using MCMC or local search (for fully observed tabular nodes only). + +<li> Constraint-based structure learning (IC/PC and IC*/FCI). +</ul> + + +<p> +<li> The source code is extensively documented, object-oriented, and free, making it +an excellent tool for teaching, research and rapid prototyping. + +</ul> + + \ No newline at end of file