Mercurial > hg > camir-aes2014
comparison toolboxes/FullBNT-1.0.7/docs/majorFeatures.html @ 0:e9a9cd732c1e tip
first hg version after svn
author | wolffd |
---|---|
date | Tue, 10 Feb 2015 15:05:51 +0000 |
parents | |
children |
comparison
equal
deleted
inserted
replaced
-1:000000000000 | 0:e9a9cd732c1e |
---|---|
1 | |
2 <h2><a name="features">Major features</h2> | |
3 <ul> | |
4 | |
5 <li> BNT supports many types of | |
6 <b>conditional probability distributions</b> (nodes), | |
7 and it is easy to add more. | |
8 <ul> | |
9 <li>Tabular (multinomial) | |
10 <li>Gaussian | |
11 <li>Softmax (logistic/ sigmoid) | |
12 <li>Multi-layer perceptron (neural network) | |
13 <li>Noisy-or | |
14 <li>Deterministic | |
15 </ul> | |
16 <p> | |
17 | |
18 <li> BNT supports <b>decision and utility nodes</b>, as well as chance | |
19 nodes, | |
20 i.e., influence diagrams as well as Bayes nets. | |
21 <p> | |
22 | |
23 <li> BNT supports static and dynamic BNs (useful for modelling dynamical systems | |
24 and sequence data). | |
25 <p> | |
26 | |
27 <li> BNT supports many different <b>inference algorithms</b>, | |
28 and it is easy to add more. | |
29 | |
30 <ul> | |
31 <li> Exact inference for static BNs: | |
32 <ul> | |
33 <li>junction tree | |
34 <li>variable elimination | |
35 <li>brute force enumeration (for discrete nets) | |
36 <li>linear algebra (for Gaussian nets) | |
37 <li>Pearl's algorithm (for polytrees) | |
38 <li>quickscore (for QMR) | |
39 </ul> | |
40 | |
41 <p> | |
42 <li> Approximate inference for static BNs: | |
43 <ul> | |
44 <li>likelihood weighting | |
45 <li> Gibbs sampling | |
46 <li>loopy belief propagation | |
47 </ul> | |
48 | |
49 <p> | |
50 <li> Exact inference for DBNs: | |
51 <ul> | |
52 <li>junction tree | |
53 <li>frontier algorithm | |
54 <li>forwards-backwards (for HMMs) | |
55 <li>Kalman-RTS (for LDSs) | |
56 </ul> | |
57 | |
58 <p> | |
59 <li> Approximate inference for DBNs: | |
60 <ul> | |
61 <li>Boyen-Koller | |
62 <li>factored-frontier/loopy belief propagation | |
63 </ul> | |
64 | |
65 </ul> | |
66 <p> | |
67 | |
68 <li> | |
69 BNT supports several methods for <b>parameter learning</b>, | |
70 and it is easy to add more. | |
71 <ul> | |
72 | |
73 <li> Batch MLE/MAP parameter learning using EM. | |
74 (Each node type has its own M method, e.g. softmax nodes use IRLS,<br> | |
75 and each inference engine has its own E method, so the code is fully modular.) | |
76 | |
77 <li> Sequential/batch Bayesian parameter learning (for fully observed tabular nodes only). | |
78 </ul> | |
79 | |
80 | |
81 <p> | |
82 <li> | |
83 BNT supports several methods for <b>regularization</b>, | |
84 and it is easy to add more. | |
85 <ul> | |
86 <li> Any node can have its parameters clamped (made non-adjustable). | |
87 <li> Any set of compatible nodes can have their parameters tied (c.f., | |
88 weight sharing in a neural net). | |
89 <li> Some node types (e.g., tabular) supports priors for MAP estimation. | |
90 <li> Gaussian covariance matrices can be declared full or diagonal, and can | |
91 be tied across states of their discrete parents (if any). | |
92 </ul> | |
93 | |
94 <p> | |
95 <li> | |
96 BNT supports several methods for <b>structure learning</b>, | |
97 and it is easy to add more. | |
98 <ul> | |
99 | |
100 <li> Bayesian structure learning, | |
101 using MCMC or local search (for fully observed tabular nodes only). | |
102 | |
103 <li> Constraint-based structure learning (IC/PC and IC*/FCI). | |
104 </ul> | |
105 | |
106 | |
107 <p> | |
108 <li> The source code is extensively documented, object-oriented, and free, making it | |
109 an excellent tool for teaching, research and rapid prototyping. | |
110 | |
111 </ul> | |
112 | |
113 |