annotate toolboxes/FullBNT-1.0.7/netlab3.3/demgmm4.m @ 0:cc4b1211e677 tip

initial commit to HG from Changeset: 646 (e263d8a21543) added further path and more save "camirversion.m"
author Daniel Wolff
date Fri, 19 Aug 2016 13:07:06 +0200
parents
children
rev   line source
Daniel@0 1 %DEMGMM4 Demonstrate density modelling with a Gaussian mixture model.
Daniel@0 2 %
Daniel@0 3 % Description
Daniel@0 4 % The problem consists of modelling data generated by a mixture of
Daniel@0 5 % three Gaussians in 2 dimensions with a mixture model using full
Daniel@0 6 % covariance matrices. The priors are 0.3, 0.5 and 0.2; the centres
Daniel@0 7 % are (2, 3.5), (0, 0) and (0,2); the variances are (0.16, 0.64) axis
Daniel@0 8 % aligned, (0.25, 1) rotated by 30 degrees and the identity matrix. The
Daniel@0 9 % first figure contains a scatter plot of the data.
Daniel@0 10 %
Daniel@0 11 % A Gaussian mixture model with three components is trained using EM.
Daniel@0 12 % The parameter vector is printed before training and after training.
Daniel@0 13 % The user should press any key to continue at these points. The
Daniel@0 14 % parameter vector consists of priors (the column), and centres (given
Daniel@0 15 % as (x, y) pairs as the next two columns). The covariance matrices
Daniel@0 16 % are printed separately.
Daniel@0 17 %
Daniel@0 18 % The second figure is a 3 dimensional view of the density function,
Daniel@0 19 % while the third shows the axes of the 1-standard deviation ellipses
Daniel@0 20 % for the three components of the mixture model.
Daniel@0 21 %
Daniel@0 22 % See also
Daniel@0 23 % GMM, GMMINIT, GMMEM, GMMPROB, GMMUNPAK
Daniel@0 24 %
Daniel@0 25
Daniel@0 26 % Copyright (c) Ian T Nabney (1996-2001)
Daniel@0 27
Daniel@0 28
Daniel@0 29 % Generate the data
Daniel@0 30
Daniel@0 31 ndata = 500;
Daniel@0 32
Daniel@0 33 % Fix the seeds for reproducible results
Daniel@0 34 randn('state', 42);
Daniel@0 35 rand('state', 42);
Daniel@0 36 data = randn(ndata, 2);
Daniel@0 37 prior = [0.3 0.5 0.2];
Daniel@0 38 % Mixture model swaps clusters 1 and 3
Daniel@0 39 datap = [0.2 0.5 0.3];
Daniel@0 40 datac = [0 2; 0 0; 2 3.5];
Daniel@0 41 datacov = repmat(eye(2), [1 1 3]);
Daniel@0 42 data1 = data(1:prior(1)*ndata,:);
Daniel@0 43 data2 = data(prior(1)*ndata+1:(prior(2)+prior(1))*ndata, :);
Daniel@0 44 data3 = data((prior(1)+prior(2))*ndata +1:ndata, :);
Daniel@0 45
Daniel@0 46 % First cluster has axis aligned variance and centre (2, 3.5)
Daniel@0 47 data1(:, 1) = data1(:, 1)*0.4 + 2.0;
Daniel@0 48 data1(:, 2) = data1(:, 2)*0.8 + 3.5;
Daniel@0 49 datacov(:, :, 3) = [0.4*0.4 0; 0 0.8*0.8];
Daniel@0 50
Daniel@0 51 % Second cluster has variance axes rotated by 30 degrees and centre (0, 0)
Daniel@0 52 rotn = [cos(pi/6) -sin(pi/6); sin(pi/6) cos(pi/6)];
Daniel@0 53 data2(:,1) = data2(:, 1)*0.5;
Daniel@0 54 data2 = data2*rotn;
Daniel@0 55 datacov(:, :, 2) = rotn' * [0.25 0; 0 1] * rotn;
Daniel@0 56
Daniel@0 57 % Third cluster is at (0,2)
Daniel@0 58 data3 = data3 + repmat([0 2], prior(3)*ndata, 1);
Daniel@0 59
Daniel@0 60 % Put the dataset together again
Daniel@0 61 data = [data1; data2; data3];
Daniel@0 62
Daniel@0 63 clc
Daniel@0 64 disp('This demonstration illustrates the use of a Gaussian mixture model')
Daniel@0 65 disp('with full covariance matrices to approximate the unconditional ')
Daniel@0 66 disp('probability density of data in a two-dimensional space.')
Daniel@0 67 disp('We begin by generating the data from a mixture of three Gaussians and')
Daniel@0 68 disp('plotting it.')
Daniel@0 69 disp(' ')
Daniel@0 70 disp('The first cluster has axis aligned variance and centre (0, 2).')
Daniel@0 71 disp('The second cluster has variance axes rotated by 30 degrees')
Daniel@0 72 disp('and centre (0, 0). The third cluster has unit variance and centre')
Daniel@0 73 disp('(2, 3.5).')
Daniel@0 74 disp(' ')
Daniel@0 75 disp('Press any key to continue.')
Daniel@0 76 pause
Daniel@0 77
Daniel@0 78 fh1 = figure;
Daniel@0 79 plot(data(:, 1), data(:, 2), 'o')
Daniel@0 80 set(gca, 'Box', 'on')
Daniel@0 81
Daniel@0 82 % Set up mixture model
Daniel@0 83 ncentres = 3;
Daniel@0 84 input_dim = 2;
Daniel@0 85 mix = gmm(input_dim, ncentres, 'full');
Daniel@0 86
Daniel@0 87 % Initialise the model parameters from the data
Daniel@0 88 options = foptions;
Daniel@0 89 options(14) = 5; % Just use 5 iterations of k-means in initialisation
Daniel@0 90 mix = gmminit(mix, data, options);
Daniel@0 91
Daniel@0 92 % Print out model
Daniel@0 93 clc
Daniel@0 94 disp('The mixture model has three components and full covariance')
Daniel@0 95 disp('matrices. The model parameters after initialisation using the')
Daniel@0 96 disp('k-means algorithm are as follows')
Daniel@0 97 disp(' Priors Centres')
Daniel@0 98 disp([mix.priors' mix.centres])
Daniel@0 99 disp('Covariance matrices are')
Daniel@0 100 disp(mix.covars)
Daniel@0 101 disp('Press any key to continue.')
Daniel@0 102 pause
Daniel@0 103
Daniel@0 104 % Set up vector of options for EM trainer
Daniel@0 105 options = zeros(1, 18);
Daniel@0 106 options(1) = 1; % Prints out error values.
Daniel@0 107 options(14) = 50; % Number of iterations.
Daniel@0 108
Daniel@0 109 disp('We now train the model using the EM algorithm for 50 iterations.')
Daniel@0 110 disp(' ')
Daniel@0 111 disp('Press any key to continue.')
Daniel@0 112 pause
Daniel@0 113 [mix, options, errlog] = gmmem(mix, data, options);
Daniel@0 114
Daniel@0 115 % Print out model
Daniel@0 116 disp(' ')
Daniel@0 117 disp('The trained model has priors and centres:')
Daniel@0 118 disp(' Priors Centres')
Daniel@0 119 disp([mix.priors' mix.centres])
Daniel@0 120 disp('The data generator has priors and centres')
Daniel@0 121 disp(' Priors Centres')
Daniel@0 122 disp([datap' datac])
Daniel@0 123 disp('Model covariance matrices are')
Daniel@0 124 disp(mix.covars(:, :, 1))
Daniel@0 125 disp(mix.covars(:, :, 2))
Daniel@0 126 disp(mix.covars(:, :, 3))
Daniel@0 127 disp('Data generator covariance matrices are')
Daniel@0 128 disp(datacov(:, :, 1))
Daniel@0 129 disp(datacov(:, :, 2))
Daniel@0 130 disp(datacov(:, :, 3))
Daniel@0 131 disp('Note the close correspondence between these parameters and those')
Daniel@0 132 disp('of the distribution used to generate the data. The match for')
Daniel@0 133 disp('covariance matrices is not that close, but would be improved with')
Daniel@0 134 disp('more iterations of the training algorithm.')
Daniel@0 135 disp(' ')
Daniel@0 136 disp('Press any key to continue.')
Daniel@0 137 pause
Daniel@0 138
Daniel@0 139 clc
Daniel@0 140 disp('We now plot the density given by the mixture model as a surface plot.')
Daniel@0 141 disp(' ')
Daniel@0 142 disp('Press any key to continue.')
Daniel@0 143 pause
Daniel@0 144
Daniel@0 145 % Plot the result
Daniel@0 146 x = -4.0:0.2:5.0;
Daniel@0 147 y = -4.0:0.2:5.0;
Daniel@0 148 [X, Y] = meshgrid(x,y);
Daniel@0 149 X = X(:);
Daniel@0 150 Y = Y(:);
Daniel@0 151 grid = [X Y];
Daniel@0 152 Z = gmmprob(mix, grid);
Daniel@0 153 Z = reshape(Z, length(x), length(y));
Daniel@0 154 c = mesh(x, y, Z);
Daniel@0 155 hold on
Daniel@0 156 title('Surface plot of probability density')
Daniel@0 157 hold off
Daniel@0 158 drawnow
Daniel@0 159
Daniel@0 160 clc
Daniel@0 161 disp('The final plot shows the centres and widths, given by one standard')
Daniel@0 162 disp('deviation, of the three components of the mixture model. The axes')
Daniel@0 163 disp('of the ellipses of constant density are shown.')
Daniel@0 164 disp(' ')
Daniel@0 165 disp('Press any key to continue.')
Daniel@0 166 pause
Daniel@0 167
Daniel@0 168 % Try to calculate a sensible position for the second figure, below the first
Daniel@0 169 fig1_pos = get(fh1, 'Position');
Daniel@0 170 fig2_pos = fig1_pos;
Daniel@0 171 fig2_pos(2) = fig2_pos(2) - fig1_pos(4) - 30;
Daniel@0 172 fh2 = figure('Position', fig2_pos);
Daniel@0 173
Daniel@0 174 h3 = plot(data(:, 1), data(:, 2), 'bo');
Daniel@0 175 axis equal;
Daniel@0 176 hold on
Daniel@0 177 title('Plot of data and covariances')
Daniel@0 178 for i = 1:ncentres
Daniel@0 179 [v,d] = eig(mix.covars(:,:,i));
Daniel@0 180 for j = 1:2
Daniel@0 181 % Ensure that eigenvector has unit length
Daniel@0 182 v(:,j) = v(:,j)/norm(v(:,j));
Daniel@0 183 start=mix.centres(i,:)-sqrt(d(j,j))*(v(:,j)');
Daniel@0 184 endpt=mix.centres(i,:)+sqrt(d(j,j))*(v(:,j)');
Daniel@0 185 linex = [start(1) endpt(1)];
Daniel@0 186 liney = [start(2) endpt(2)];
Daniel@0 187 line(linex, liney, 'Color', 'k', 'LineWidth', 3)
Daniel@0 188 end
Daniel@0 189 % Plot ellipses of one standard deviation
Daniel@0 190 theta = 0:0.02:2*pi;
Daniel@0 191 x = sqrt(d(1,1))*cos(theta);
Daniel@0 192 y = sqrt(d(2,2))*sin(theta);
Daniel@0 193 % Rotate ellipse axes
Daniel@0 194 ellipse = (v*([x; y]))';
Daniel@0 195 % Adjust centre
Daniel@0 196 ellipse = ellipse + ones(length(theta), 1)*mix.centres(i,:);
Daniel@0 197 plot(ellipse(:,1), ellipse(:,2), 'r-');
Daniel@0 198 end
Daniel@0 199 hold off
Daniel@0 200
Daniel@0 201 disp('Note how the data cluster positions and widths are captured by')
Daniel@0 202 disp('the mixture model.')
Daniel@0 203 disp(' ')
Daniel@0 204 disp('Press any key to end.')
Daniel@0 205 pause
Daniel@0 206
Daniel@0 207 close(fh1);
Daniel@0 208 close(fh2);
Daniel@0 209 clear all;
Daniel@0 210