view toolboxes/FullBNT-1.0.7/nethelp3.3/olgd.htm @ 0:e9a9cd732c1e tip

first hg version after svn
author wolffd
date Tue, 10 Feb 2015 15:05:51 +0000
parents
children
line wrap: on
line source
<html>
<head>
<title>
Netlab Reference Manual olgd
</title>
</head>
<body>
<H1> olgd
</H1>
<h2>
Purpose
</h2>
On-line gradient descent optimization.

<p><h2>
Description
</h2>
<CODE>[net, options, errlog, pointlog] = olgd(net, options, x, t)</CODE> uses 
on-line gradient descent to find a local minimum of the error function for the
network
<CODE>net</CODE> computed on the input data <CODE>x</CODE> and target values
<CODE>t</CODE>. A log of the error values
after each cycle is (optionally) returned in <CODE>errlog</CODE>, and a log
of the points visited is (optionally) returned in <CODE>pointlog</CODE>.
Because the gradient is computed on-line (i.e. after each pattern)
this can be quite inefficient in Matlab.

<p>The error function value at final weight vector is returned
in <CODE>options(8)</CODE>.

<p>The optional parameters have the following interpretations.

<p><CODE>options(1)</CODE> is set to 1 to display error values; also logs error 
values in the return argument <CODE>errlog</CODE>, and the points visited
in the return argument <CODE>pointslog</CODE>.  If <CODE>options(1)</CODE> is set to 0,
then only warning messages are displayed.  If <CODE>options(1)</CODE> is -1,
then nothing is displayed.

<p><CODE>options(2)</CODE> is the precision required for the value
of <CODE>x</CODE> at the solution. If the absolute difference between
the values of <CODE>x</CODE> between two successive steps is less than
<CODE>options(2)</CODE>, then this condition is satisfied.

<p><CODE>options(3)</CODE> is the precision required of the objective
function at the solution.  If the absolute difference between the
error functions between two successive steps is less than
<CODE>options(3)</CODE>, then this condition is satisfied.
Both this and the previous condition must be
satisfied for termination. Note that testing the function value at each
iteration roughly halves the speed of the algorithm.

<p><CODE>options(5)</CODE> determines whether the patterns are sampled randomly
with replacement. If it is 0 (the default), then patterns are sampled
in order.

<p><CODE>options(6)</CODE> determines if the learning rate decays.  If it is 1
then the learning rate decays at a rate of <CODE>1/t</CODE>.  If it is 0
(the default) then the learning rate is constant.

<p><CODE>options(9)</CODE> should be set to 1 to check the user defined gradient
function.

<p><CODE>options(10)</CODE> returns the total number of function evaluations (including
those in any line searches).

<p><CODE>options(11)</CODE> returns the total number of gradient evaluations.

<p><CODE>options(14)</CODE> is the maximum number of iterations (passes through
the complete pattern set); default 100.

<p><CODE>options(17)</CODE> is the momentum; default 0.5.

<p><CODE>options(18)</CODE> is the learning rate; default 0.01.

<p><h2>
Examples
</h2>
The following example performs on-line gradient descent on an MLP with
random sampling from the pattern set.
<PRE>

net = mlp(5, 3, 1, 'linear');
options = foptions;
options(18) = 0.01;
options(5) = 1;
net = olgd(net, options, x, t);
</PRE>


<p><h2>
See Also
</h2>
<CODE><a href="graddesc.htm">graddesc</a></CODE><hr>
<b>Pages:</b>
<a href="index.htm">Index</a>
<hr>
<p>Copyright (c) Ian T Nabney (1996-9)


</body>
</html>