newrb
Design radial basis network
Description
takes two of these arguments:net = newrb(P,T,goal,spread,MN,DF)
P—R-by-Qmatrix ofQinput vectorsT—S-by-Qmatrix ofQtarget class vectorsgoal— Mean squared error goalspread— Spread of radial basis functionsMN— Maximum number of neuronsDF— Number of neurons to add between displays
Radial basis networks can be used to approximate functions. newrb
adds neurons to the hidden layer of a radial basis network until it meets the specified mean
squared error goal.
The larger spread is, the smoother the function approximation. Too
large a spread means a lot of neurons are required to fit a fast-changing function. Too
small a spread means many neurons are required to fit a smooth function, and the network
might not generalize well. Call newrb with different spreads to find the
best value for a given problem.
Examples
Input Arguments
Output Arguments
Algorithms
newrb creates a two-layer network. The first layer has
radbas neurons, and calculates its weighted inputs with
dist and its net input with netprod. The second layer
has purelin neurons, and calculates its weighted input with
dotprod and its net inputs with netsum. Both layers
have biases.
Initially the radbas layer has no neurons. The following steps are
repeated until the network’s mean squared error falls below goal.
The network is simulated.
The input vector with the greatest error is found.
A
radbasneuron is added with weights equal to that vector.The
purelinlayer weights are redesigned to minimize error.
Version History
Introduced before R2006a