The neural network stops before it starts? (minimum gradient reached ) error
Show older comments
Hello every body. I am using Neural network in my research graduation project " Pattern Identification ". I have built a custom neural network, and I used the function 'trainlm' .
The Problem is that, once I start the training process, the training window appears and tells me that the minimum gradient is reached and the neural network stops even before it starts !
( The architecture of my NN is complicated, and I'm using three inputs with 13 layers each of them carries different neurons ! )
I wish I can find someone helping me
Thanks in advance !
2 Comments
EanX
on 30 Sep 2015
I have a similar problem in designing a NARNET.
Currently best performance was obtained with 15 hidden nodes (1 hidden layer), mapminmax for processing function for inputs (ymin=0, ymax=10), tansig as transfer function for layer 1 and poslin for output layer.
I noticed that if I change output transfer function to the more commonly employed purelin I can avoid "minimum gradient reached" issue, but I used poslin to obtain forecasts always positive.
Can anyone enlighten me on this topic?
Thanks in advance !
Greg Heath
on 30 Sep 2015
Edited: Greg Heath
on 30 Sep 2015
Do not change defaults until you have obtained the best default result.
Are you looping over 10 or more weight initializations and getting premature MinGrad reached on all of them?
Accepted Answer
More Answers (0)
Categories
Find more on Deep Learning Toolbox in Help Center and File Exchange
Products
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!