How to use different transfer functions within the same layer of a neural network?
2 views (last 30 days)
Show older comments
For example, hidden layer with 3 different neurons: 1.tansig, 2.logsig, 3.elliotsig. I understand that this influences normalization range and that training might not resault in good performance but i need it to test some other aspects. Thank you.
0 Comments
Answers (1)
Greg Heath
on 24 Apr 2013
If it were possible with the NNTBX, you would have to design a custom net.
The only way I can see doing it is to have 3 hidden layers that are only connected to the input and output but not to each other.
See the custom network section in the documentation.
Hope this helps.
Thank you for formally accepting my answer
Greg
See Also
Categories
Find more on Sequence and Numeric Feature Data Workflows in Help Center and File Exchange
Products
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!