How to use different transfer functions within the same layer of a neural network?
2 visualizzazioni (ultimi 30 giorni)
Mostra commenti meno recenti
For example, hidden layer with 3 different neurons: 1.tansig, 2.logsig, 3.elliotsig. I understand that this influences normalization range and that training might not resault in good performance but i need it to test some other aspects. Thank you.
0 Commenti
Risposte (1)
Greg Heath
il 24 Apr 2013
If it were possible with the NNTBX, you would have to design a custom net.
The only way I can see doing it is to have 3 hidden layers that are only connected to the input and output but not to each other.
See the custom network section in the documentation.
Hope this helps.
Thank you for formally accepting my answer
Greg
Vedere anche
Categorie
Scopri di più su Sequence and Numeric Feature Data Workflows in Help Center e File Exchange
Prodotti
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!