Abstract
Feedforward neural network structures have extensively been considered in the literature. In a significant volume of research and development studies hyperbolic tangent type of a neuronal nonlinearity has been utilized. This paper dwells on the widely used neuronal activation functions as well as two new ones composed of sines and cosines, and a sinc function characterizing the firing of a neuron. The viewpoint here is to consider the hidden layer(s) as transforming blocks composed of nonlinear basis functions, which may assume different forms. This paper considers 8 different activation functions which are differentiable and utilizes Levenberg-Marquardt algorithm for parameter tuning purposes. The studies carried out have a guiding quality based on empirical results on several training data sets.
| Original language | English |
|---|---|
| Pages (from-to) | 63-79 |
| Number of pages | 17 |
| Journal | Neural Processing Letters |
| Volume | 28 |
| Issue number | 2 |
| DOIs | |
| Publication status | Published - Oct 2008 |
| Externally published | Yes |
Keywords
- Activation functions
- Dynamical system identification
- Levenberg-Marquardt algorithm
Fingerprint
Dive into the research topics of 'Novel neuronal activation functions for feedforward neural networks'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver