Incremental extreme learning machine
This new theory shows that in order to let SLFNs work as universal approximators, one may simply randomly choose input-to-hidden nodes, and then we only need to adjust the output weights linking the hidden layer and the output layer. In such SLFNs implementations, the activation functions for additi...
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis |
Published: |
2008
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/3804 |