Abstract
A comparison was carried out of several learning algorithms for training feedforward neural networks with linear threshold units. These learning algorithms do not require an a priori network architecture, but add neurons at will during training. The performance of these algorithms was compared by using training sets with a particular correlation of the input patterns over the full range of possible correlations. For binary input patterns we present a fast method for the selection of input patterns that can be identified by a single neuron. This method is not based on the perceptron learning rule.
| Original language | English |
|---|---|
| Pages (from-to) | 555-559 |
| Number of pages | 5 |
| Journal | Europhysics Letters |
| Volume | 18 |
| Issue number | 6 |
| Publication status | Published - 15 Mar 1992 |
Keywords
- LAYERED NETWORKS
Fingerprint
Dive into the research topics of 'A FAST PARTITIONING ALGORITHM AND A COMPARISON OF BINARY FEEDFORWARD NEURAL NETWORKS'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver