The Cascade-Correlation Learning Architecture

Scott E. Fahlman and Christian Lebiere


Reviewer

John Trevithick (j.trevithick@auckland.ac.nz)


Reference

Fahlman, S. E. and Lebiere, C. (1990) "The Cascade-Correlation Learning Architecture", in Advances in Neural Information Processing Systems 2, D. S. Touretzky (ed.), Morgan Kaufmann Publishers, Los Altos CA, pp.524-532.
Downloaded from http://www.cs.cmu.edu/~sef/sefPubs.htm


Keywords

Cascade-Correlation, ANNs, feature-detector, constructive algorithm, topology, back-propagation.


Related Papers

Fahlman, S. E. (1991) "The Recurrent Cascade-Correlation Architecture" in Advances in Neural Information Processing Systems 3, D. S. Touretzky (ed.), Morgan Kaufmann, Los Altos CA, pp. 190-196.
Fahlman, S. E. (1988) "Faster-Learning Variations on Back-Propagation: An Empirical Study" in Proceedings, 1988 Connectionist Models Summer School, Morgan-Kaufmann, Los Altos CA.
S. Baluja and S.E. Fahlman, “Reducing network depth in the cascade-correlation”. Technical Report CMU-CS-94-209, Carnegie Mellon University (1994).

Summary

  1. Why is Back-Propagation Learning So Slow? The authors propose some key causes behind the need for BPL's many epochs.
  2. Description of Cascade-Correlation. Describes the algorithm with reference to cascade architecture and correlation of residual error signal.
  3. Benchmark Results. Shows outcome of testing against two known problems: the two-spirals problem and the N-input parity problem.
  4. Discussion. Briefly summarises perceived advantages of the algorithm.
  5. Relation To Other Work. The authors identify the origins of many of their base ideas and approaches.

Abstract:
Cascade-Correlation is a new architecture and supervised learning algorithm for artificial neural networks. Instead of just adjusting the weights in a network of fixed topology, Cascade-Correlation begins with a minimal network, then automatically trains and adds new hidden units one by one, creating a multi-layer structure. Once a new hidden unit has been added to the network, its input-side weights are frozen. This unit then becomes a permanent feature-detector in the network, available for producing outputs or for creating other, more complex feature detectors. The Cascade-Correlation architecture has several advantages over existing algorithms: it learns very quickly, the network determines its own size and topology, it retains the structures it has built even if the training set changes, and it requires no back-propagation of error signals through the connections of the network.


Evaluation

Strengths of the paper:

Weaknesses of the paper: