C++ Neural Networks and Fuzzy Logic C++ Neural Networks and Fuzzy Logic
by Valluru B. Rao
M&T Books, IDG Books Worldwide, Inc.
ISBN: 1558515526   Pub Date: 06/01/95
  

Previous Table of Contents Next


Chapter 10
Adaptive Resonance Theory (ART)

Introduction

Grossberg’s Adaptive Resonance Theory, developed further by Grossberg and Carpenter, is for the categorization of patterns using the competitive learning paradigm. It introduces a gain control and a reset to make certain that learned categories are retained even while new categories are learned and thereby addresses the plasticity–stability dilemma.

Adaptive Resonance Theory makes much use of a competitive learning paradigm. A criterion is developed to facilitate the occurrence of winner-take-all phenomenon. A single node with the largest value for the set criterion is declared the winner within its layer, and it is said to classify a pattern class. If there is a tie for the winning neuron in a layer, then an arbitrary rule, such as the first of them in a serial order, can be taken as the winner.

The neural network developed for this theory establishes a system that is made up of two subsystems, one being the attentional subsystem, and this contains the unit for gain control. The other is an orienting subsystem, and this contains the unit for reset. During the operation of the network modeled for this theory, patterns emerge in the attentional subsystem and are called traces of STM (short-term memory). Traces of LTM (long-term memory) are in the connection weights between the input layer and output layer.

The network uses processing with feedback between its two layers, until resonance occurs. Resonance occurs when the output in the first layer after feedback from the second layer matches the original pattern used as input for the first layer in that processing cycle. A match of this type does not have to be perfect. What is required is that the degree of match, measured suitably, exceeds a predetermined level, termed vigilance parameter. Just as a photograph matches the likeness of the subject to a greater degree when the granularity is higher, the pattern match gets finer when the vigilance parameter is closer to 1.

The Network for ART1

The neural network for the adaptive resonance theory or ART1 model consists of the following:

  A layer of neurons, called the F1 layer (input layer or comparison layer)
  A node for each layer as a gain control unit
  A layer of neurons, called the F2 layer (output layer or recognition layer)
  A node as a reset unit
  Bottom-up connections from F1 layer to F2 layer
  Top-down connections from F2 layer to F1 layer
  Inhibitory connection (negative weight) form F2 layer to gain control
  Excitatory connection (positive weight) from gain control to a layer
  Inhibitory connection from F1 layer to reset node
  Excitatory connection from reset node to F2 layer

A Simplified Diagram of Network Layout


Figure 10.1  simplified diagram of the neural network for an ART1 model.

Processing in ART1

The ART1 paradigm, just like the Kohonen Self-Organizing Map to be introduced in Chapter 11, performs data clustering on input data; like inputs are clustered together into a category. As an example, you can use a data clustering algorithm such as ART1 for Optical Character Recognition (OCR), where you try to match different samples of a letter to its ASCII equivalent. Particular attention is made in the ART1 paradigm to ensure that old information is not thrown away while new information is assimilated.

An input vector, when applied to an ART1 system, is first compared to existing patterns in the system. If there is a close enough match within a specified tolerance (as indicated by a vigilance parameter), then that stored pattern is made to resemble the input pattern further and the classification operation is complete. If the input pattern does not resemble any of the stored patterns in the system, then a new category is created with a new stored pattern that resembles the input pattern.

Special Features of the ART1 Model

One special feature of an ART1 model is that a two-thirds rule is necessary to determine the activity of neurons in the F1 layer. There are three input sources to each neuron in layer F1. They are the external input, the output of gain control, and the outputs of F2 layer neurons. The F1neurons will not fire unless at least two of the three inputs are active. The gain control unit and the two-thirds rule together ensure proper response from the input layer neurons. A second feature is that a vigilance parameter is used to determine the activity of the reset unit, which is activated whenever there is no match found among existing patterns during classification.

Notation for ART1 Calculations

Let us list the various symbols we will use to describe the operation of a neural network for an ART1 model:


wij Weight on the connection from the ith neuron in the F1 layer to the jth neuron in the F2 layer
vji Weight on the connection from the jth neuron in the F2 layer to the ith neuron on the F1 layer
ai Activation of the ith neuron in the F1 layer
bj Activation of the jth neuron in the F2 layer
xi Output of the ith neuron in the F1 layer
yj Output of the jth neuron in the F2 layer
zi Input to the ith neuron in F1 layer from F2 layer
ρ Vigilance parameter, positive and no greater than 1 (0<ρ ≤ 1)
m Number of neurons in the F1 layer
n Number of neurons in the F2 layer
I Input vector
Si Sum of the components of the input vector
Sx Sum of the outputs of neurons in the F1 layer
A, C, D Parameters with positive values or zero
L Parameter with value greater than 1
B Parameter with value less than D + 1 but at least as large as either D or 1
r Index of winner of competition in the F2 layer


Previous Table of Contents Next

Copyright © IDG Books Worldwide, Inc.