Simple Recurrent Networks Learn Context-Free and Context-Sensitive Languages by Counting
- 1 September 2001
- journal article
- Published by MIT Press in Neural Computation
- Vol. 13 (9) , 2093-2118
- https://doi.org/10.1162/089976601750399326
Abstract
It has been shown that if a recurrent neural network (RNN) learns to process a regular language, one can extract a finite-state machine (FSM) by treating regions of phase-space as FSM states. However, it has also been shown that one can construct an RNN to implement Turing machines by using RNN dynamics as counters. But how does a network learn languages that require counting? Rodriguez, Wiles, and Elman (1999) showed that a simple recurrent network (SRN) can learn to process a simple context-free language (CFL) by counting up and down. This article extends that to show a range of language tasks in which an SRN develops solutions that not only count but also copy and store counting information. In one case, the network stores information like an explicit storage mechanism. In other cases, the network stores information more indirectly in trajectories that are sensitive to slight displacements that depend on context. In this sense, an SRN can learn analog computation as a set of interdependent counters. This demonstrates how SRNs may be an alternative psychological model of language or sequence processing.Keywords
This publication has 19 references indexed in Scilit:
- Connectionist sentence processing in perspectiveCognitive Science, 1999
- Toward a Connectionist Model of Recursion in Human Linguistic PerformanceCognitive Science, 1999
- Watching The Transients: Viewing a Simple Recurrent Network As a Limited CounterBehaviormetrika, 1999
- Dynamical recognizers: real-time language recognition by analog computersTheoretical Computer Science, 1998
- Analysis of Dynamical RecognizersNeural Computation, 1997
- On the Computational Power of Neural NetsJournal of Computer and System Sciences, 1995
- Tail-recursive Distributed Representations and Simple Recurrent NetworksConnection Science, 1995
- Finding structure in timeCognitive Science, 1990
- Finite State Automata and Simple Recurrent NetworksNeural Computation, 1989
- Inference of Reversible LanguagesJournal of the ACM, 1982