We use local input representation. A feedforward neural net cannot learn the problem, due to the arbitrary time lags that may occur. Instead, we need something like a recurrent net (e.g. [18] [37] [17] [40] [15] [39] [38] [20] [7] ).
A conventional fully recurrent net with
5 input units, 5 hidden units, 1 output unit, 1 ``bias unit'',
and a learning rate of 0.5,
needed less than 10000 training sequences to
come up with correct classifications in
100% of all test sequences.
(class 1: output activation
,
class 2: output activation
at the end of a sequence).
This sets the stage for the more difficult task described in the next subsection. It involves highly redundant sequences and will be harder to learn.