next up previous contents index
Next: Update Functions Up: Working with Partial Previous: The Initialization Function

Learning Functions

By deleting all recurrent links in a partial recurrent network, a simple feedforward network remains. The context units have now the function of input units, i.e. the total network input consists of two components. The first component is the pattern vector, which was the only input to the partial recurrent network. The second component is a state vector. This state vector is given through the next--state function in every step. By this way the behavior of a partial recurrent network can be simulated with a simple feedforward network, that receives the state not implicitly through recurrent links, but as an explicit part of the input vector. In this sense, backpropagation algorithms can easily be modified for the training of partial recurrent networks in the following way:

  1. Initialization of the context units. In the following steps, all recurrent links are assumed to be not existent, except in step 2(f).
  2. Execute for each pattern of the training sequence the following steps:
  3. Only off--line training: weight adaption

In this manner, the following learning functions have been adapted for the training of partial recurrent networks like Jordan and Elman networks:

The parameters for these learning functions are the same as for the regular feedforward versions of these algorithms (see section gif).

For training a network with one of these functions a method called teacher forcing is used. That means that during the training phase the output units propagate the teaching output instead of their produced output to successor units (if there are any).



next up previous contents index
Next: Update Functions Up: Working with Partial Previous: The Initialization Function



Niels Mache
Wed May 17 11:23:58 MET DST 1995