In this section we show how Contrastive Hebbian Learning (CHL) [Movellan, 1990] needs to be modified to accommodate units with relative phase angles. We follow the derivation in Movellan closely.
Movellan defines a continuous Hopfield Energy function
where E reflects the constraints imposed by the weights in the network and S the tendency to drive the activations to a resting value. For our network S is the same as for a network with no phase angles:
where n is the number of units in the network, is the activation of unit i, is the activation function for unit i, and .
However, E becomes
where is the weight connecting units i and j and is the coupling function associated with units i and j. In what follows we will abbreviate as .
The coupling function must be differentiable and satisfy the following:
When the network is stable, the inverse of the activation function for each unit is equal to the input into that unit:
where ( ) represents equilibrium and is the input to unit i. Furthermore, when the network is stable, the phase angle of each unit no longer changes:
Movellan defines the contrastive function J as
and shows that the CHL rule minimizes J. We follow his derivation for the case where units have phase angles.
The energy of the network E at equilibrium is
Extracting the terms with a term,
Differentiating with respect to a single weight and considering that is the only weight depending on ,
From Equation 10, we have
and
Substituting these into Equation 16,
From 11 and 12, we have the following for the case where . Since there are no self-recurrent connections in our network, we need only consider this case.
From 12, the last term is 0, and we have
From Equation 7,
and from Equation 11, we have
making
which shows that the modified CHL rule
descends in the J function.