Relative entropy and learning rules

Abstract
The dynamics of a probabilistic neural network is characterized by the distribution ν(x’‖x) of successor states x’ of an arbitrary state x of the network. A prescribed memory or behavior pattern is represented in terms of an ordered sequence of network states x(1),x(2),...,x(l). A successful procedure for learning this pattern must modify the neuronal interactions in such a way that the dynamical successor of x(s) is likely to be x(s+1), with x(l+1)=x(1). The relative entropy G of the probability distribution δx(s+1),x’ concentrated at the desired successor state, evaluated with respect to the dynamical distribution ν(x’‖x(s)), is used to quantify this criterion, by providing a measure of the distance between actual and ideal probability distributions. Minimization of G subject to appropriate resource constraints leads to ‘‘optimal’’ learning rules for pairwise and higher-order neuronal interactions. The degree to which optimality is approached by simple learning rules in current use is considered, and it is found, in particular, that the algorithm adopted in the Hopfield model is more effective in minimizing G than the original Hebb law.