Unnamed: 0
int64
0
7.24k
id
int64
1
7.28k
raw_text
stringlengths
9
124k
vw_text
stringlengths
12
15k
7,100
869
Two Iterative Algorithms for Computing the Singular Value Decomposition from Input / Output Samples Terence D. Sanger Jet Propulsion Laboratory MS 303-310 4800 Oak Grove Drive Pasadena, CA 91109 Abstract The Singular Value Decomposition (SVD) is an important tool for linear algebra and can be used to invert or approximate matrices. Although many authors use "SVD" synonymously with "Eigenvector Decomposition" or "Principal Components Transform", it is important to realize that these other methods apply only to symmetric matrices, while the SVD can be applied to arbitrary nonsquare matrices. This property is important for applications to signal transmission and control. I propose two new algorithms for iterative computation of the SVD given only sample inputs and outputs from a matrix. Although there currently exist many algorithms for Eigenvector Decomposition (Sanger 1989, for example), these are the first true samplebased SVD algorithms. 1 INTRODUCTION The Singular Value Decomposition (SVD) is a method for writing an arbitrary nons quare matrix as the product of two orthogonal matrices and a diagonal matrix. This technique is an important component of methods for approximating nearsingular matrices and computing pseudo-inverses. Several efficient techniques exist for finding the SVD of a known matrix (Golub and Van Loan 1983, for example). 144 Singular Value Decomposition p r- --------------------------- U __--I~ s 1I _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ __ I Figure 1: Representation of the plant matrix P as a linear system mapping inputs u into outputs y. LT SR is the singular value decomposition of P. However, for certain signal processing or control tasks, we might wish to find the SVD of an unknown matrix for which only input-output samples are available. For example, if we want to model a linear transmission channel with unknown properties, it would be useful to be able to approximate the SVD based on samples of the inputs and outputs of the channel. If the channel is time-varying, an iterative algorithm for approximating the SVD might be able to track slow variations. 2 THE SINGULAR VALUE DECOMPOSITION The SVD of a nonsymmetric matrix P is given by P = LT SR where Land Rare matrices with orthogonal rows containing the left and right "singular vectors" , and S is a diagonal matrix of "singular values". The inverse of P can be computed by inverting S, and approximations to P can be formed by setting the values of the smallest elements of S to zero. For a memoryless linear system with inputs u and outputs y = Pu, we can write y LT SRu which shows that R gives the "input" transformation from inputs to internal "modes", S gives the gain of the modes, and LT gives the "output" transformation which determines the effect of each mode on the output. Figure 1 shows a representation of this arrangement. = The goal of the two algorithms presented below is to train two linear neural networks Nand G to find the SVD of P. In particular, the networks attempt to invert P by finding orthogonal matrices Nand G such that NG ~ p- 1 , or P NG = I. A particular advantage of using the iterative algorithms described below is that it is possible to extract only the singular vectors associated with the largest singular values. Figure 2 depicts this situation, in which the matrix S is shown smaller to indicate a small number of significant singular values. There is a close relationship with algorithms that find the eigenvalues of a symmetric matrix, since any such algorithm can be applied to P pT = LT S2 Land pT P = RT S2 R in order to find the left and right singular vectors. But in a behaving animal or operating robotic system it is generally not possible to compute the product with pT, since the plant is an unknown component of the system. In the following, I will present two new iterative algorithms for finding the singular value decomposition 145 146 Sanger p , - - -- -- -- -- -' I I I I I I__ _ _ _ __ _ _ _ _ _ J y u N G Figure 2: Loop structure of the singular value decomposition for control. The plant is P = LT SR, where R determines the mapping from control variables to system modes, and LT determines the outputs produced by each mode . The optimal sensory network is G L, and the optimal motor network is N RT S-l. Rand L are shown as trapezoids to indicate that the number of nonzero elements of S (the "modes") may be less than the number of sensory variables y or motor variables u. = = of a matrix P given only samples of the inputs u and outputs y. 3 THE DOUBLE GENERALIZED HEBBIAN ALGORITHM The first algorithm is the Double Generalized Hebbian Algorithm (DGHA), and it is described by the two coupled difference equations b-.G = l(zyT - LT[zzT]G) (1) b-.N T = l(zuT - LT[zzT]N T ) (2) where LT[ ] is an operator that sets the above diagonal elements of its matrix argument to zero, y Pu, z Gy, and I is a learning rate constant. = = Equation 1 is the Generalized Hebbian Algorithm (Sanger 1989) which finds the eigenvectors of the autocorrelation matrix of its inputs y. For random uncorrelated inputs u, the autocorrelation of y is E[yyT] LT S2 L, so equation 1 will cause G to converge to the matrix of left singular vectors L . Equation 2 is related to the Widrow-Hoff (1960) LMS rule for approximating u T from z, but it enforces orthogonality of the columns of N. It appears similar in form to equation 1, except that the intermediate variables z are computed from y rather than u. A graphical representation of the algorithm is given in figure 3. Equations 1 and 2 together RT S-l L is an cause N to converge to RT S-l , so that the combination N G approximation to the plant inverse. = = Theorem 1: (Sanger 1993) If y = Pu, z = Gy, and E[uuT ] = I, then equations 1 and 2 converge to the left and right singular vectors of P . Singular Value Decomposition u y p IGHA Figure 3: Graphic representation of the Double Generalized Hebbian Algorithm. G learns according to the usual G HA rule, while N learns using an orthogonalized form of the Widrow- Hoff LMS Rule. Proof: After convergence of equation 1, E[zzT] will be diagonal, so that E[LT[zzT]] E[zzT]. Consider the Widrow-Hoff LMS rule for approximating uT from z: ~NT = 'Y(zu T - zzT NT). = (3) After convergence of G, this will be equivalent to equation 2, and will converge to the same attractor. The stable points of 3 occur when E[uz T - NzzT] 0, for which N = RT 5- 1 = ? The convergence behavior of the Double Generalized Hebbian Algorithm is shown in figure 4. Results are measured by computing B GP N and determining whether B is diagonal using a score "L...I~) ...... b~. I) = ?= L . b?2 1 1 The reduction in ? is shown as a function of the number of (u, y) examples given to the network during training, and the curves in the figure represent the average over 100 training runs with different randomly-selected plant matrices P. Note that the Double Generalized Hebbian Algorithm may perform poorly in the presence of noise or uncontrollable modes. The sensory mapping G depends only on the outputs y, and not directly on the plant inputs u. So if the outputs include noise or autonomously varying uncontrollable modes, then the mapping G will respond to these modes. This is not a problem if most of the variance in the output is due the inputs u, since in that case the most significant output components will reflect the input variance transmitted through P. 4 THE ORTHOGONAL ASYMMETRIC ENCODER The second algorithm is the Orthogonal Asymmetric Encoder (OAE) which is described by the equations (4) 147 148 Sanger 0.7 Double Generalized Hebbian Algorithm 0.& 0.5 j 0.4 j I! is 0.3 0.2 0.1 ... ... .. . . . . . . .. . . . . . . . ~~~-~~~~~~~~~-~~~~~ Exomple Figure 4: Convergence of the Double Generalized Hebbian Algorithm averaged over 100 random choices of 3x3 or 10xlO matrices P. (5) where z = NT u. This algorithm uses a variant of the Backpropagation learning algorithm (Rumelhart et al. 1986). It is named for the "Encoder" problem in which a three-layer network is trained to approximate the identity mapping but is forced to use a narrow bottleneck layer. I define the "Asymmetric Encoder Problem" as the case in which a mapping other than the identity is to be learned while the data is passed through a bottleneck. The "Orthogonal Asymmetric Encoder" (OAE) is the special case in which the hidden units are forced to be uncorrelated over the data set. Figure 5 gives a graphical depiction of the algorithm. Theorem 2: (Sanger 1993) Equations vectors of P. 4 and 5 converge to the left and right singular Proof: Suppose z has dimension m. If P = LT SR where the elements of S are distinct, and E[uu T ] I, then a well-known property of the singular value decomposition (Golub and Van Loan 1983, , for example) shows that = E[IIPu - C T NT ullJ (6) is minimized when = LrnU, NT = V Rm , and U and V are any m x m matrices 1mS/;;". (L~ and Rm signify the matrices of only the first m for which UV columns of LT or rows of R.) If we want E[zzT] to be diagonal, then U and V must be diagonal. OAE accomplishes this by training the first hidden unit as if m = 1, the second as if m = 2, and so on. = = CT For the case m 1, the error 6 is minimized when C is the first left singular vector of P and N is the first right singular vector. Since this is a linear approximation problem, there is a single global minimum to the error surface 6, and gradient descent using the backpropagation algorithm will converge to this solution. Singular Value Decomposition u y p I , ~as!pr.2Pa~ti~ I Figure 5: The Orthogonal Asymmetric Encoder algorithm computes a forward approximation to the plant P through a bottleneck layer of hidden units. After convergence, the remaining error is E[II(P - GT N T plant matrix as )ull1. If we decompose the i=l where Ii and ri are the rows of Land R, and the remaining error is Si are the diagonal elements of S, then n P2 = LlisirT i=2 which is equivalent to the original plant matrix with the first singular value set to zero. If we train the second hidden unit using P2 instead of P, then minimization of E[IIP2 u - GT NT ull1 will yield the second left and right singular vectors. Proceeding in this way we can obtain the first m singular vectors. Combining the update rules for all the singular vectors so that they learn in parallel leads to the governing equations of the OAE algorithm which can be written in matrix form as equations 4 and 5 . ? (Bannour and Azimi-Sadjadi 1993) proposed a similar technique for the symmetric encoder problem in which each eigenvector is learned to convergence and then subtracted from the data before learning the succeeding one. The orthogonal asymmetric encoder is different because all the components learn simultaneously. After convergence, we must multiply the learned N by S-2 in order to compute the plant inverse. Figure 6 shows the performance of the algorithm averaged over 100 random choices of matrix P. Consider the case in which there may be noise in the measured outputs y. Since the Orthogonal Asymmetric Encoder algorithm learns to approximate the forward plant transformation from u to y, it will only be able to predict the components of y which are related to the inputs u. In other words, the best approximation to y based on u is if ~ Pu, and this ignores the noise term. Figure 7 shows the results of additive noise with an SNR of 1.0. 149 150 Sanger 0 .7 Orthogonal Asymmetric Encoder 0.8 " 0.5 I!! c? 0.4 c j ii c 0.3 0 GO .l!I 0 0.2 , 0.1 .. . . . ..... . . . . . ..... 0 0 50 100 150 200 250 300 350 400 450 500 550 800 850 700 750 800 850 900 950 Example Figure 6: Convergence of the Orthogonal Asymmetric Encoder averaged over 100 random choices of 3x3 or 10xlO matrices P. Acknowledgements This report describes research done within the laboratory of Dr. Emilio Bizzi in the department of Brain and Cognitive Sciences at MIT. The author was supported during this work by a National Defense Science and Engineering Graduate Fellowship, and by NIH grants 5R37 AR26710 and 5ROINS09343 to Dr. Bizzi. References Bannour S., Azimi-Sadjadi M. R., 1993, Principal component extraction using recursive least squares learning, submitted to IEEE Transactions on Neural Networks. Golub G. H., Van Loan C. F., 1983, Matrix Computations, North Oxford Academic P., Oxford. Rumelhart D. E., Hinton G. E., Williams R. J., 1986, Learning internal representations by error propagation, In Parallel Distributed Processing, chapter 8, pages 318-362, MIT Press, Cambridge, MA. Sanger T. D., 1989, Optimal unsupervised learning in a single-layer linear feedforward neural network, Neural Networks, 2:459-473. Sanger T. D., 1993, Theoretical Elements of Hierarchical Control in Vertebrate Motor Systems, PhD thesis, MIT. Widrow B., Hoff M. E., 1960, Adaptive switching circuits, In IRE WESCON Conv. Record, Part 4, pages 96-104. Singular Value Decomposition 2 OAE with 50% Added Noise 1? ? 3x3 101110 ! .? j 1 j I CI ... ... . . . . . . ... . . . oL-.---,--,--~~~:;:::::;:::::;~~ o 50 100 150 200 250 300 350 400 450 500 650 &00 860 700 750 &00 860 000 860 ElIIlmple Figure 7: Convergence of the Orthogonal Asymmetric Encoder with 50% additive noise on the outputs, averaged over 100 random choices of 3x3 or 10xlO matrices P. 151
869 |@word effect:1 approximating:4 true:1 indicate:2 oae:5 uu:1 zyt:1 arrangement:1 symmetric:3 laboratory:2 memoryless:1 nonzero:1 added:1 rt:5 usual:1 decomposition:14 diagonal:8 during:2 gradient:1 reduction:1 m:2 generalized:8 propulsion:1 score:1 uncontrollable:2 decompose:1 nt:6 relationship:1 trapezoid:1 si:1 must:2 written:1 mapping:6 realize:1 predict:1 additive:2 lm:3 nih:1 motor:3 smallest:1 bizzi:2 succeeding:1 update:1 nonsymmetric:1 selected:1 unknown:3 currently:1 significant:2 perform:1 cambridge:1 largest:1 uv:1 record:1 tool:1 descent:1 minimization:1 situation:1 ire:1 mit:3 hinton:1 oak:1 rather:1 stable:1 arbitrary:2 behaving:1 operating:1 varying:2 pu:4 depiction:1 surface:1 gt:2 inverting:1 autocorrelation:2 certain:1 learned:3 narrow:1 behavior:1 uz:1 brain:1 ol:1 transmitted:1 minimum:1 able:3 below:2 p2:2 nand:2 pasadena:1 hidden:4 accomplishes:1 converge:6 vertebrate:1 conv:1 signal:2 ii:3 circuit:1 emilio:1 hebbian:8 jet:1 animal:1 eigenvector:3 special:1 academic:1 hoff:4 finding:3 transformation:3 extraction:1 ng:2 pseudo:1 variant:1 ti:1 unsupervised:1 synonymously:1 extract:1 minimized:2 rm:2 report:1 represent:1 control:5 unit:4 grant:1 invert:2 randomly:1 acknowledgement:1 simultaneously:1 before:1 national:1 engineering:1 want:2 signify:1 fellowship:1 singular:27 determining:1 switching:1 plant:11 attractor:1 oxford:2 attempt:1 sr:4 might:2 multiply:1 golub:3 presence:1 uncorrelated:2 intermediate:1 graduate:1 feedforward:1 averaged:4 land:3 row:3 supported:1 enforces:1 grove:1 recursive:1 x3:4 backpropagation:2 orthogonal:12 roins09343:1 bottleneck:3 whether:1 van:3 orthogonalized:1 defense:1 theoretical:1 passed:1 distributed:1 word:1 curve:1 column:2 dimension:1 computes:1 sensory:3 author:2 forward:2 ignores:1 close:1 cause:2 operator:1 adaptive:1 transaction:1 writing:1 useful:1 generally:1 rare:1 equivalent:2 snr:1 eigenvectors:1 approximate:4 go:1 williams:1 graphic:1 global:1 robotic:1 wescon:1 exist:2 rule:5 iterative:5 track:1 channel:3 learn:2 write:1 variation:1 terence:1 together:1 coupled:1 ca:1 pt:3 zzt:7 suppose:1 thesis:1 reflect:1 i__:1 dgha:1 us:1 containing:1 pa:1 element:6 rumelhart:2 dr:2 cognitive:1 s2:3 noise:7 asymmetric:10 run:1 inverse:4 gy:2 respond:1 north:1 named:1 depicts:1 slow:1 autonomously:1 r37:1 depends:1 wish:1 azimi:2 layer:4 ct:1 parallel:2 learns:3 theorem:2 trained:1 formed:1 algebra:1 square:1 occur:1 variance:2 orthogonality:1 zu:1 ri:1 yield:1 uut:1 ar26710:1 argument:1 chapter:1 produced:1 ci:1 phd:1 train:2 forced:2 distinct:1 drive:1 department:1 submitted:1 according:1 combination:1 lt:14 smaller:1 describes:1 encoder:12 xlo:3 associated:1 proof:2 gp:1 transform:1 nonsquare:1 gain:1 pr:1 samplebased:1 determines:3 advantage:1 eigenvalue:1 equation:13 ut:1 ma:1 goal:1 propose:1 yyt:1 identity:2 product:2 appears:1 loop:1 combining:1 available:1 loan:3 except:1 poorly:1 apply:1 rand:1 hierarchical:1 principal:2 done:1 governing:1 svd:12 subtracted:1 sru:1 convergence:9 double:7 transmission:2 original:1 internal:2 remaining:2 include:1 propagation:1 graphical:2 widrow:4 mode:9 sanger:10 measured:2 quare:1
7,101
87
233 HIGH ORDER NEURAL NETWORKS FOR EFFICIENT ASSOCIATIVE MEMORY DESIGN I. GUYON?, L. PERSONNAZ?, J. P. NADAL?? and G. DREYFUS? ? Ecole Superieure de Physique et de Chimie Industrielles de la Ville de Paris Laboratoire d'Electronique 10, rue Vauquelin 75005 Paris (France) ?? Ecole Normale Superieure Groupe de Physique des Solides 24, rue Lhomond 75005 Paris (France) ABSTRACT We propose learning rules for recurrent neural networks with high-order interactions between some or all neurons. The designed networks exhibit the desired associative memory function: perfect storage and retrieval of pieces of information and/or sequences of information of any complexity. INTRODUCTION In the field of information processing, an important class of potential applications of neural networks arises from their ability to perform as associative memories. Since the publication of J. Hopfield's seminal paper 1, investigations of the storage and retrieval properties of recurrent networks have led to a deep understanding of their properties. The basic limitations of these networks are the following: - their storage capacity is of the order of the number of neurons; - they are unable to handle structured problems; - they are unable to classify non-linearly separable data. ? American Institute of Physics 1988 234 In order to circumvent these limitations, one has to introduce additional non-linearities. This can be done either by using "hidden", non-linear units, or by considering multi-neuron interactions2 . This paper presents learning rules for networks with multiple interactions, allowing the storage and retrieval, either of static pieces of information (autoassociative memory), or of temporal sequences (associative memory), while preventing an explosive growth of the number of synaptic coefficients. AUTOASSOCIATIVE MEMORY The problem that will be addressed in this paragraph is how to design an autoassociative memory with a recurrent (or feedback) neural network when the number p of prototypes is large as compared to the number n of neurons. We consider a network of n binary neurons, operating in a synchronous mode, with period t. The state of neuron i at time t is denoted by (Ji(t), and the state of the network at time t is represented by a vector ~(t) whose components are the (Ji(t). The dynamics of each neuron is governed by the following relation: (Ji(t+t) = sgn vi(t). (1 ) In networks with two-neuron interactions only, the potential vi(t) is a linear function of the state of the network: For autoassociative memory design, it has been shown 3 that any set of correlated patterns, up to a number of patterns p equal to 2 n, can be made the stable states of the system, provided the synaptic matrix is computed as the orthogonal projection matrix onto the subspace spanned by the stored vectors. However, as p increases, the rank of the family of prototype vectors will increase, and finally reach the value of n. In such a case, the synaptic matrix reduces to the identity matrix, so that all 2 n states are stable and the energy landscape becomes flat. Even if such an extreme case is avoided, the attractivity of the stored states decreases with increasing p, or, in other terms, 235 the number of fixed points which are not the stored patterns increases; this problem can be alleviated to a large extent by making a useful use of these "spurious" fixed points4. Another possible solution consists in "gardening" the state space in order to enlarge the basins of attraction of the fixed points5. Anyway, no dramatic improvements are provided by all these solutions since the storage capacity is always O(n). We now show that the introduction of high-order interactions between neurons, increases the storage capacity proportionally to the number of connections per neuron. The dynamical behaviour of neuron i is still governed by (1). We consider two and three-neuron interactions, extension to higher order are straightforward. The potential vi (t) is now defi ned as It is more convenient, for the derivation of learning rules, to write the potential in the matrix form: ~(t) = C ;t(t), where :?(t) is an m dimensional vector whose components are taken among the set of the (n 2+n)/2 values: a1 , ... , an' a1 a2 , ... , aj al ' ... , a n-1 an. As in the case of the two-neuron interactions model, we want to compute the interaction coefficients so that the prototypes are stable and attractor states. A condition to store a set of states Q:k (k=1 to p) is that y'k= Q:k for all k. Among the solutions, the most convenient solution is given by the (n,m) matrix c=I,r l (2) where I, is the (n,p) matrix whose columns are the Q:k and rl is the (p,m) pseudoinverse of the (m,p) matrix r whose columns are the { . This solution satisfies the above requirements, up to a storage capacity which is related to the dimension m of vectors :?. Thus, in a network with three-neuron 236 interactions, the number of patterns that can be stored is O(n 2). Details on these derivations are published in Ref.6. By using only a subset of the products {aj all. the increase in the number of synaptic coefficients can remain within acceptable limits, while the attractivity of the stored patterns is enhanced, even though their number exceeds the number of neurons ; this will be examplified in the simulations presented below. Finally, it can be noticed that, if vector ~contains all the {ai aj}' i=1, ... n, j=1, ... n, only, the computation of the vector potential ~=C~can be performed after the following expression: where ~ stands for the operation which consists in squaring all the matrix coefficients. Hence, the computation of the synaptic coefficients is avoided, memory and computing time are saved if the simulations are performed on a conventional computer. This formulation is also meaningful for optical implementations, the function ell being easily performed in optics 7. In order to illustrate the capabilities of the learning rule, we have performed numerical simulations which show the increase of the size of the basins of attraction when second-order interactions, in addition to the first-order ones, are used. The simulations were carried out as follows. The number of neurons n being fixed, the amount of second-order interactions was chosen ; p prototype patterns were picked randomly, their components being ?1 with probability 0.5 ; the second-order interactions were chosen randomly. The synaptic matrix was computed from relation (2). The neural network was forced into an initial state lying at an initial Hamming distance Hi from one of the prototypes {!k ; it was subsequently left to evolve until it reached a stable state at a distance Hf from {!k. This procedure was repeated many times for each prototype and the Hf were averaged over all the tests and all the prototypes. Figures 1a. and 1b. are charts of the mean values of Hf as a function of the number of prototypes, for n = 30 and for various values of m (the dimension of 237 vector ':/.). These curves allowed us to determine the maximum number of prototype states which can be stored for a given quality of recall. Perfect recall implies Hf =0 ; when the number of prototypes increases, the error in recall may reach Hf =H i : the associative memory is degenerate. The results obtained for Hi In =10% are plotted on Figure 1a. When no high-order interactions were used, Hf reached Hi for pIn = 1, as expected; conversely, virtually no error in recall occured up to pIn = 2 when all second-order interactions were taken into account (m=465). Figure 1b shows the same quantities for Hi=20 0/0 ; since the initial states were more distant from the prototypes, the errors in recall were more severe. 1.2 1.2 1.0 1.0 0.8 0.8 0.6 :f A 0.6 :f A f f v (8) 0.4 v 0.2 0.2 0.0 0.0 2 0 3 (b) 0.4 2 0 3 pIn pIn Fig. 1. Improvement of the attractivity by addition of three-neuron interactions to the two-neuron interactions. All prototypes are always stored exactly (all curves go through the origin). Each point corresponds to an average over min(p,10) prototypes and 30 tests for each prototype. [] Projection: m = n = 30; ? 1 a: Hi I n =10 % ; m = 120 ; ? m = 180; 0 m = 465 (all interactions) 1 b : Hi In =20%. TEMPORAL SEQUENCES (ASSOCIATIVE MEMORY) The previous section was devoted to the storage and retrieval of items of information considered as fixed points of the dynamics of the network (autoassociative memory design). However, since fully connected neural networks are basically dynamical systems, they are natural candidates for 238 storing and retrieving information which is dynamical in nature, i.e., temporal sequences of patterns8: In this section, we propose a general solution to the problem of storing and retrieving sequences of arbitrary complexity, in recurrent networks with parallel dynamics. Sequences consist in sets of transitions between states {lk_> Q:k+ 1, k=1, ... , p. A sufficient condition to store these sets of transitions is that y..~ Q:k+ 1 for all k. In the case of a linear potential y"=C Q:, the storage prescription proposed in ref.3 can be used: C=r,+r,I, where r, is a matrix whose columns are the Q:k and r,+ is the matrix whose columns are the successors Q:k+ 1 of Q:k. If P is larger than n, one can use high-order interactions, which leads to introduce a non-linear potential Y..=C ';f. , with ';f. as previously defined. We proposed in ref. 10 the following storage prescription : (3) The two above prescriptions are only valid for storing simple sequences, where no patterns occur twice (or more). Suppose that one pattern occurs twice; when the network reaches this bifurcation point, it is unable to make a decision according the deterministic dynamics described in (1), since the knowledge of the present state is not sufficient. Thus, complex sequences require to keep, at each time step of the dynamics, a non-zero memory span. The vector potential Y..=C':J. must involve the states at time t and t-t, which leads to define the vector ';f. as a concatenation of vectors Q:(t), ~(t-t), Q:(t)?Q:(t), Q:(t)?Q:(t-t), or a suitable subset thereof. The subsequent vector Q:(t+t) is still determined by relation (1). In this form, the problem is a generalization of the storage of patterns with high order interactions, as described above. The storage of sequences can be still processed by relation (3). The solution presented above has the following features: i) Sequences with bifurcation points can be stored and retrieved. ii) The dimension of the synaptic matrix is at most (n,2(n 2+n)), and at least (n,2n) in the linear case, so that at most 2n(n2+n) and at least 2n2 synapses are required. 239 iii) The storage capacity is O(m), where m is the dimension of the vector ';t . iv) Retrieval of a sequence requires initializing the network with two states in succession. The example of Figure 2 illustrates the retrieval performances of the latter learning rule. We have limited vector ';t to Q:(t}?Q:(t-t). In a network of n=48 neurons, a large number of poems have been stored, with a total of p=424 elementary transitions. Each state is consists in the 6 bit codes of 8 letters. ALOUETTE JETE PLUMERAI ALOUETTE GENTILLE ALOUETTE ALOUETTE JETE PLUMERAI JE NE OLVMERAI AQFUETTE JEHKILLE SLOUETTE ALOUETTE JETE PLUMERAI Fig. 2. One of the stored poems is shown in the first column. The network is initialized with two states (the first two lines of the second column). After a few steps, the network reaches the nearest stored sequence. LOCAL LEARNING Finally, it should be mentioned that all the synaptic matrices introduced in this paper can be computed by iterative, local learning rules. For autoassociative memory, it has been shown analytically9 that the procedure: with Cij(O) = 0, which is a Widrow-Hoff type learning rule, yields the projection matrix, when 240 the number of presentations of the prototypes {~k} goes to infinity, if the latter are linearly independent. A derivation along the same lines shows that, by repeated presentations of the prototype transitions, the learning rules: lead to the exact solutions (relations (2) and (3) respectively), if the vectors }< are Ii nearly independent. GENERALIZATION TASKS Apart from storing and retrieving static pieces of information or sequences, neural networks can be used to solve problems in which there exists a structure or regularity in the sample patterns (for example presence of clumps, parity, symmetry ... ) that the network must discover. Feed-forward networks with multiple layers of first-order neurons can be trained with back-propagation algorithms for these purposes; however, one-layer feed-forward networks with mUlti-neuron interactions provide an interesting alternative. For instance, a proper choice of vector ':I. (second-order terms only) with the above learning rule yields a perfectly straightforward solution to the exclusive-OR problem. Maxwell et al. have shown that a suitable high-order neuron is able to exhibit the "ad hoc network solution" for the contiguity problem 11. CONCLUSION The use of neural networks with high-order interactions has long been advocated as a natural way to overcome the various limitations of the Hopfield model. However, no procedure guaranteed to store any set of information as fixed points or as temporal sequences had been proposed. The purpose of the present paper is to present briefly such storage prescriptions and show 241 some illustrations of the use of these methods. Full derivations and extensions will be published in more detailed papers. REFERENCES 1. 2. J. J. Hopfield, Proc. Natl. Acad. Sci. (USA) la, 2554 (1982). P. Peretto and J. J. Niez, BioI. Cybern. M, 53 (1986). P. Baldi and S. S. Venkatesh, Phys. Rev. Lett . .5.6 , 913 (1987). For more references see ref.6. 3. L. Personnaz, I. Guyon, G. Dreyfus, J. Phys. Lett. ~ , 359 (1985). L. Personnaz, I. Guyon, G. Dreyfus, Phys. Rev. A ~ , 4217 (1986). 4. I. Guyon, L. Personnaz, G. Dreyfus, in "Neural Computers", R. Eckmiller and C. von der Malsburg eds (Springer, 1988). 5. E. Gardner, Europhys. Lett . .1, 481 (1987). G. Poppel and U.Krey, Europhys. Lett.,.1, 979 (1987). 6. L. Personnaz, I. Guyon, G. Dreyfus, Europhys. Lett. .1,863 (1987). 7. D. Psaltis and C. H. Park, in "Neural Networks for Computing", J. S. Denker ed., (A.I.P. Conference Proceedings 151, 1986). 8. P. Peretto, J. J. Niez, in "Disordered Systems and Biological Organization", E. Bienenstock, F. Fogelman, G. Weisbush eds (Springer, Berlin 1986). S. Dehaene, J. P. Changeux, J. P. Nadal, PNAS (USA)~, 2727 (1987). D. Kleinfeld, H. Sompolinsky, preprint 1987. J. Keeler, to appear in J. Cog. Sci. For more references see ref. 9. 9. I. Guyon, L. Personnaz, J.P. Nadal and G. Dreyfus, submitted for publication. 10. S. Diederich, M. Opper, Phys. Rev. Lett . .5.6, 949 (1987). 11. T. Maxwell, C. Lee Giles, Y. C. Lee, Proceedings of ICNN-87, San Diego, 1987.
87 |@word briefly:1 lett:6 simulation:4 dramatic:1 initial:3 contains:1 ecole:2 must:2 distant:1 subsequent:1 numerical:1 designed:1 item:1 along:1 retrieving:3 consists:3 paragraph:1 baldi:1 introduce:2 expected:1 multi:2 considering:1 increasing:1 becomes:1 provided:2 discover:1 linearity:1 nadal:3 contiguity:1 temporal:4 growth:1 exactly:1 unit:1 appear:1 local:2 limit:1 acad:1 twice:2 conversely:1 limited:1 clump:1 averaged:1 procedure:3 projection:3 alleviated:1 convenient:2 onto:1 storage:14 cybern:1 seminal:1 conventional:1 deterministic:1 straightforward:2 go:2 rule:9 attraction:2 spanned:1 handle:1 anyway:1 enhanced:1 suppose:1 diego:1 exact:1 origin:1 defi:1 preprint:1 initializing:1 connected:1 sompolinsky:1 decrease:1 mentioned:1 complexity:2 dynamic:5 trained:1 easily:1 hopfield:3 represented:1 various:2 derivation:4 forced:1 europhys:3 whose:6 larger:1 solve:1 ability:1 associative:6 hoc:1 sequence:14 propose:2 interaction:20 product:1 degenerate:1 regularity:1 requirement:1 perfect:2 illustrate:1 recurrent:4 widrow:1 nearest:1 advocated:1 implies:1 saved:1 subsequently:1 disordered:1 sgn:1 successor:1 require:1 behaviour:1 generalization:2 icnn:1 investigation:1 biological:1 elementary:1 extension:2 keeler:1 lying:1 considered:1 a2:1 purpose:2 proc:1 psaltis:1 always:2 normale:1 publication:2 improvement:2 rank:1 squaring:1 spurious:1 hidden:1 relation:5 bienenstock:1 france:2 fogelman:1 among:2 denoted:1 ell:1 bifurcation:2 hoff:1 field:1 equal:1 enlarge:1 park:1 nearly:1 few:1 randomly:2 attractor:1 explosive:1 organization:1 severe:1 physique:2 extreme:1 natl:1 devoted:1 orthogonal:1 iv:1 initialized:1 desired:1 plotted:1 instance:1 classify:1 column:6 giles:1 subset:2 stored:11 lee:2 physic:1 von:1 american:1 account:1 potential:8 de:6 coefficient:5 vi:3 ad:1 piece:3 performed:4 picked:1 reached:2 hf:6 capability:1 parallel:1 chart:1 succession:1 yield:2 landscape:1 basically:1 poppel:1 published:2 submitted:1 synapsis:1 reach:4 phys:4 synaptic:8 ed:3 diederich:1 energy:1 thereof:1 static:2 hamming:1 recall:5 knowledge:1 occured:1 back:1 feed:2 maxwell:2 higher:1 formulation:1 done:1 though:1 until:1 propagation:1 kleinfeld:1 mode:1 quality:1 aj:3 usa:2 hence:1 poem:2 dreyfus:6 ji:3 rl:1 ai:1 had:1 stable:4 operating:1 retrieved:1 apart:1 store:3 binary:1 der:1 additional:1 attractivity:3 determine:1 period:1 ii:2 multiple:2 full:1 pnas:1 reduces:1 exceeds:1 long:1 retrieval:6 prescription:4 a1:2 basic:1 addition:2 want:1 addressed:1 laboratoire:1 virtually:1 dehaene:1 presence:1 iii:1 perfectly:1 prototype:16 synchronous:1 expression:1 autoassociative:6 deep:1 useful:1 proportionally:1 involve:1 detailed:1 amount:1 processed:1 vauquelin:1 per:1 write:1 eckmiller:1 ville:1 letter:1 family:1 guyon:6 decision:1 acceptable:1 bit:1 layer:2 hi:6 guaranteed:1 occur:1 optic:1 infinity:1 krey:1 flat:1 min:1 span:1 separable:1 optical:1 ned:1 structured:1 according:1 remain:1 rev:3 making:1 taken:2 previously:1 pin:4 operation:1 denker:1 alternative:1 malsburg:1 personnaz:6 noticed:1 quantity:1 occurs:1 exclusive:1 exhibit:2 subspace:1 distance:2 unable:3 sci:2 capacity:5 concatenation:1 berlin:1 extent:1 code:1 illustration:1 cij:1 design:4 implementation:1 proper:1 perform:1 allowing:1 neuron:22 peretto:2 arbitrary:1 introduced:1 venkatesh:1 paris:3 required:1 connection:1 able:1 dynamical:3 pattern:10 below:1 memory:14 suitable:2 natural:2 circumvent:1 ne:1 gardner:1 carried:1 understanding:1 evolve:1 fully:1 interesting:1 limitation:3 basin:2 sufficient:2 storing:4 parity:1 institute:1 niez:2 electronique:1 feedback:1 opper:1 dimension:4 curve:2 stand:1 transition:4 valid:1 overcome:1 preventing:1 forward:2 made:1 san:1 avoided:2 keep:1 pseudoinverse:1 iterative:1 nature:1 symmetry:1 complex:1 rue:2 linearly:2 n2:2 repeated:2 ref:5 allowed:1 fig:2 je:1 lhomond:1 candidate:1 governed:2 cog:1 changeux:1 consist:1 exists:1 illustrates:1 led:1 springer:2 corresponds:1 satisfies:1 bioi:1 identity:1 presentation:2 determined:1 total:1 la:2 meaningful:1 latter:2 arises:1 superieure:2 correlated:1
7,102
870
Robot Learning: Exploration and Continuous Domains David A. Cohn MIT Dept. of Brain and Cognitive Sciences Cambridge, MA 02139 The goal of this workshop was to discuss two major issues: efficient exploration of a learner's state space, and learning in continuous domains. The common themes that emerged in presentations and in discussion were the importance of choosing one's domain assumptions carefully, mixing controllers/strategies, avoidance of catastrophic failure, new approaches with difficulties with reinforcement learning, and the importance of task transfer. 1 Domain assumptions Andrew Moore (CMU) discussed the problem of standardizing and making explicit the set of assumptions that researcher makes about his/her domain. He suggested that neither "fewer assumptions are better" nor "more assumptions are better" is a tenable position, and that we should strive to find and use standard sets of assumptions. With no such commonality, comparison of techniques and results is meaningless. Under Moore's guidance, the group discussed the possibility of designing an algorithm which used a number of well-chosen assumption sets and switched between them according to their empirical validity. Suggestions were made to draw on the AI approach of truth maintenance systems. This theme of detecting failure of an assumption set/strategy was echoed in the discussion on mixing controllers and avoiding failure (described below). 2 Mixing controllers and strategies Consensus appeared to be against using single monolithic approaches, and in favor of mixing controllers. Spatial mixing resulted in local models, as advocated by Stefan Schaal (MIT) using locally weighted regression. Controllers could also be mixed over the entire domain. Jeff Schneider (Rochester) discussed mixing a na'ive feedback controller with a "coaching signal." Combining the coached controller with an un coached one further improved performance. During the main conference, Satinder Singh (MIT) described a controller that learned by reinforcement to mix the strategies of two "safe" but suboptimal controllers, thus avoiding unpleasant surprises and catastrophic failure. 1169 1170 Cohn 3 Avoiding failure The issue of det,ecting impending failure and avoiding catastrophic failure clarified differences in several approaches. A learning strategy that learns in few trials is useless on a real robot if the initial trials break the robot by crashing it into walls. Singh's approach has implicit failure avoidance, but may be hampered by an unnecessarily large margin of safety. Terry Sanger (JPL) discussed a trajectory extension algorithm by which a controller could smoothly "push the limits" of its performance, and detect impending failure of the control strategy. 4 Reinforcement learning Reinforcement learning seems to have come into its own, with people realizing the diverse ways in which it may be applied to problems. Long-Ji Lin (Siemens) described his group's unusual but successful application of reinforcement learning in landmark-based navigation. The Siemens RatBot uses reinforcement to select landmarks on the basis of their recognizability and their value to the eventual precision of position estimation. The reinforcement signal is simply the cost and the robot's final position error after it has used a set of landmarks. Jose del R. Millan (JRC) presented an approach similar to Schneider's, but training a neural controller with reinforcement learning. As the controller's performance improves, it supplants the mobile robot's reactive "instincts," which are designed to prevent catastrophic failure. With new applications, however, come new pitfalls. Leemon Baird (WPAFB) showed how standard reinforcement learning approaches can fail when adapted to exploration in continuous time. He then described the "advantage updating" algorithm which was designed to work in noisy domains with continuous or small time steps. The issue of exploration in continuous space, especially with noise, has not been as easily addressed. Jiirgen Schmidhuber (TUM) described the approach one should take if interested solely in exploration: use prior information gain as a reinforcement signal to decide on an "optimal" action. The ensuing discussion centered on the age-old and intractable tradeoff between exploration and exploitation. Final consensus was that we, as a group, should become more familiar with the literature on dual control, which addresses exactly this issue. 5 Task transfer An unorchestrated theme that emerged from the discussion was the need to address, or even define, task transfer. As with last year's workshop on Robot Learning, it was generally agreed that "one-task learning" is not a suitable goal when designing a learning robot. During the discussion, Long-Ji Lin (Siemens) and Lori Pratt (CSM) described several types of task transfer that are considered in the literature. These included model learning for multiple tasks, hierarchical control and learning, and concept (or bias) sharing across tasks.
870 |@word especially:1 validity:1 trial:2 exploitation:1 come:2 concept:1 seems:1 safe:1 moore:2 strategy:6 exploration:6 centered:1 during:2 ensuing:1 landmark:3 initial:1 wall:1 consensus:2 extension:1 useless:1 considered:1 common:1 echoed:1 major:1 ji:2 commonality:1 csm:1 designed:2 discussed:4 he:2 estimation:1 fewer:1 cambridge:1 ai:1 realizing:1 weighted:1 stefan:1 mit:3 detecting:1 clarified:1 robot:7 mobile:1 become:1 coaching:1 david:1 own:1 showed:1 schaal:1 schmidhuber:1 learned:1 detect:1 nor:1 address:2 brain:1 suggested:1 below:1 entire:1 pitfall:1 schneider:2 appeared:1 her:1 signal:3 interested:1 multiple:1 issue:4 mix:1 terry:1 dual:1 suitable:1 difficulty:1 spatial:1 long:2 lin:2 unnecessarily:1 regression:1 maintenance:1 controller:12 cmu:1 exactly:1 control:3 few:1 prior:1 literature:2 resulted:1 safety:1 monolithic:1 local:1 addressed:1 limit:1 crashing:1 mixed:1 suggestion:1 familiar:1 meaningless:1 solely:1 age:1 switched:1 possibility:1 navigation:1 pratt:1 last:1 suboptimal:1 bias:1 tradeoff:1 det:1 old:1 empirical:1 guidance:1 feedback:1 made:1 reinforcement:10 action:1 cost:1 generally:1 satinder:1 successful:1 locally:1 instinct:1 continuous:5 avoidance:2 un:1 impending:2 his:2 transfer:4 diverse:1 group:3 na:1 us:1 domain:7 designing:2 prevent:1 neither:1 main:1 cognitive:1 updating:1 tenable:1 strive:1 noise:1 year:1 standardizing:1 jose:1 baird:1 jrc:1 decide:1 precision:1 theme:3 position:3 explicit:1 draw:1 break:1 learns:1 rochester:1 singh:2 adapted:1 learner:1 basis:1 jpl:1 workshop:2 easily:1 intractable:1 ecting:1 importance:2 leemon:1 trajectory:1 researcher:1 push:1 margin:1 according:1 lori:1 surprise:1 smoothly:1 choosing:1 sharing:1 simply:1 across:1 emerged:2 failure:10 against:1 ive:1 making:1 jiirgen:1 favor:1 wpafb:1 noisy:1 gain:1 truth:1 final:2 ma:1 advantage:1 discus:1 goal:2 presentation:1 improves:1 fail:1 agreed:1 carefully:1 eventual:1 jeff:1 combining:1 unusual:1 tum:1 included:1 mixing:6 improved:1 recognizability:1 hierarchical:1 catastrophic:4 implicit:1 siemens:3 select:1 millan:1 unpleasant:1 cohn:2 hampered:1 people:1 del:1 andrew:1 reactive:1 sanger:1 dept:1 advocated:1 avoiding:4
7,103
871
Comparisoll Training for a Resclleduling Problem ill Neural Networks Didier Keymeulen Artificial Intelligence Laboratory Vrije Universiteit Brussel Pleinlaan 2, 1050 Brussels Belgium Martine de Gerlache Prog Laboratory Vrije Universiteit Brussel Pleinlaan 2, 1050 Brussels Belgium Abstract Airline companies usually schedule their flights and crews well in advance to optimize their crew pools activities. Many events such as flight delays or the absence of a member require the crew pool rescheduling team to change the initial schedule (rescheduling). In this paper, we show that the neural network comparison paradigm applied to the backgammon game by Tesauro (Tesauro and Sejnowski, 1989) can also be applied to the rescheduling problem of an aircrew pool. Indeed both problems correspond to choosing the best solut.ion from a set of possible ones without ranking them (called here best choice problem). The paper explains from a mathematical point of view the architecture and the learning strategy of the backpropagation neural network used for the best choice problem. We also show how the learning phase of the network can be accelerated. Finally we apply the neural network model to some real rescheduling problems for the Belgian Airline (Sabena). 1 Introduction Due to merges, reorganizations and the need for cost reduction, airline companies need to improve the efficiency of their manpower by optimizing the activities of their crew pools as much as possible. A st.andard scheduling of flights and crews is usually made well in advance but many events, such as flight delays or the absence of a crew member make many schedule cha.nges (rescheduling) necessary. 801 802 Keymeulen and de Gerlache Each day, the CPR1 team of an airline company has to deal with these perturbations. The problem is to provide the best answer to these regularly occurring perturbations and to limit their impact on the general schedule. Its solution is hard to find and usually the CPR team calls on full reserve crews. An efficient rescheduling tool taking into account the experiences of the CPR team could substantially reduce the costs involved in rescheduling notably by limit.ing the use of a reserve crew. The paper is organized as follow. In the second section we describe the rescheduling task. In the third section we argue for the use of a neural network for the rescheduling task and we apply an adequate architecture for such a network. Finally in the last section, we present results of experiments with schedules based on actual schedules used by Sabena. 2 Rescheduling for an Airline Crew Pool When a pilot is unavailable for a flight it becomes necessary to replace him, e.g. to reschedule the crew. The rescheduling starts from a list of potential substitute pilots (PSP) given by a scheduling program based generally on operation research or expert syst.em technology (Steels, 1990). The PSP list obtained respects legislation and security rules fixing for example t.he number of flying hours per month, the maximum number of consecutive working hour and the number of training hours per year and t.heir schedule. From the PSP list, the CPR team selects the best candidat.es taking into account t.he schedule stability and equity. The schedule stability requires that possible perturbations of the schedule can be dealt with with only a minimal rescheduling effort. This criterion ensures work stability t.o the crew members and has an important influence on their social behavior. The schedule equity ensures the equal dist.ribution of the work and payment among the crew members during the schedule period. One may think to solve this rescheduling problem in t.he same way as the scheduling problem itself using software t.ools based on operational research or expert system approach. But t.his is inefficient. for t.wo reasons, first., the scheduling issued from a scheduling system and its adapt.at.ion t.o obt.ain an acceptable schedule takes days. Second this system does not t.ake into account the previous schedule. It follows that the updat.ed one may differ significantly from the previous one after each perturbation. This is unaccept.able from a pilot's point of view. Hence a specific procedure for rescheduling is necessary. 3 Neural Network Approach The problem of reassigning a new crew member to replace a missing member can be seen as the problem of finding the best pilot in a pool of potential substitute pilots (PSP), called the best choice problem. To solve the best choice problem, we choose the neural network approach for two reasons. First the rules llsed by the expert. are not well defined: to find the best PSP, lCrew Pool Rescheduler Comparison Training for a Rescheduling Problem in Neural Networks the expert associates implicit.ly a score value to each profile. The learning approach is precisely well suited to integrate, in a short period of time, t.he expert knowledge given in an implicit form. Second, t.he neural network approach was applied with success to board-games e.g. the Backgammon game described by Tesauro (Tesauro and Sejnowski, 1989) and the Nine l\ll en's Morris game described by Braun (Braun and al., 1991). These two games are also exa.mples of best choice problem where the player chooses the best move from a set of possible ones. 3.1 Profile of a Potential Substitute Pilot To be able to use the neural network approach we have to identify the main features of the potential substitute pilot and to codify them in terms of rating values (de Gerlache and Keymeulen, 1993). We based our coding scheme on the way the expert solves a rescheduling problem. He ident.ifies the relevant parameters associated with the PSP and the perturbed schedule. These parameters give three types of information. A first type describes the previous, present and future occupation of the PSP. The second t.ype represents information not in the schedule such as the human relationship fadars. The assocjat.ed values of t.hese two t.ypes of parameters differ for f'ach PSP. The last t.ype of paramet.ers describes the context of the rescheduling, namely t.he characteristics of t.he schedule . This last type of parameters are the same for all the PSP. All t.hese paramet.ers form the profile of a PSP associated to a perturbed schedule. At each rescheduling problem corresponds one perturbed schedule j and a group of 11 PSpi to which we associate a Projile~ (PSpi, PertU1?berLSchedulej) . Implicitly, the expert associates a rating value between 0 and 1 to each parameter of the P1'ojile; based on respectively its little or important impact on the result.ing schedule if the P S pi was chosen. The rating value reflects the relative importance of the parameters on the stability and the equity of the resulting schf'dnle obt.ained after the pilots substitution. = 3.2 Dual Neural Network It would have been possible to get more information from the expert than only the best profile. One of the possibilities is to ask him to score every profile associated with a perturbed planning. From this associat.ion we could immediately construct a scoring function which couples each profile with a specific value, namely its score. Another possibility is to ask the expert to rank all profiles associated with a perturbed schedule. The corresponding ranking function couples each profile with a value such that the values associat.ed with the profiles of the same perturbed schedule order the profiles according t.o t.heir rank. The decision making process used by the rescheduler team for the aircrew rescheduling problem does not consist in the evaluation of a scoring or ranking function . Indeed only the knowledge of the best profile is useful for the rescheduling process. From a neura.l net.work architectural point of view, because the ranking problem is a generalization of the best choice problem, a same neural net.work architecture can be used. But the difference between the best choice problem and t.he scoring problem is such that two different neural network architectures are associated to them. As we show in this section, although a backpropagatian network is sufficient to learn a scoring function, its architecture, its learning and its retrieval procedures must be 803 804 Keymeulen and de Gerlache adapted to learn the best profile. Through a mathematical formulation of the best choice problem, we show that the comparison paradigm of Tesauro (Tesauro, 1989) is suited to the best choice problem and we suggest how to improve the learning convergence. 3.2.1 Comparing Function For the best choice problem the expert gives the best profile Projilefest associated with the perturbed schedule j and that for m pert.urbed schedules. The problem consists then to learn the mapping of the m * n profiles associated with the m perturbed schedules into the m best profiles, one for each pert.urbed schedule. One way to represent this association is through a comparing function. This function has as input a profile, represented by a vector xj, and returns a single value. When a set of profiles associated with a perturbed schedule are evaluated by the function, it returns the lowest value for the best profile. This comparing function integrates the information given by the expert and is sufficient to reschedule any perturbed schedule solved in the past by the expert. Formally it is defined by: Comp(J.1>e) = C(Projile)) (1) C ompare jBest < C ompcl1>C ,j. {V)' Vi=fBest with)' = 1, ... ,111. with i=l, ... ,n The value of Comp(J.1>e) are not known a priori and have only a meaning when they are compared to the value Comp(J.1>ef est of the comparing function for the best profile. 3.2.2 Geometrical Interpretation To illustrate the difference between the neural network learning of a scoring function and a comparing function, we propose a geometrical interpretation in the case of a linear network having as input vect.ors (profiles) XJ, ... ,XJ, ... ,Xp associated with a perturbed schedule j. The learning of a scoring function which associat.es a score Score; with each input vector xj consists in finding a hyperplane in the input vector space which is tangent to the circles of cent.er and radius SC01>e{ (Fig. 1). On the contrary the learning of a comparing function consists t.o obt.ain t.he equation of an hyperplane such that the end-point of the vector Xfest is nearer the hyperplane than the end-points of the other input vectors associated with the same perturbed schedule j (Fig. 1). xf XJ 3.2.3 Learning We use a neural network approach to build the comparing function and the mean squared error as a measure of the quality of t.he approximation. The comparing function is approximated by a non-linear function: C(P1>ojile;) = N?(W,Xj) where W is the weight. vector of the neural network (e.g backpropagat.ion network). The problem of finding C which has the property of (1) is equivalent to finding the function C that minimizes the following error function (Braun and al., 1991) where <I> is the sigmoid function : Comparison Training for a Rescheduling Problem in Neural Networks ,-- --_ .. _... """", , ,, \ ,, , , x,, I I , ,,, ,, , .,' I " -- ,.,1' x..,,, ', ....... --,' W( Wl ,w2'''')' .. ,W L) wlh .w. - 1 Figure 1: Geometrical Interpretation of the learning of a Scoring Function (Rigth) and a Comparing Function (Left) n I: (2) = i 1 i -::f Best To obtain t.he weight vector which mll1UTIlzes the error funct.ion (2), we use the property that t.he -gr~ld?~(W) point.s in the direct.ion in which the error function will decrease at the fastest possible rate. To update t.he weight we have thus to calculate the partial derivative of (2) with each components of the weight vector ltV: it is made of a product of three factors. The evaluation of the first two factors (the sigmoid and the derivative of the sigmoid) is immediate. The third factor is the partial derivative of the non-linear function N ?, which is generally calculated by using the generalized delta rule learning law (Rumelhart. and McClelland, 1986), Unlike the linear associator network, for the backpropagation network, the error e3t function (2) is not equivalent to the error function where the difference is associated with the input vector of the backpropagation network because: Xl X; (3) By consequence to calculate t.he three factors of the partial derivative of (2), we have to introduce separately at the bottom of the network t.he input vector of the best profile X!e3t and the input vector of a less good profile Then we have to memorize theIl' partial contribution at each node of the network and multiply their contributions before updating the weight . Using this way to evaluate the derivative of (2) and to update t.he weight, the simplicity of the generalized delta rule learning law has disappeared . XJ. 805 806 Keymeulen and de Gerlache 3.2.4 Architecture Tesauro (Tesauro and Sejnowski, 1989) proposes an architecture, t.hat we call dual neural network, and a learning procedure such that the simplicity of the generalized delta rule learning law can still be used (Fig. 2). The same kind of architecture, called siamese network, was recently used by Bromley for the signature verification (Bromley and al., 1994). The dual neural network architecture and the learning strategy are justified mathematically at one hand by the decomposition of the partial derivative of the error function (2) in a sum of two terms and at the other hand by the asymmetry property of the sigmoid and its derivative. The architecture of the dual neural network consists to duplicate the multi-layer network approximating the comparing function (1) and to connect the output of both to a unique output node through a positive unit weight for the left network and negative unit weight. for the right network. During the learning a couple of profiles is presented to the dual neural network: a best profile X e3t and a less good profile X!. The desired value at the output node of the dual neural network is 0 when the left network has for input the best profile and the right network has for input a less good profile and 1 when these profiles are permuted. During the recall we work only with one of the two multi-layer networks, suppose the left one (the choice is of no importance because they are exactly the same). The profiles JY~ associated with a perturbed schedule j are presented at the input of the left. multi-layer network. The best profile is the one having the lowest value at the output of the left multi-layer network. f Through this mathematical formulation we can use the suggestion of Braun to improve the learning convergence (Braun and al., 1991). They propose to replace the positive and negative unit weight het.ween the output node of the multi-layer networks and the output. node of the dual neural network by respect.ively a weight value equal to V for the left net.work and - V for the right. network. They modify the value of V by applying the generalized delt.a rule which has no significant impact on the learning convergence. By manually increasing the factor V during the learning procedure, we improve considerably the learning convergence due to its asymmetric impact on the derivative of ?<I>(W) with W: the modification of the weight vector is greater for couples not yet learned than for couples already learned. 4 Results The experiments show the abilit.y of our model to help the CPR team of the Sabena Belgian Airline company to choose the best profile in a group of PSPs based on the learned expertise of the team. To codify the profile we identify 15 relevant parameters. They constitute the input of our neural network. The training data set was obtained by analyzing the CPR team at work during 15 days from which we retain our training and test perturbed schedules. We consider that the network has learned when the comparing value of the best profile is less than the comparing value of t.he other profiles and that for all training perturbed schedules. At that time ?cJ>(W) is less t.han .5 for every couple of profiles. The left graph of Figure 3 shows t.he evolution of t.he mean error over t.he couples Comparison Training for a Rescheduling Problem in Neural Networks 807 Dual Neural Network ?1 C_ .. n C_,.n Belt. .1Ir.(~.l 'J BeltoJ ) .11r.(~. t ZMulI.La,'"" -'-od ..... .) ,~ Nouol NoI",ort.o 10.6lo.~ -+ (X -+ .X But~ toJ 10,910", 10.110.710.11 ) ? wdb XIJ B.... I63IO.2!O-.1 b.61 0.21 0.91 0.21 10.11 0,91 0.71 0.91 0.11 0.11 0,61 Wllb X 2?1 X B.... I =X But.l it.t ) 10.610.710.1/ 0.11 0.711.01 0.21 ,~ XB.... I 10.310.210.11 0.61 0.21 0.91 0.21 X 2?1 10,616316.91 631 0.11 6.71 0.21 XB..... lo. 1 Io.9Io.7! X I .. X I.. 0.910.110.11 0.&1 10.610.710.1( 0.11 0.711.0 I 0.11 XB..... 'S .. = X Be... Figure 2: The training of a dual neural network. during the training . The right graph shows the improvement of the convergence when the weight V is increased regularly during the training process. lnclasilll V OJ or the Dull NaIl'll Ndworll o.s OA 0.? 0.1 OJ M~--~~~---------------------- 02 ~--~~------------------------ 0.1 01 Nwnbcr of'lhilitg liIl) liIO ~~~~~~~~~~~~~~~~ 111(1) IlO) Figure 3: Convergence of the dual neural network architecture. The network does not converge when we introduce contradictory decisions in our training set. It is possible to resolve them by adding new context parameters in the coding scheme of the profile. After learning, our network shows generalization capacity by retrieving the best profile for a new perturbed schedule that is similar to one which has already been learned. The degree of similarity required for the generalization remains a topic for further study. Nunt.crci nW~ 808 Keymeulen and de Gerlache 5 Conclusion In conclusion, we have shown that the rescheduling problem of an airline crew pool can be stated as a decision making problem, namely the identification of the best potential substitute pilot. We have stressed the importance of the codification of the information used by the expert to evaluate the best candidate. We have applied the neural network learning approach to help the rescheduler team in the rescheduling process by using the experience of already solved rescheduling problems. By a mathematical analysis we have proven the efficiency of the dual neural network architecture. The mathematical analysis permits also to improve the convergence of the network. Finally we have illustrated the method on rescheduling problems for the Sabena Belgian Airline company. Acknowledgments We thank the Scheduling and Rescheduling team of Mr. Verworst at Sabena for their valuable information given all along this study; Professors Steels and D'Hondt from the VUB and Professors Pastijn, Leysen and Declerck from the Military Royal Academy who supported this research; Mr. Horner and Mr. Pau from the Digital Europe organization for their funding. We specially thank Mr. Decuyper and Mr. de Gerlache for their advices and attentive reading. References H. Braun, J. Faulner & V. Uilrich. (1991) Learning strategies for solving the problem of planning using backpropagation. In Proceedings of Fourth International Conference on Neural Networks and their Applications, 671-685. Nimes, France. J. Bromley, I. Guyon, Y . Lecun, E. Sackinger, R. Shah . (1994). Signature verification using a siamese delay neural network. In J. Cowan, G. Tesauro & J. Alspector (eds.), Advances in Neural Information Processing Systems 1. San Mateo, CA: Morgan Kaufmann. M. de Gerlache & D. Keymeulen. (1993) A neural network learning strategy adapted for a rescheduling problem. In Proceedings of Fourth International Conference on Neural Networks and their Applications, 33-42 . Nimes, France. D. Rumelhart & J. McClelland. (1986) Parallel Distributed Processing: Explorations in the Microstructure of Cognition I [1 II. Cambridge, MA: MIT Press. L. Steels. (1990) Components of expertise. AI Maga.zine, 11(2):29-49 . G. Tesauro. (1989) Connectionist learning of expert preferences by comparison training. In D. S. Touretzky (ed.), Advances in Neural Information Processing Systems 1, 99-106. San Mateo, CA: Morgan Kaufmann. G. Tesauro & T.J. Sejnowski. (1989) A parallel network that learns to play backgammon. Artificial Intelligence, 39:357-390.
871 |@word llsed:1 cha:1 decomposition:1 ld:1 reduction:1 initial:1 substitution:1 score:5 past:1 comparing:12 od:1 yet:1 must:1 heir:2 update:2 intelligence:2 short:1 didier:1 node:5 preference:1 belt:1 mathematical:5 along:1 direct:1 retrieving:1 consists:4 introduce:2 notably:1 indeed:2 alspector:1 p1:2 dist:1 planning:2 multi:5 behavior:1 company:5 resolve:1 actual:1 little:1 increasing:1 becomes:1 lowest:2 kind:1 substantially:1 minimizes:1 finding:4 every:2 braun:6 exactly:1 unit:3 ly:1 before:1 positive:2 modify:1 limit:2 consequence:1 io:2 analyzing:1 mateo:2 fastest:1 unique:1 acknowledgment:1 lecun:1 backpropagation:4 procedure:4 significantly:1 suggest:1 get:1 scheduling:6 context:2 influence:1 applying:1 optimize:1 equivalent:2 missing:1 simplicity:2 immediately:1 rule:6 his:1 stability:4 suppose:1 mples:1 play:1 associate:3 rumelhart:2 approximated:1 updating:1 asymmetric:1 bottom:1 solved:2 calculate:2 ensures:2 decrease:1 noi:1 valuable:1 hese:2 signature:2 solving:1 funct:1 flying:1 efficiency:2 pspi:2 represented:1 ompare:1 describe:1 sejnowski:4 artificial:2 choosing:1 solve:2 think:1 itself:1 net:3 propose:2 product:1 relevant:2 academy:1 convergence:7 ach:1 asymmetry:1 disappeared:1 help:2 illustrate:1 fixing:1 solves:1 crew:14 memorize:1 differ:2 radius:1 exploration:1 human:1 explains:1 require:1 microstructure:1 generalization:3 mathematically:1 paramet:2 cpr:5 c_:2 pleinlaan:2 bromley:3 mapping:1 nw:1 cognition:1 reserve:2 consecutive:1 belgium:2 integrates:1 ain:2 him:2 wl:1 tool:1 reflects:1 mit:1 ribution:1 ltv:1 improvement:1 backgammon:3 rank:2 toj:1 france:2 selects:1 among:1 ill:1 dual:11 priori:1 proposes:1 equal:2 construct:1 having:2 manually:1 represents:1 reassigning:1 future:1 connectionist:1 duplicate:1 vrije:2 phase:1 organization:1 possibility:2 multiply:1 evaluation:2 ilo:1 xb:3 ively:1 partial:5 belgian:3 necessary:3 experience:2 circle:1 desired:1 minimal:1 increased:1 vub:1 military:1 cost:2 delay:3 gr:1 connect:1 answer:1 perturbed:16 considerably:1 chooses:1 st:1 international:2 retain:1 pool:8 squared:1 martine:1 choose:2 expert:14 inefficient:1 derivative:8 return:2 syst:1 account:3 potential:5 de:8 coding:2 ranking:4 vi:1 view:3 universiteit:2 start:1 parallel:2 contribution:2 ir:1 kaufmann:2 characteristic:1 who:1 correspond:1 identify:2 dealt:1 identification:1 comp:3 expertise:2 touretzky:1 ed:5 attentive:1 involved:1 associated:12 couple:7 pilot:9 ask:2 recall:1 knowledge:2 organized:1 schedule:34 cj:1 day:3 follow:1 formulation:2 evaluated:1 implicit:2 flight:5 working:1 hand:2 sackinger:1 ident:1 quality:1 evolution:1 hence:1 dull:1 laboratory:2 codify:2 illustrated:1 deal:1 ll:2 game:5 during:7 criterion:1 generalized:4 reschedule:2 geometrical:3 meaning:1 ef:1 recently:1 funding:1 sigmoid:4 permuted:1 association:1 he:21 interpretation:3 significant:1 cambridge:1 ai:1 han:1 similarity:1 europe:1 ort:1 optimizing:1 tesauro:11 issued:1 success:1 scoring:7 seen:1 morgan:2 greater:1 mr:5 converge:1 paradigm:2 period:2 ween:1 ii:1 siamese:2 full:1 ing:2 xf:1 adapt:1 retrieval:1 jy:1 impact:4 ained:1 represent:1 ion:6 justified:1 separately:1 w2:1 unlike:1 airline:8 specially:1 cowan:1 member:6 regularly:2 contrary:1 call:2 xj:7 psps:1 architecture:12 reduce:1 rescheduling:28 effort:1 wo:1 nine:1 constitute:1 backpropagat:1 adequate:1 generally:2 useful:1 obt:3 brussel:2 morris:1 mcclelland:2 xij:1 delta:3 per:2 group:2 graph:2 year:1 sum:1 legislation:1 fourth:2 prog:1 guyon:1 nail:1 architectural:1 decision:3 acceptable:1 layer:5 exa:1 activity:2 adapted:2 precisely:1 software:1 according:1 brussels:2 psp:10 describes:2 em:1 abilit:1 making:2 modification:1 ypes:1 equation:1 liil:1 payment:1 remains:1 end:2 operation:1 permit:1 apply:2 ype:2 shah:1 hat:1 substitute:5 cent:1 nimes:2 neura:1 build:1 approximating:1 jbest:1 move:1 already:3 strategy:4 thank:2 oa:1 capacity:1 topic:1 argue:1 reason:2 reorganization:1 relationship:1 negative:2 stated:1 steel:3 ake:1 vect:1 immediate:1 team:11 perturbation:4 het:1 rating:3 namely:3 required:1 security:1 merges:1 learned:5 hour:3 nearer:1 horner:1 able:2 usually:3 reading:1 solut:1 program:1 oj:2 royal:1 event:2 scheme:2 improve:5 technology:1 tangent:1 occupation:1 relative:1 law:3 suggestion:1 proven:1 digital:1 integrate:1 degree:1 sufficient:2 xp:1 verification:2 pi:1 lo:2 supported:1 last:3 taking:2 distributed:1 calculated:1 pert:2 made:2 san:2 social:1 implicitly:1 learn:3 associator:1 ca:2 operational:1 unavailable:1 main:1 profile:37 associat:3 fig:3 advice:1 en:1 board:1 xl:1 candidate:1 third:2 learns:1 specific:2 er:3 list:3 consist:1 adding:1 importance:3 ools:1 occurring:1 pau:1 updat:1 suited:2 corresponds:1 ma:1 month:1 fbest:1 replace:3 absence:2 professor:2 change:1 hard:1 hyperplane:3 contradictory:1 called:3 e:2 la:1 equity:3 player:1 est:1 formally:1 stressed:1 accelerated:1 evaluate:2
7,104
872
Exploiting Chaos to Control the Future Gary W. Flake* Guo-Zhen Sun t Yee-Chun Lee t Hsing-Hen Chen t Institute for Advance Computer Studies University of Maryland College Park, MD 20742 Abstract Recently, Ott, Grebogi and Yorke (OGY) [6] found an effective method to control chaotic systems to unstable fixed points by using only small control forces; however, OGY's method is based on and limited to a linear theory and requires considerable knowledge of the dynamics of the system to be controlled. In this paper we use two radial basis function networks: one as a model of an unknown plant and the other as the controller. The controller is trained with a recurrent learning algorithm to minimize a novel objective function such that the controller can locate an unstable fixed point and drive the system into the fixed point with no a priori knowledge of the system dynamics. Our results indicate that the neural controller offers many advantages over OGY's technique. 1 Introduction Recently, Ott, Grebogi and Yorke (OGY) [6] proposed a simple but very good idea. Since any small perturbation can cause a large change in a chaotic trajectory, it is possible to use a very small control force to achieve a large trajectory modification. Moreover, due to the ergodicity of chaotic motion, any state in a chaotic *Department of Computer Science, peyote@umiacs.umd.edu tLaboratory for Plasma Research 647 648 Flake, Sun, Lee, and Chen attractor can be reached by a small control force . Since OGY published their work, several experiments and simulations have proven the usefulness of OGY's method. One prominent application of OGY's method is the prospect of controlling cardiac chaos [1] . We note that there are several unfavorable constraints on OGY's method. First, it requires a priori knowledge of the system dynamics, that is, the location of fixed points. Second, due to the limitation of linear theory, it will not work in the presence of large noise or when the control force is as large as beyond the linear region from which the control law was constructed. Third, although the ergodicity theory guarantees that any state after moving away from the desired fixed point will eventually return to its linear vicinity, it may take a very long time for this to happen, especially for a high dimensional chaotic attractor. In this paper we will demonstrate how a neural network (NN) can control a chaotic system with only a small control force and be trained with only examples from the state-space. To solve this problem, we introduced a novel objective function which measures the distance between the current state and its previous average. By minimizing this objective function, the NN can automatically locate the fixed point. As a preliminary step , a training set is used to train a forward model for the chaotic dynamics. The work of Jordan and Rumelhart [4] has shown that control problems can be mapped into supervised learning problems by coupling the outputs of a controller NN (the control signals) to the inputs of a forward model of a plant to form a multilayer network that is indirectly recurrent. A recurrent learning a.lgorithm is used to train the controller NN. To facilitate learning we use an extended radial basis function (RBF) network for both the forward model and the controller. To benchmark with OGY's result, the Himon map is used as a numerical example. The numerical results have shown the preliminary success of the proposed scheme. Details will be given in the following sections. In the next section we give our methodology and describe the general form of the recurrent learning algorithm used in our experiments. In Section 3, we discuss RBF networks and reintroduce a more powerful version. In Section 4, the numerical results are presented in detail. Finally, in Section 5, we give our conclusions. 2 Recurrent Learning for Control Let kC) denote a NN whose output, tit, is composed through a plant, l(?), with unknown dynamics. The output of the unknown plant (the state), it+l' forms part of the input for the NN a.t the next time step, hence the recurrency. At each time step the state is also passed to an output function , gC), which computes the sensation, Yt+l. The time evolution of this system is more accurately described by fit it+l Yt+l k(it ,!h+l'W) {(it, fit) g(Xt+I), where ii7+1 is the desired sensation for time t + 1 and W represents the trainable weights for the network . Additionally, we define the temporally local and global Exploiting Chaos to Control the Future error functionals J t = ~11Y7 - Ytl1 2 and E = L~I Ji, where N is the final time step for the system. The real-time recurrent learning (RTRL) algorithm [9] for training the network weights to minimize E is based on the fair assumption that minimizing the local error functionals with a small learning rate at each time step will correspond to minimizing the global error. To derive the learning algorithm, we can imagine the system consisting of the plant, controller, and error functionals as being unfolded in time. From this perspective we can view each instance of the controller NN as a separate NN and thus differentiate the error functionals with respect to the network weights at different times. Hence, we now add a time index to Wt to represent this fact. However, when we use W without the time index, the term should be understood to be time invariant. We can now define the matrix f t =~ a~t = ~it a aiI-t- 1 + (axt aiI L.J a....t- 1 + a....ait ) ?) - i=O UWi ?) UUt-l Wt-I ?)- UUt-l Xt-I Xt-l ft-I, (1) which further allows us to define aJi aw aE aw (2) (3) Equation 2 is the gradient equation for the RTRL algorithm while Equation 3 is for the backpropagation through time (BPTT) learning algorithm [7]. The gradients defined by these equations are usually used with gradient descent on a multilayer perceptron (MLP) . We will use them on RBF networks. 3 The CNLS Network The Connectionist Normalized Local Spline (eNLS) network [3] is an extension of the more familiar radial basis function network of Moody and Darken [5]. The forward operation of the network is defined by (4) where (5) All of the equations in this section assume a single output. Generalizing them for multiple outputs merely adds another index to the terms. For all of our simulations, we choose to distribute the centers, iii, based on a sample of the input space. 649 650 Flake, Sun, Lee, and Chen Additionally, the basis widths, f3i' are set to an experimentally determined constant . Because the output, <p, is linear in the terms Ii and d~, training them is very fast. To train the CNLS network on a prediction problem we, can use a quadratic error function of the form E = ~(y(i) - qj(i?2, where y(i) is the target function that we wish to approximate. We use a one-dimensional Newton-like method [8] which yields the update equations If + 7J (y(i) - <P(i?L'~~~i)' ~ + 7J (y(x) - <p(x?~=---=-!J...?...!..lo..::....L-- The right-most update rules form the learning algorithm when using the CNLS network for prediction, where 7J is a learning rate that should be set below 1.0. The left-most update rules describe a more general learning algorithm that can be used when a target output is unknown. When using the CNLS network architecture as part of a recurrent learning algorithm we must be able to differentiate the network outputs with respect to the inputs. Note that in Equations 1 and 2 each of the terms aXt/aUt-l, aUt-daxt-l, ait/Bit- 1 , and Biii/ aii can either be exactly solved or approximated by differentiating a CNLS network. Since the CNLS output is highly nonlinear in its inputs, computing these partial derivatives is not quite as elegant as it would be in a MLP . Nevertheless, it can be done. We skip the details and just show the end result: ann a: = ~ d~Pi(X) + 2~(pj (x) qj f3j l=l 4 (aj - i)) - 2<p(x)::;, (6) J=l Adaptive Control By combining the equations from the last two sections, we can construct a recurrent learning scheme for RBF networks in a similar fashion to what has been done with MLP networks. To demonstrate the utility of our technique, we have chosen a wellstudied nonlinear plant that has been successfully modeled and controlled by using non-neural techniques. Specifically, we will use the Henon map as a plant, which has been the focus of much of the research of OGY [6]. We also adopt some of their notation and experimental constraints. 4.1 The Himon Map The Henon map [2] is described by the equations (7) (8) Exploiting Chaos to Control the Future = = where A Ao + p and p is a control parameter that may be modified at each time step to coerce the plant into a desirable state. For all simulations we set Ao 1.29 and B 0.3 which gives the above equations a chaotic attracter that also contains an unstable fixed point. Our goal is to train a CNLS network that can locate and drive the map into the unstable fixed point and keep it there with only a minimal amount of information about the plant and by using only small values of p. = The unstable fixed point (XF, YF) in Equations 7 and 8 can be easily calculated as XF = YF ~ 0.838486. Forcing the Henon map to the fixed point is trivial if the controller is given unlimited control of the parameter. To make the problem more realistic we define p* as the maximum magnitude that p can take and use the rule below on the left if Ipi < p* _ {p if Ipl < p* if p > p* Pn 0 if Ipl > p* if p < -p* while OGY use the rule on the right. The reason we avoid the second rule is that it cannot be modeled by a CNLS network with any precision since it is step-like. The next task is to define what it means to "control" the Henon map. Having analytical knowledge of the fixed point in the attracter would make the job of the controller much easier, but this is unrealistic in the case where the dynamics of the plant to control are unknown. Instead, we use an error function that simply compares the current state of the plant with an average of previous states: et=21 [(Xt-(x)r) 2+(Yt-(Y)r) 2] , (9) where (.)r is the average of the last T values of its argument. This function approaches zero when the map is in a fixed point for time length greater than T. This function requires no special knowledge about the dynamics of the plant, yet it still enforces our constraint of driving the map into a fixed point. The learning algorithm also requires the partial derivatives of the error function with respect to the plant state variables, which are oet!f)xt = Xt - (x}r and oet!oYt = Yt - (Y)r. These two equations and the objective function are the only special purpose equations used for this problem. All other equations generalize from the derivation of the algorithm. Additionally, since the "output" representation (as discussed earlier) is identical to the state representation, training on a distinct output function is not strictly necessary in this case. Thus, we simplify the problem by only using a single additional model for the unknown next-state function of the Henon map. 4.2 Simulation To facilitate comparison between alternate control techniques, we now introduce the term f6t where 6t is a random variable and f is a small constan~ which specifies the intensity of the noise. We use a Gaussian distribution for bt such that the distribution has a zero mean, is independent, and has a variance of one. In keeping with [6], we discard any values of 6t which are greater in magnitude than 10. For training we set f 0.038. However, for tests on the real controller, we will show results for several values of f. = 651 652 Flake, Sun, Lee, and Chen (a) ? ? ? ? (b) ? ? (c) ? (f) ? ? ? """ 'r (d) ? ? - (e) . ? . Figure 1: Experimental results from training a neural controller to drive the Himon map into a fixed point. From (a) to (f), the values of fare 0.035, 0.036, 0.038, 0.04,0 .05, and 0.06, respectively. The top row corresponds to identical experiments performed in [6]. We add the noise in two places. First, when training the model, we add noise to the target output of the model (the next state). Second, when testing the controller on the real Henon map, we add the noise to the input of the plant (the previous state). In the second case, we consider the noise to be an artifact of our fictional measurements; that is, the plant evolves from the previous noise free state. Training the controller is done in two stages: an off-line portion to tune the model and an on-line stage to tune the controller. To train the model we randomly pick a starting state within a region (-1.5, 1.5) for the two state variables. We then iterate the map for one hundred cycles with p = 0 so that the points will converge onto the chaotic attractor. Next, we randomly pick a value for p in the range of (-p*, p*). The last state from the iteration is combined with this control parameter to compute a target state. We then add the noise to the new state values. Thus, the model input consists of a clean previous state and a control parameter and the target values consist of the noisy next state. We compute 100 training patterns in this manner. Using the prediction learning algorithm for the CNLS network we train the model network on each of the 100 patterns (in random order) for 30 epochs. The model quickly converges to a low average error. In the next stage, we use the model network to train the controller network in two ways. First, the model acts as the plant for the purposes of computing a next state. Additionally, we differentiate the model for values needed for the RTRL algorithm. We train the controller for 30 epochs, where each epoch consists of 50 cycles. At the beginning of each epoch we initialize the plant state to some random values (not necessarily on the chaotic attracter ,) and set the recurrent history matrix, Exploiting Chaos to Control the Future ... - .... _ ... -. _ .._..... ... _-- .. _.... ... _. -.-_ ..... .. (a) (b) . -.- -_... _---.-- . .... _... ...... __ .- -. - ..... . .-.-.. _. _- .... _... .. (c) Figure 2: Experimental results from [6]. From left to right, the values of f. are 0.035, 0.036, and 0.038, respectively. r t, to zero. Then, for each cycle, we feed the previous state into the controller as input. This produces a control parameter which is fed along with the previous state as input into the model network, which in turn produces the next state. This next state is fed into the error function to produce the error signal. At this point we compute all of the necessary values to train the controller for that cycle while maintaining the history matrix. In this way, we train both the model and control networks with only 100 data points, since the controller never sees any of the real values from the Henon map but only estimates from the model. For this experiment both the control and model RBF networks consist of 40 basis functions. 4.3 Summary Our results are summarized by Figure 1. As can be seen, the controller is able to drive the Henon Map into the fixed point very rapidly and it is capable of keeping it there for an extended period of time without transients. As the level of noise is increased, it can be seen that the plant maintains control for quite some time . The first visible spike can be observed when f. = 0.04. These results are an improvement over the results generated from the best nonneural technique available for two reasons: First, the neural controller that we have trained is capable of driving the Henon map into a fixed point with far fewer transients then other techniques. Specifically, alternate techniques , as illustrated in Figure 2, experience numerous spikes in the map for values of f. for which our controller is spike-free (0.035 - 0.038). Second, our training technique has smaller data requirements and uses less special purpose information. For example, the RBF controller was trained with only 100 data points compared to 500 for the nonneural. Additionally, non-neural techniques will typically estimate the location of the fixed point with an initial data set. In the case of [6] it was assumed that the fixed point could be easily discovered by some technique, and as a result all of their experiments rely on the true (hard-coded) fixed point. This, of course, could be discovered by searching the input space on the RBF model, but we have instead allowed the controller to discover this feature on its own. 653 654 Flake, Sun, Lee, and Chen 5 Conclusion and Future Directions A crucial component of the success of our approach is the objective function that measures the distance between the current state and the nearest time average. The reason why this objective function works is that during the control stage the learning algorithm is minimizing only a small distance between the current point and the "moving target." This is in contrast to minimizing the large distance between the current point and the target point, which usually causes unstable long time correlation in chaotic systems and ruins the learning. The carefully designed recurrent learning algorithm and the extended RBF network also contribute to the success of this approach. Our results seem to indicate that RBF networks hold great promise in recurrent systems. However, further study must be done to understand why and how NNs could provide more useful schemes to control real world chaos. Acknowledgements We gratefully acknowledge helpful comments from and discussions with Chris Barnes, Lee Giles, Roger Jones, Ed Ott, and James Reggia. This research was supported in part by AFOSR grant number F49620-92-J-0519. References [1] A. Garfinkel, M.L. Spano, and W.L. Ditto. Controlling cardiac chaos. Science, 257(5074):1230, August 1992. [2] M. HEmon. A two-dimensional mapping with a strange attractor. Communications in Mathematical Physics, 50:69-77, 1976. [3] R.D. Jones, Y.C. Lee, C.W. Barnes, G.W. Flake, K. Lee, P.S. Lewis, and S. Qian. Function approximation and time series prediction with neural network. In Proceedings of the International Joint Conference on Neural Networks, 1990. [4] M.1. Jordan and D.E. Rumelhart. Forward models: Supervised learning with a distal teacher. Technical Report Occasional Paper #40, MIT Center for Cognitive Science, 1990. [5] J. Moody and C. Darken. Fast learning in networks of locally-tuned processing units. Neural Computation, 1:281-294, 1989. [6] E. Ott, C. Grebogi, and J .A. Yorke. Controlling chaotic dynamical systems. In D.K. Campbell, editor, CHAOS: Soviet-American Perspectives on Nonlinear Science, pages 153-172. American Institute of Physics, New York, 1990. [7] F.J. Pineda. Generalization of back-propagation to recurrent neural networks. Physical Review Letters, 59:2229-2232, 1987. [8] W.H. Press, B.P. Flannery, S.A. Teukolsky, and W.T. Vetterling. Numerical Recipes. Cambridge University Press, Cambridge, 1986. [9] R.J. Williams and D. Zipser. Experimental analysis of the real-time recurrent learning algorithm. Connection Science, 1:87-111, 1989.
872 |@word version:1 cnls:9 bptt:1 simulation:4 pick:2 initial:1 contains:1 series:1 tuned:1 current:5 yet:1 must:2 realistic:1 visible:1 numerical:4 happen:1 designed:1 update:3 fewer:1 beginning:1 contribute:1 location:2 ipi:1 ditto:1 mathematical:1 along:1 constructed:1 consists:2 manner:1 introduce:1 automatically:1 unfolded:1 discover:1 moreover:1 notation:1 what:2 guarantee:1 act:1 axt:2 exactly:1 control:29 unit:1 grant:1 understood:1 local:3 limited:1 range:1 enforces:1 testing:1 backpropagation:1 chaotic:12 aji:1 radial:3 cannot:1 onto:1 yee:1 ogy:11 map:17 yt:4 center:2 williams:1 starting:1 qian:1 rule:5 searching:1 controlling:3 imagine:1 target:7 us:1 rumelhart:2 approximated:1 observed:1 ft:1 solved:1 region:2 reintroduce:1 sun:5 cycle:4 prospect:1 dynamic:7 trained:4 tit:1 basis:5 easily:2 aii:3 joint:1 soviet:1 derivation:1 train:10 distinct:1 fast:2 effective:1 describe:2 hsing:1 whose:1 quite:2 solve:1 coerce:1 noisy:1 final:1 differentiate:3 advantage:1 pineda:1 analytical:1 combining:1 rapidly:1 achieve:1 recipe:1 exploiting:4 requirement:1 produce:3 converges:1 coupling:1 recurrent:13 derive:1 nearest:1 job:1 skip:1 indicate:2 direction:1 sensation:2 transient:2 ao:2 generalization:1 preliminary:2 extension:1 strictly:1 hold:1 ruin:1 great:1 mapping:1 driving:2 adopt:1 purpose:3 successfully:1 mit:1 gaussian:1 modified:1 pn:1 avoid:1 focus:1 improvement:1 contrast:1 helpful:1 nn:8 vetterling:1 bt:1 typically:1 kc:1 priori:2 special:3 initialize:1 construct:1 never:1 having:1 identical:2 represents:1 park:1 jones:2 future:5 connectionist:1 spline:1 simplify:1 report:1 randomly:2 composed:1 familiar:1 uwi:1 consisting:1 attractor:4 mlp:3 highly:1 wellstudied:1 oet:2 capable:2 partial:2 necessary:2 experience:1 desired:2 minimal:1 instance:1 increased:1 earlier:1 giles:1 ott:4 hundred:1 usefulness:1 aw:2 teacher:1 combined:1 nns:1 international:1 lee:8 off:1 physic:2 quickly:1 moody:2 choose:1 cognitive:1 american:2 derivative:2 return:1 distribute:1 summarized:1 performed:1 view:1 reached:1 portion:1 maintains:1 minimize:2 variance:1 correspond:1 yield:1 generalize:1 accurately:1 trajectory:2 drive:4 published:1 history:2 ed:1 james:1 knowledge:5 f3j:1 carefully:1 campbell:1 back:1 feed:1 supervised:2 methodology:1 y7:1 done:4 ergodicity:2 just:1 stage:4 roger:1 correlation:1 nonlinear:3 propagation:1 yf:2 artifact:1 aj:1 facilitate:2 normalized:1 lgorithm:1 true:1 evolution:1 vicinity:1 hence:2 illustrated:1 distal:1 during:1 width:1 prominent:1 demonstrate:2 motion:1 chaos:8 novel:2 recently:2 ji:1 physical:1 discussed:1 fare:1 measurement:1 cambridge:2 gratefully:1 moving:2 fictional:1 add:6 own:1 perspective:2 forcing:1 discard:1 success:3 seen:2 aut:2 greater:2 additional:1 converge:1 period:1 signal:2 ii:1 multiple:1 desirable:1 technical:1 xf:2 offer:1 long:2 coded:1 controlled:2 prediction:4 multilayer:2 ae:1 controller:26 iteration:1 represent:1 crucial:1 umiacs:1 umd:1 comment:1 elegant:1 seem:1 jordan:2 zipser:1 presence:1 iii:1 iterate:1 fit:2 architecture:1 f3i:1 idea:1 qj:2 utility:1 passed:1 york:1 cause:2 useful:1 tune:2 amount:1 locally:1 specifies:1 promise:1 nevertheless:1 pj:1 clean:1 merely:1 letter:1 powerful:1 place:1 strange:1 bit:1 quadratic:1 barnes:2 constraint:3 unlimited:1 argument:1 department:1 alternate:2 flake:6 smaller:1 cardiac:2 rtrl:3 evolves:1 modification:1 invariant:1 equation:14 discus:1 eventually:1 turn:1 needed:1 fed:2 end:1 available:1 operation:1 occasional:1 away:1 indirectly:1 reggia:1 recurrency:1 top:1 maintaining:1 newton:1 especially:1 objective:6 spike:3 md:1 gradient:3 distance:4 separate:1 maryland:1 mapped:1 chris:1 unstable:6 trivial:1 reason:3 length:1 index:3 modeled:2 yorke:3 minimizing:5 unknown:6 darken:2 benchmark:1 acknowledge:1 descent:1 extended:3 communication:1 locate:3 gc:1 perturbation:1 discovered:2 august:1 intensity:1 introduced:1 connection:1 beyond:1 able:2 usually:2 below:2 pattern:2 dynamical:1 unrealistic:1 force:5 rely:1 scheme:3 temporally:1 numerous:1 zhen:1 epoch:4 review:1 acknowledgement:1 hen:1 law:1 afosr:1 plant:18 limitation:1 proven:1 editor:1 pi:1 lo:1 row:1 course:1 summary:1 supported:1 last:3 keeping:2 free:2 understand:1 perceptron:1 institute:2 ipl:2 differentiating:1 f49620:1 calculated:1 world:1 computes:1 forward:5 adaptive:1 spano:1 far:1 functionals:4 approximate:1 keep:1 global:2 assumed:1 constan:1 why:2 additionally:5 necessarily:1 noise:9 ait:2 fair:1 allowed:1 fashion:1 precision:1 wish:1 third:1 xt:6 uut:2 chun:1 biii:1 consist:2 magnitude:2 chen:5 easier:1 flannery:1 generalizing:1 simply:1 gary:1 corresponds:1 lewis:1 teukolsky:1 goal:1 ann:1 rbf:9 considerable:1 change:1 experimentally:1 hard:1 determined:1 specifically:2 wt:2 experimental:4 plasma:1 unfavorable:1 college:1 guo:1 trainable:1
7,105
873
x                    !#" & $ %('   ) *  ,+ } y z { | ~ -/.1032547680:9<;7=?>@A=?BC6 ; DFEHGJIHKMLNPOMQRSITEVUXWZY[]\^E W8\`_aNb I K N O IHKdc e If N Yg7h`E Yikj \PKml Yn O c oPppPqsr tu OMvPw NE N L \yxYi K N O lZz w N r NPE1N|{N8}~N?|??P?FoT????N? N? ? ? ??H?)?a?1?  D?? O?? \? K \ lM\?E O ? N ? \ K ? } \ ? K?I ? IP? \ N E \?E O?K?I ?? L \ O ?HIH? O I O K N E ? GJIXK t ?m? O ??? QR E O \ K ??N???K?\ ? Km\^l?\ E O N O Y?k?HE R e??H\???E O?KdI ? c G z1?XW^???Y I EYu ? ? \??)E \ ? }(?g]? ??? Kd\?l ? \?W O?O?I O?? \ l OMN O \?? Ga? Y?k?7??\?? z EHYR O ? O ? N O Q? ? ? YR ? O \ K E?N?? Kd? ? Km\^l?\?E O N?dY? I E)? e ?H\ Y??E O \ K E?NP? K \ ? K \^l \?E O N O Q?SI E?W?NE b1\ O?? N EH? G ? K L\ ? b c W8?XNEX{H?? E { O????S? ? vP? N?^? \ O \???? G I K O ? \?? E O K I ? c G z E W O Y? I E??~e ? z?? ? O?? \ ? ??NE ?dG I K?L N O Y? I E tn ? K \Z??\ K?K \ ? O ?? ?????m?d????A?8?|?Z????? ? ? ? ? e ? \/?? E O \ K E?N ? K \ ? K \ ? ? E O N O Yi I ? W N? b \`? K?v ? ? GJI KdL \ ? N W^W ? K?? Y?SEX? O ? {H?? j \?E??1? ?X? ??\ L lZ??? c ? K N EH?M? I?KF? t? ?X{ O?? ? Y? E O \ K E?N ? K \ ? K \8ld\?E O N O Yi ? EY? E O I O?? \ ? t? ? t? L z?? \PE O K ? ? ? K \ ? K \8ld\ ? ? N O Y? I ? ? } ? W NE ?H?1? O NPYn]E f \ K E ? ??E \ ?d} I ? f ?^? ?Z? NP?]?]\ K EH\Z?d} I?KM? l }(Y? OZ? \Z?????]Y? W^Y? O t? ?H? O \ ? )K \ O N O Y I E `E O?? \ I O ? \ K? NE ? ? b c W ? v E?{ t EX{N ?1? K ? ? K YiSN O \Z? c ? O  \ ? N K N ? \ O \ K ? ? }?\ W?N E ? b O N Yu EY? E O \ K?? \ ?) Q N Od? YiSE O \ K E?N ? K \^? K \^? \ ? O N O t I E ? Gm I O ? \?Yn L ? KmI j \ ?{ \?EX\ K N ??Y u N O Yi I E??  \ N ??? ? g \ ? ? ? \ \ E O K I ? c ? \ ? ? IH? OFI NE N  O ? \ EHW I?? ? K N? ?  \ ? z?W^W \Z\ ? \ ? t E I b O N tu EHYR ? { w \ K ?X\Z??E \Z? } IXK ?|? N  \?E? K I ? c  DdE N ?1? Y O Y I E ?  \ N ??????Yu \ ? O??H\  \ O ? I  Y RkO ? l L v ?  t ?]E O \ K E  O?I O ? ? \ GFKm"\ ! z \^E W ? Y??? \ E O Y#%$ W?N O t& IHE ? Km? b ??\?L NPE ? } \ W I z ? ? IHbX?FN t' E )? ( N  t/ 0 N O tR ? ? ? \ ? GmI 2K 1 N ? W84\ 365 ? Y? ? ? t8? 7 W N ? :? 9<; +Y * t ,.- \ ? ?H\ O  I ? f ? } ? I ?M? {\ E?\ Kd ? ? \P=K u IHK O I O ? \`? \PKAG?IH2K >N E W^\ b c ? O NE ? N K ? b?N W f? ? K I ? NP{NPO Yi IE ? ? C KLF B - R_^ 6 Ra`ZR V12 ; 2 Tb?0V MONPM Q 6SR 4?2"T U]0V0)WYXV 2ZR 6[V]\  ? \ I?G`O?? \LI ? ?Y R d L c IKdO NE O W ? N K?NPW O \PK?YR ? ? t e YR E OM? \ ? \?NHK?? Y? E ? ? cEX\[ ? N ? ?gfh?} ? K w ? Y? l O ?1v O E ? O }?I :K i ? W?N ? WPKm\?N O \ N ? ? K I ? KmYj]N???\YSEH?F\TK?E?NP??Kd\ ? ? \ ld\P? O N O tk I ? ? YlSE O ? \nm ?go_p+q r \I[s Ot ?m\ @ ABDC EGFIH6J u"vw       !#"%$%&(' )+*,  -.0/1& *    32+4 2 5  16 17  89  5;: <&>= 9 @? ,A B 5 9  C  DFE  &  G 4   4IH 9J  * - K ,ML NPO = Q *R SUTWV0 X &&Y   &M& *;& -Z 9\[<] J     - ^  _& 9_& [< 4 2!`aY D : 2  1 2!b  c01de=gfh&Fikjmlno<prq `Rs t 17 h2 Iu VIvxw y 2 5 &Mz 6{ ` r,|/  &!= ` 2@} , A h~A ! h& P`2| `<2 ?5 & 2 `  d1? $ = fr 6?? -? b `?  ??+? 2>- ? u ? 2?;?I2  q*  &  6  ? ` e  6@? `???x? ,%, f `E 0b D  ?R1?%?g? 2  f< C ?ar ?? `1?F ? 9 2 ? ???o?? :??U????o?? z  b ?? ??A `  A;*R?  ?o<d?& f  Y   2   ?d O s 6 f  ? 1&  2 Ih 2  4 2  ,   &1& -? O  : ?  ? [ -? A f & f   ? ?+?? 2 r  2  4? ` ?  C h2: &@? ?? 2  ? ]?J  /  t 7 6 V  ???? , A ? , 5 l? `  2  ,  & ??>? X 9B  ? 5  ` & 1?& ` ??h& ? A   Y $ d    2 z 6 Rb G?? 2  d  2 b {? 1? 6 f ???E z Y  .7 $ `2 A f  2  A &  2 ? , &> . A , ` s `?? &>? ?   p? .o? &  2   ??  4 2  ,   6  & ? `?U? ,;? ?r?m? ????? ? ?<?1?;? ? ? ? ?+?& ?U?  G ,?A `  & ?;? & T & [ `a1?   A &  .oC?`s  $ 4  4 2 -?5 & `?br??   h? 6 2>`4h? s E??? A &> ? `  s ` 2 & f ~-? B 7>? 2 /  Y 2  4?2  ,   &(z 6 -? ` / 1 b & `???`?2 D ?  &     ? & 2 ` ? $e?  &@f O p?& ` &M>1??;? `2>D & [  ?? / & ? ?/Iz g2>4 2 *,  / & 6{? ` ??? ? ` 2?b  ?  d & ` d  ? C?  4?2?`?J  ? q `?2 &  2 d  &+,;? s?`2? ? 1?? ?  ? : 6M? ?? : L?  4I2 9 C1 D */ & 9 s & f ?|-? 1&   4I2? 7>? - G Y  ? &!$ `Rs /  7 ? `2 ? q a ??  C l.` 2 , ` 2 ? `b  ?o d?q6 2 1??   .o , ?  p 6 ' ?!   D?4   `C1 D   & `Rs &  d   h2  l 1&   O / ?  2 s?` 2!D 1/<? ? ?`  ?o 6  2!4 2? & ? ? 4  {? &  $?6 [ - n & 12 /3 z    R4      &1& -? `  : 13b~  & ? ` 2 ?hq J  f  C  ? O<2 , : &    B & 2 ` 4 5 `  b?J  ? - ?  8 D"!#  b%$ &  6 [< ` ? f  2 [ 1rpIT s `2?& [  -  ? 4I2 ` C RbFd   R2   ? ' h& -8 ` / 4  2 s ` 2 D zh/<?  : & f )( +* 2 `+, $ 5;[ ` ?rb  4 4?2?` 4 .? -.oh&@  $ a ???f  u / d b  ? ? `?2 b ?   & ` d  .oC   4 2 9 a ?o ??, t ?102 34I?;?5476 894?? 5:1;=< & 2 ` 4 $ML >  &????*A@ 4%B z ?0 M& f   1 & ? ` 4 DC ?  A 6 {E `  T1F 5> b -G B?& - 8,IH  4   ? J K 2  5 y  ?R? & ` &  * f  . b p   F  -R &  ? 7  X C  .k& $ : LUTTWVYX S [ bdcegf Z+[ \  9 d Z+[ ] :  , b NOQP ? b ? - ?k& f ^._a` [ i fX ,   ` ? ??  ?  ;b zhA &> Q1  j  `hs9k l &Mf [ - mb b  B  n 6 <b &Mf  ,  ????h&+  ` o P 8 ? `  Qp j [ * f<r brb ?tsr fu & , ^wv [ xzyy *R/ ?r?? &.?{`}|~%f ? . ,  /?& 2 9 4 p -? 5?2> s R2 ? ? be& ` ?? ? ?????}??? ??????+? ??? ?w???{?=? T a  ?R  5MF& [ h/h& 2@`14 pY? E <?R&>. ` B ?, b }?   b ? -?& f 2 ;5 H  ? & & ` & [  . /?&?? ? ? ? B3  - ?  4+? 2??F - ? & u &+ X `  ? ? &Mf  ? ,  1& 2+`14 p -? 5 ?? ?  -t ? f R '  b : `   p `  F[  Rop b  E  ? &? . q & E ?  p `  : = frl?   ?   &Mf ? ` &;f ? 2 ' ?  . p b<  ?r??? & , z 2  & E 2  b `? J $ D F  & -? 4  5 & 2?`  d ? If - ? a  Xk&>9 2 p ? O /   ?&  c `  ? ??? ??? &  7[ ` &;f  2 [ z 3p :  .s * 1& 2 ` H p  ?, ?? ? ??  R   b#:   Y&Mf  [3-? b??<R ?3 ? & ,??1d? ? ( ? ? ?  p ? ?@-  C z & ?Ay ? s * ?& 2 O 4 p  G , 5 E? A  ?  1& q p b3;? 2  ]  b :I` ? p  , D ?   ID?a  2 `?s f -? b1?Q?   E  8 & , . <  & F 2 d? ?b ` ?T = [  m   h YI7 f g` & [h2 F  - u &>q  2  `?  ?<b  ` &  5+b?s?`<2 4I2 `b  A -?o  ` F ? 4 E j , ? [ ,}? & [ ? ,  & 2>` 4 p s E / ?R& ?? ` / ? z1 J ? ? q b ???Yr ?  & ? ? * E r  ?; , ,  2 p [  ?ob%?   E ??-?o& q 7 ` a * ?  Y  ? D { ? ? ? p :   ? 7 ` ? `  q & 2  ? & q - ? D 4 0   & ? 9 2 ? ,A? * h ? [2 " `C1 2 z q ? ?Q? ???14??}? : 5? ??? ? : 4 :1? 34I? ? 5 ??6 ? ?? 5 ? ?;? 4 ? ? ? ? ? :14 > R? E , ? ?  G }? p `  & ? .- ???`? 2  ? p  ] ` ? & [ P& 2   , s?`32 ?? &@-  ` / ` s  ?  &+* 2  zRV 2 4 2 ( ,  1&?? ? 6  ? `  q? p & [    ?d? & ? ? Iz  * /1& 2>`14 p ? s &Mf Y? G 17 ? ? 3   /1& 2 ` 4 p  8 , D  m / ??Q? - ???b : `   p `   f?Q??Q?   E ? & . E  oE , ?? ?  ? QzR& ? b Jpe  l?  4 & 4  & & * 2    &  ?  ?  8   D  . ?&  2 3  * /1& 2 9 4 p :  m & ? ? * f  u b b<? .?? E  -8 +* 5  2 *?zR?& - ?Q?h& ;b a p he . B 4 E & 4 1&+&+ 2  >  & F 5  O ` ?  6I? -? d E 2 D? :2 4 2  ,?? ??? ? &  ?  d & [ | ? 5 6  &  9 ?  ?Q? 6 ? 2 B3 Y?2 ( y 2  ,.? B 6  6?. 9? ? ? s ? ( D f? / -. ? f  ( 6 [<  o16 ?  z Y  / 6}? `4 $T ? ? ? ? O a 6  ? t  & [ *  ? 17 *  ?  1Y 2  4I2  ,   6  ?  ? O "? ?o& [ F 8 3?  E ?  & 2+`4 p &  6 f ` &?f 2 f z?Ip :  u s?? ?m/h& 2 `+? p - . , ? D z ?  . ?? ? ? * ? T  m &Mf  f - kbIp   ?r ?? &+5g 2 ? ?IE ??  $ ??&+ ? C h&  ?=?g?   b p3?o& -?o9?&>9?& = O  ? & 2 * ?  2! y 2? , * B* ?&  ? ?  ,;T ? ?? B ? z C1 D E ? &> 4 ?  ?o &   ?  p3?o &   ? /6  2 r ? ?  H 2? q ? /1& z 6 -? ` f ?Q?}?                !#"%$'& (   *)   +  /1023546879:7<;>=?:@BA>CED F GO ? GH ?EIJ 7 IK5LM8N ? IPO I C J IRQ I H ? N ?SM A H Q ?? / NETVUW 7 9 7YX H ? J A[ZD f R? ]?v? s r \1] ^`_Eab8cedf Eq8r fvu u ] w cylz ylc| q~} c ? t? c nE? q u |R? q fhc |?gj? ilc kP co m? nEs ? m o u ps8? ? u ? mp u s8?qt x n ? c?p _ u nu x ? ] 4 n ms {c y8??ylc s m o ? t? nu x? umBs8?  n s ? u? u s_q~c? ??| c ? q q c f?? ? q c??u c n s m x t n m ? ? t ? s c b ? c ? '? ? s c ylc | q c o c n s m s t ?? no q x ? ?c?? ? ? ? r m??t m s8? n ? t? a?r? ^ r ac???zB? ? ? | y u? c ?%? y???????? ?%?P???? ? ?E? ? y ? ??R? ?? ? ??v?l? ?? ???? ???? ? ? y ?[? ? y ??5yl?? ?:? ? ? ? ??v? ?1? ?? ? ???? ??? y??%?? ? ? ? ?y ? ? ? ? ? ??? ? ???[?????? ? ?? ? ???? ??1??? ? ? ?Y? ? ? ?? ??? ? y ?5? ? ? ? ?y ??? ? ?? ? ? ?? ?%? ? ?? ? ?B? ?#??>? ? ? ?[? ? ?? ? ?%?? ?5???? ???????? ? ? ? ? ? ? ? ?5?#??? ?? ? ??? ?? ?? ? ? ? ? ? y ? ? ?? ? ? ? y ? ? ? ? ? ? ? ?  ? ????  ??? ? y ?  ? ? ? B?   ? ? ? ?? ? ?? ? ? ? ? >? ?? ? ? ? ? ? ???  ?~? ? ?j? ?! ?   ? ? ? ? ? ?#" %$ ? ?[? y ? ? ? ? ? ?'& (*) ?,+ ? ? ? ? ?? ? ??? ? ? ?? ?[?>? . ? %???? ? / ?0 ? ? 1 ?%? ?R?.>?3?2 ? y ? ? y + ? ? ? ? ? ??? ? ?? ?? ? y4 ? ? ?? ? ? y ? - ? y ? ?2 ? ?6587 ( ? ? ?? ?:9 ? ; ? ? ? ? ? ? ? ?? >A@ ?? ? ?B @ ? ? ? ? y ? ? ? ? ?C ?EDGF ? & "5?3HJI ? ? y ? > ? ?? ?K#L ?NMPO ? ??RQ ?  ? ? ? ? ? ? ? ? ? 0 ?~? ? = < ? ? ? ? ? ? ? 0  ? ?? 1 /TS (  'o U ? ? ? V ?[?W ? ?X ? ?Y??Z ? ? ? ?? ? ? ? y ? ?N[ ? ? ??y ? ? ? ?]\ ?  ? ? ?^ ? ?_ ? ?? ? ? F ?E? ? G? ` ? ?[? ? ? ? acb ? y  ? ?d ? ?  ? O ? ? e ? ?  ? ??? ? ??? ? ? ? f ?[??? ?%j? ?? ? ?hgikjmli'npoq r ss i l%t l q r uwv ? ? ? ? ? ? ???5? ? ? x ? ? ? 9 ? ? cy y ?z ? ? |? {? y?y ? ? ? ? ? } ? gi'j li n l i'~8??Gj g 2 /TS ( ? ? y ?  ? ?  y ??? ? ??? ?? ?? ?P? ??? b ? ? ? 0 ? ? ? ? \ ? y b ?? ?  ? " ? ?? gi'j|??i n o q ??ss i'l?? l3? ? ~!?l???#? v q ? v|? ? ? ?  G? a ??? ? ? ;*? ?? ? Z ? ? ? y ? ? ?? ? ? ? ? ? ? ? ?[? y ? ? y ? ? ?  N? ? \ ? ? ? ?  ? b ??? ? ? ? ? y?? ?   ? ? S?? ? ? ? ? ? ? y?y ? ?? ? ? ?? ? ? ? ? ?? ?  - ??? ? ?B ? ? y ? ? ? ? ? ? - ? ? ? y ? R? ?? ? ?E? ???? ? ? ? ? ?[?R? ? ? ? ? y  ? ? ? ? ? ? o ? x:????? N? |? ?? ? y*? ? ? ? ? ? ?' ? ? ? `? ? ? ? ? ? ? y G? ? ? ? ? ? ? ? ?R?'? ? ?R???*? ? ?? ? ? ?  ? ?? ? *y ? ? ? ?E?? ? ? ?? ?'?? j'? ? j q ? ? ~?i|?R?? ? ?? \ ? ?y ? ? - ??  ? ? ? y ? ? ? ? 4 ? 8? ? y???< ? ? ?? ?e? y ? x ? ? Z :? ? ? y ? & ?d ? ? ? ? ? ? ? ? ? y ? ? y ? ? ?y ? ?? ? |? ? ? ? ? ? ? ? ? ? ??  ? ? ? y ?  @ ?? ?y ? ? ? y ? ? y ? ? ? ? ? ? ? ?? ? ? ? ? ? ? - 0 ? :? ? ? ? ?? ? Y ?  ?? ? ? ? ? ? 4?? ? ? ? ? ? ? ? ?  ? ? ? ? ? ? ? ? ? ?? ? ?? ? ? ?R?(? ? ??? ?  3? ? ? ?? ? ? ? ? ? ? ? 0| ? ?? ? ? ?%? ?f ? ? ? ?  ? ? ? ?  ? ? ? ? ? ? ? ? ? ?*? ? ?? ?  ? ? M ? y ? ??? ? y ?? y |? ? ? ? ? '? ?E? ? ? ?? ? ? ?? ? ? ? ? ?>?  " ?  ? ? y ? ? y ? ? ? ? ? ? ? ? ? ???? ?  ? ? ?[? ? ? ? ? y ? ? y ? ? ?:? ?? ??? ? ? ? y ? ? ? ? ? ? ? ?E?y ? ? a ? ? ? ? ? U ?? ? ? ? ? ? ?R? ?  y ?? ? ? ? ? ?? ? ?  ? ? ? ? ? ? y ? '? ? < > ? ? ??? '? ?>? -? ? ? ? ? ? ? y ? ? ?? ?? c? ?? ??? ?? ? ?? ? ? ? ?? ? ? ?B ??  ??  ? ? ? ?j?? ? - ? ? ?  ? ? ? ? ? ? ? ??? ? |? ? ? ? "??Y? ?e? ?? '? ? y ?? ?  ? ?R? ?  ? y ? )G?R???Ey ? ? ? ; ?? ?? ? ? ? ?  ?>?? ??? ? ?<? ?? ?d ? y ? ?  ? \ ??? ? ? ? ?? ? ? ? y ???? ?? ? ? ? y ? ? y ? ? ? ? ? ? ? ?? ? ? ? ? ?>? ? ? 6 ? ? ? ? ? y ?  G? ? ? ? ? ? ? ? ??- ? y ? ? ??? ? ? ? ? ? y ? ? ?R? ? ? ? ?z ??% ,-.        78-9 !"$#&%(')'+* ,-/.0132  '+4  65 :;=<?>A@B)CEDGFH<I)@KJ LNMPO=QR+MPQ+S=STVUW MX MAY+Z []\+S+^E_a`bVYcU de_ `feg QhZ i jk\mlonqpr sc\HXEZ]\Et]puwvxt t-M Yyr3z{Qh|}+~ Sxt- S?A?mQ Z `?e\+??e??-? ?]Y Z O{U ????? ?|Z?U ?  Y??? M??-\ t nq?HT Q?Z M?r scM M Y+Z t  S ^ _qM? O  X? \ t ? Z?Q?Y X QHt X z Q?|? ~?? t-?c??A? ? Z ???\H??? ? ? v=SSx\???M Z s{Q Z6??Y M Z?? \ t?} U ?e? | cn S \ ? M?3?? Z???t M M(? Q ^ M t-????` Y ?v r?H| \y?3? M Z?`? Z?U ??? p(O U ????VXcM?Y QhY? \ v ? S v r?? Q? M t ? O{M Y ? U ? X ?=??YN?=YmU ?VZ ? Q [?M XM Y  Z MX z ^N??? ? QyY X? ??? ??? r?r Mht n U ? YcQ+? ?q?+?????? j?? WP? |  ? l M ur ` ?  l{??? [ \+_ ` ?VYcSvHr?? r \ ? U ??X ? M?l? Y U ? r ? QHt M0X M lc? r]M?XN????? ? ? ??? Q YcX | ? Y?? M |?????  lc? ? [ ? n s{`? X=X{M Yq?=Y{`? ? ??Z \?\c?H? ? ? Z ? YHU ? r ??Q t?M X?M Y \ Z MX z ?a?P? ? ? ? U? XXcM???? Y U? Z?S t \ X vc| M ? ?+` ? \ ? Z S??+Z ????? ? ? ?H????)???c? ??? ? ? OcM t M ? ? ? ? ? ??? ??? ? ? ?{???? ? ?k? i  M t  ? `e? M l ? _ M?t ?M TVM n ` ? ?(` ` ? Z M ? scM  M ? ? ? `V?(Q  Z s MwT Mhn MY r \ QHl `? Y S? r S Q?Z Z ? ? M X r M ? ? ? r = X M  $ M S Q?Z Z t Y?Q U e? s ? U " Y | Ube\+Y ? Y ? z ^ ? ! \+U#eX ) ?&%(' ? ? ?*), + . /*021436547 8 Y M Y+Z t \ ? ? ?]? lc|Z?`#? Y?\Hl |w\ _ S Mr `9 r `b ?hM s U: X=X{M l?? Q^ Mh[ `;e??X M ? Ycp X=< ? $ >  L ?? ? )A@ ? Q r U deY ? M Y r t ?HS ^gV v { U ced M4f?M Y+Z?` ? ? | Z `? \ Y ?Q ^ ? p [ ? ? M O ?kj?M ? ?? ? ? ?) Uh?r P  [ M ? S?M|r Z ?P|\ Yi M | r U? \ Y ? ? [ ? n L l l ? @nm ) ?st ? lir q ? s M t ? ? ???? zT| }~ ?o? ? U ????? ? Z O U d?? ??? Yc| r `b \ Y z |  ? Y ? M??r `? \ Y ?? t?\^?G|\ _ S MAZ v ? ??6 ? ? ? \^? ? | \ Yi M |?Z]U? \ Y{? ? t ? X=M Tr Q t?? ? M ? Z c? ? v"=|wZ?`??\HY z \h??? ? | ? ?u l y p 0??N? ?"? u ????? s U? X?X MAY ? ? v ?xw l ? ?^{ ? ? ? ? ? ? z U jeY S? Z ?  ? ?H l ? @o y ? %(EGF ? ? P P ? n z M t \4V| \ _ ? M Z U Wkr?U X ??M O UWZY c `U ? r s M Y  X M Y\[cY^] _ ? a? ` O b ?D ? B 3&K MNM?? QhY XRQTS ? ? ? ?C B ? HJI ? ? ? ? O /?? ? ? u ??? %?? ? ?6? S X QhZ M t??+TVM ? | Q+Y z M ? ?mn?? H Q t U ? g M X?? -? \????V\ ? k? ???U ; t ? r ? ? \ t U? ? U ? jhM  U ???=? M Y??YHU  Z]? Z \  vZ?Sx? Z v { X MT Z?Q t v^???J? _qv ? r z M Y `? r ? ?  Y ? ?  ? ` _ U ??Y S=v r ? Yc`9?Z]? r \ | ? n ? Mr W` r U j?? M cU j?X XcM Y v Yc` r ???U Y 4 Q ? ? ?=`kr ` d \ l?Z  _ ?m??Z  Mi? lH|  [?S?\ [?Q r?MN? Q?? l ?) q????? ? ??? ? @???? ??? ?? ?? ? ? ?\?????i? l ? ? ? ? ? ?(?^? f O? ? ?? S Y ?AZ ? t??H?VM ? M?? ? ? ?Z O Q+Z U ?VY Q XX U j?Z??? \ YZ  Z s M M t]t H \ t _ U? Y `# n U b g QhZ `?  Y ) ? M Y+Z ? \+S+^ ? ? ??? ? M M X Uj ? z n U ?eY? ? n?? ? g M X \ tn?? U ? _ U ?x?  Y Z ? M |\{? t(? M \A?)Z?O M?? pAQ+t YHU ? Y ? z ? u O=Q+Y ? `?eY ? ? H O M????+t??4??? M ?? ? t ? z ??? ? u ??Z? 0 ??? ?s ? ?x?         !  ! " #%$'&)(+* - ,  .  / , ! "  45 4 0132 687:9<;>=?A@8BCD>D =FEHG EI JLKNM EPOQ?-RST U S VXWZY%[\ ]^<_%`:acb ^<dfehgi+jkdl `mPnoqp)rfst^vuwW ^3x ey:ozo dke|{hjk} a8~vo ep)j x??? { ?? e?tg??d?? d:???3???? ? ?}??o d:l3e? ? j:g ???f? m? ?<?L? ^|?c??????Q?3?Q?P?|?????? ??? ?3?? o eA? W ?)? g? ^3?Qa?o?:} ? rk?? d??3yo?} { j? o<x ? { o j ?%ot{d)?? y g? xQxltd ? d g? ez_k? _%?+? W?? W3????^?p r st^ j ? Wo { sklx?e yFr e ? a b g? d ??+b g? ????d)?No%dke{ j } ak? ? o?? rk? j??P??? ? \ ?L? W3bN?<?+? d:o3e?? j? ? ??_ ?? y)g+? e ? o ? e? ? o<W ? ? jb } r { lx??X?? e???:{h??+?t?? ?)r ? jt? ^%{ W g ? ? o3x?d:o3e ? j{ ? r {?? y g? ?h?%? ? e ?:?hot?8?d r ???? e^?j? ~ g? d g ~ ? b otd e|{?j } ? m r W b ? ???+? { d ?:? ? ok{ j  p)?? xQ?:o%d ? d g? ? W eot? ?:o<? ? jc? o ?? ^3? ? j } {?j x ? ? ^3? e ?{?? o%e W ? ? ov? rf? ? ep:o ? o u ? g? e Wv? ?|? ?Q?3???   ? ? ?  ?  j x o ? o { ?  ?oto { ? ? ?  ? ?g!#" W % m $ ' p:? ? xFx   l & ??? o ?? ? e3{?j )x (k? ox?ep)^ ? r *{ g! rkd ? o j ? ?? d } ? ,e +,-.! ` xwo%? ?jd)d o ?0/21354 dwW ^ ? r {hH? gO?+? ? ; ^?j  ?? ? =} <?> ? p g ? xQx:lt? ? ? j d ? o ? e g @ j d W g? W ? W ?xcejcbql<? ?3? { o y j ? ? ? g? ?ko dBA g1DCFEHG ? ^JI K9?7L 69?NM ? 85e: W ?1 Wz? }QP ? e|? ? e  j {z} {?jx u:? ? dR8e r {TS o ehW ? j ? { o ; e ? aU 6 p g?FV ? r { ?W r ? ? o g ? W ?:Who%x ? o ?%?NXZY ? ^Z?? e 8  ?%W ? b ^Ul d?j b Who ? s lx?gi ? j u {eo [ } l { ?fhg ^ dfeW ? y j? i p gf x)x:o d ? d ?? ? W } ??r a g ? d ? e kmlFnpoqsr } j T? t \^ [ ?hJl ] _? ^`#a  ? c  e _ d u g v jt? ? el ? x x w j y)?? ^ e y oA? Y { ? o s%r { g? rt? ? ^ m ? j b?} r { lx ? g? \e z ?j e3? ? ok{ ? d g? b }wj:{ e {| }~ ?e y g? xQ?:?^ ? ?  ? ? e 8 ? y:?? x x:l<d ? d ?q e ?? W?x \? d:\^ ? ? a ?Qd ?^ ?hW ? ?? K p l ? r { g W ? d ?? ^ ???e?? ?,?'j ? ? ?? ? ?? ?? ? ?? ?? ?s? ??? m? *? ? ?N? ? ? ?? ?? ? ? ? ?F? ???? ? z K ^ ? ^?? ^? ? ep:o?dFu)b? o { j ?? g? xQx l d ? d ?? eW m)? ? ? ? ?! W?r d?? s o { r ? o j<? o { r ??>ep:o ?j ?#? ? ^ ? e g? ?? W g??d ? j  ? e???>? ?? ? ? `? d u=? g ? e ? ? ? ?! e y e p ? ? ?  o { ? o ? y:?q x x:o ? uFd ?? e V ? ?\g r ? ? ? o e ? j:{, W<? ? rU?+?^xB?Z?m?\???Z??F? ? ? ? ?|??????\?e?m ; ?k? ?'?Qovj b ? ?<?? dw^3? ? ? p j W ? } of{?? j {hb rk? ? o ? g?h? ? ? o3W3 } o ? e e j e y o o ? {j { ? ??g? ? g? g  g?s? r ? e g?Lj ? g! ? ? jb?? ? l3el ?La?le? ?g? s _<?Lo ? ee j e? o jw{?g? ? .? d r ? d:o3e ? j{ ?tW ? ? e\? ? r ? r { ? o ? ?Qb ? ? { PN? 8 K ?? xQ?? ? ? ? g? / ? ??? ? ? <jfs<of{m ? o ? r ?8^ r W ?? ? ? ??s? ? lt{}){?oe e p:o ? o<p r ?wg? j { ? ? ?? ot{ d o ? d)o e ? j{  ? m ? o ? ? uW l e y o?? :H? ",?N? P ?_? W g? ? o ^q ? W b ?<?+??9? g! d r ? ? ? ?Pe y g? ? }Q{ j ? o WW ?f j ? e r g? d g? d ? r V b r ? ? d:o eT?? ? {?'?k{ ? z?T??Z? eo ? /?F? g ? l ^?+W??u W / e y oN? ? ? ? ?? ?  ? N? ? ? ) ?  ? ?%? ? ?U?s? ?  ? ? ? ?  ???   ;  O  ; T  ; K ?6 7  ;>= ?"! = K$#&% 9' S (?lv_*) ? ? o+  ? / pwlvbxe y j,?ej.-vd)^|e?? j ? ? g/ d $10 ?? ; p e y g2 ? e35476st^ g f d } ? ?m y g! x)x:o ?98k? ? ? j:<; ? ) u e L ? ^? e ? n ^ ? ^1= b }k?j ? ^ ` l d l\/CB j ? ? b uDe l [F_ ? e ?FE { ^ } ? jx ?? o 6 ? ^ r ? }:p r5G ^ e?IH /?e o ? W ??KJL ? ? A > @ M ? e jw? e}Fue ? d g V eW$WYX g! d ; l?e y o?x IZ\[ lf{?otd ? o G ^ e ? o ^dNep ^ WoX? oe ? ^ ? WA] g ^ ^ W$_N_ ? ? m ? ? LOg `bNQadPOc5R e f g P` S B PUg? T" ?ihej  ^ x . k [ ot{?^%d ?o ? o<e B lo<d j e y ^ ? ?L^e e ^ { W3? / y lWho ??^ eel ?h? r {o o [:}Q^ ; eml3xNej ?Ql ? jNg??on ^ ? W^ ? ^1 ? ehj??qW3b Y ? ?+o { dpFb ?F^ { j  pF? xQx ^??8? d g? ehW ? V qsr? ? ?? qstou l<dke{ j} a ? ? W ? l ?tg ? ?3?wl? ? ? ? ? r d ? ?? d ? e ?wv o } r { rkb?oeo {yx ?{z dke gw| j }5}~ ? W xQ^ ?%{ ^fYD l ` ? ? Y ? u r ?+? ??? ep:o } r ? r g ^3eo { ? r W8?/ d ? ? o ? ox ? V y ? $ ? g! ? ? ?? ? ? ? ? n????? a ?)W g? ? ? ? ??g gw? ?? ? p g? xo:? o ? ? d ^? e??fd r ? :} W ??+W ?? o j ? W ? o {?? ox j ?:? a?? p { ???? ?? ? j:{ yF?? x x ? ? ?d?5?? ehW?? j? o<d ???? j}5? b?^3e? ? ? x? 6 y u:W ? r?? ^ {hdoe?d:o3e $ j{_ ? W ?j b ? } j ? ? x j  e p {?o o ?%o {,?wo3? z ? ?? x???? nd ? d ?+? e W ? ? ? g ? Yk? ? g ` $ g e ? ? e - ?? { e ? 4 ? s<oz?F?q ?Qx:? ? Fp d?I! " W j  e y o jw{ ? ? ?? ? ? ? ?o3? ? j {  ??? ? ep o j ej?? ^<{ y Y ^ h ? ? ? a?L ? ?Z???? ? e k8 ?:` ? ? ` ? ? ? ??? } {?j } r ? _ e ?? j d m:b rk? a?p g? x x ofd ? d g@ e?W ?k{o r ? e g1 ? r%? o3x ? r d C ? ? p ? We 0 ^ g? ?t j:{ ? ?I? c? r? j ? L)}Qj d g? d)}f? e?} r ehe ? { d W r { ^ ? g? We { .? ? ? e o<? j s ot{ b?r d a?y ?! x x otd ? d g? ?W ?9? ??\? : ? ? ^ ? ? p P ? ? D P g ? ^? ? ?? d ? ?d)^e?? j { ? r dFxB^%{*d:l\? d oe ? ? {?  j?g ? e y  r ? e j o<d ?%j x)o { ? V W a j? ? ? ?cW ^ ?? {?jb e?)? ^sQ? ?ku ? l5? i y o'dF? b ? l { j  pF?\? xtg d ? d g? eWv?? ? xFo ; {?^ ? o ?  {?jb e y ?? { e a? ? s??\P? e j e 0 { o ^ ? ? u)W g ? ? R?e y ^v?t^ ? d:l ?? g` x)?ofd ? ? ??g ? e ? ?)? ? a W g +? W? ? m d a ? ? F ? E w D 3 3 ^ ^ L ? ^ ? ?g ? ? ? ?fg?f e y lzb?^%_td I? d ? j y g? ? x ofdc? d g? ? D m doe ? j:{, WA?'o { o ?j ? W ? ?$? ?I???  ?8j???I? P ? ?? " p ? ^ n g   ? y g? Q? ? o ? uFd g ?+eW m ?- ? ? e ? ??? j u /? ? e W m? ?<d:l { r e ^ ? ? ? e y o d ? e ? ?:{  W ? ? ? n ? Y ? ? ??? ?s? ? } ^? ?kb g? ? ? l ` ? ? ^ ? j L ??? W ^l e p Y e>e y:^ { ?5? l j  p g? ? xF^d ? dI?I? W ?jw? ? `vo$? ? ? g! ? ?? e ? a ? o x??$? ^{?  ^? d ^ ? ?         "! D &# & ')( *+ ,./1 - * 0 465 -67  $"% 32 8 <"= 96: ; #?> -?@ DEGF + / ' ' F ; H A 5JI KKQ 0 : S & NL MPO R L UT"V L * F KXW Q MXY * [ )- Z F /P\ H ] L ;d  ^"_ /a`bPc --e- @ L gf K / / QihkjelM Sk 0 hPm n o S p , ; Z / A L  kq s C    t_ u _ C v"w  Vkq H ,r r H x ? @ ? ? ? ? ?  ? ? ? ? y{z |~}?i?g?? ?????e???k???? ?X???? z?  z ?  ??? d ? ???????? ? ??? ? ? ?????? ?k ?? ? ? ? ?k? ? ?  ? ?z @ ? ;???? ??6? ?i?k?P? ? ? ? ??? ? ? ? ? ?6? ?? ? ? ??? ? ? ? ? ? ? ?? ? | ? ? A B"C GI\ H ? lb ?? ? ?N??? \ ? K ? ?? p ???? ? ?? ?h ?? 8 ?? ?? K ? ? ? ?? ?p ?? ?? ? p , m ?? ?? ? ? ? ? ?  ?J? ? ? ?? p ? W8 ? ? K ??? ? ? ? ? ? p ?  ? ? n? p p?~?? ? ?? ?J? ? J ?{? Z?? 0? ??{???k????? ? ? ? ?? E ?? / ?? 8 ? ?6???????? ? ? ?  ?    F ? ? Z?   ?      p ? p"!$# %'& ( %*) +,  -/.  ? ?21 ?iz 687kz Z ? ?? ?;: =< ? ? ? ?  ? : ?< ? ?  B ? ? ? ? ? C M0 Z354 ?9 ? ???> ? @ @? ?=A ? ? ? ? @ Z ?? ? ? ?D ? ? 8 ? ?                 !#"%$'&   )(*,+.-  '0/ 354687'9;:0<=;>?@Bq AC4DFEHGJI.KMLN K DFOPQ KRTS)UCV GDXWY9 S[Z5G D\J] ?l m fnpo ]rqtsvuw x KU#y RzbF{| } ]#b u 6~ KZ K? K???jM???K?U???0K D#S K R ?????!???? ? W8?M????? ?B???[?0?? ? ??'???8? ??'?T? ? ?8? ? ?8? ?? ? ? ? ?? ? ?? ? ??'? ??? ? ? ? ?8? ? ? ? ? R df ghiTjMk Z#^ _ SF`Ha8VcbFK 9 e f? R f q )b O ?4~??G k i O g b fP GTR?]? ??? ? ?%?*? ?T? ??? ????? ?????????????!???Y?? ? ? ??? ? ? ?0?v??? ?H? ?????????????????????c? ?*? ? ? ?? ??? ???.???c??? ?!?J? ???????T??? ? ? ?J?v?8? ? ?M? W ? ? ? ? ? W ?  ??  ? ? ? ?!"# ? ?%$ ?&(')*  ? + ,.-0/ 21 ? W43 W5 7? 6 9? 8  ? 87: ? ? ; ? ? : ? -< ? ? ?  3 ? $ ??W>= ? @? ? A ?C? ? ? ? ?CB2D? ? E ? GFIHKJ W E7L ?M ? N ? : ? ?OQP4RTSVUXW ? ? ? YZF E ?(:   W [ E 45 ? ? 87\ ? 5 ? ? ]^ _ ? ?4` ? ? ?  ? ? W 3 ? ?? 5b? a W?? [ ? ? ? ? W8?  ? F  ???cd2\ ?  ?feTg ? ?h ? ? 8 ? i 5j 3 ? ? ? 8 ? ?klm ? ???n ? ? ? 5 porq 8fs ?9t  ?uwv ` x ? u ?; ? ? ?zy ? ? U|{ ?} ?JW ? ?? ?h W~f?T? ??"?? ? ??? : ? ? ; ? W ? ? N ? ? ? \? ?? ? ? 3 5JWY? ?9? ? h m  ? ?? ? W ? ? V? ? A \ W : 5 : ? W?? ? ? ? ? ? $ 5?? F ??? SV!? ? ? \ ? ?? W43??f???2??? ? W ? 8?9? ? ? - W :? ? ? ? W ? ? ? ? ??4? ? D? ? j ? ` ? ? ?M? 9? ? ? ? ?  ? ?n ? ? W ~ ?? ?@? ? ? ?h ? W q o ? ? ? ` u? ? W :I?? %? ????? ??? ? ? W ? ?? ???????} ?|???? ? W?? ? ?O ? F ?  ? < ? 5 ? 8 ? ? - ?W : O ? ? ? ? ? ?? WY? W ? ? 5 ? ? ? z?  ?  ?8? ? ? ? ? ? ??W?? ? < ? ? ?  ? ? ? ?0??? ? ? ?? ? ? ? ? ?2?? ? ? ?"? ? ? ? ?? ? ?.? ? < ? ?8? ? ? ~ ? ? ?? ? ?K? ?? ? ? ?  ? ? ? W ?? W 3 ? ? ? ? W ? ? ? ? $ ?a ? ? ?H? ? W ?  f? ?.??? ?0? :I? \ ? ? ? ? ? u ? ? y%? W ??O z? ??? ? ? ? ? U ?? ? n ? ? :  :? ??? N ? ? ? ? W ? ??J? ? ? ? ? ? ? ?? ? ? ? ???? ?? ?? ? ? ` ?M4? ?b??? ? ? ? ? 8 ? ? ? ? W ? ? ? ?zu|?  ` ? U z?  ? ? ` ?  ? ? ? ?" ? ? y ? `?W ? E ? W ?T?#? D?? ? ~ F4? ?  y  ?  ` ? ?? <??? ? y ~ ? ` ~ - W ?? ? ? W I? ?E ?Zg2??? ?? ? W ? ? ? ? :D?? ? : ? 0? 3 ? ? ? ?V? ? ? W ? ? ? ? ?? 3 W ? ? ? ? 5???: ?  ? ` ? ?? ? ? `? ?G? U ? :  ? 8 ?? ?? W ?? P E ? ? ? ? < ? ? y ? W ? ? q o?? : ? \? N ?  ?  5 < ?n sK5  7? ? ? ? F? ? ? 8 ? ? ? ? ? W ?? ?  g ? ? ? ?    ? ? ?? W ? ? ?n  ? ? ? ?ry??? ?2P 8 ? ? "  ? ? U \ ?0 ? ? 3 ? ? ? W 5 ? W ? u \ ? ? ? ?  ? ? ? ?5 ? 8 W? W 6 : ? W@??, ?  ? ~ ?  ? ! ? ?? ? L 5"$#%'& ? W  & ? ? ? ? q ? - ? ? )? (? ? F? +W * ?? ??? ?-, 5/? . ? ? 0 ? ? ? \ ? ? 2? 1 ?M? `8? ? W ? -354 ?G? ? ? ?  W ? ADCE \? ? ??? ? ? ? ? ? q 8 ? ?G? G? F ? ? H ? ? @? ? ? ? ? ? 6 ? 8? 79 D:<; ?>=  ? ? ? ?  ? ? ? ? 8? @?  ? ? B n ? J \  ? ? ? U ?  ? ? 6 ? W ?  5 ?z`Y??? \?B F q o 5z?ML ? 8 ?NPO ?JW ? \ ? ? ? ` ? ? N W ?, ? ???~ ? W \ CW j I ? ? ` ? ? ?K ? ? ~ ? Q'RTS ? "E?? ? * ? ??? ? ? ? ? I?  ? ?h ??U ? Dn ? ?V*  W ? ?? ? ?W ? ? ? ? ?V`8? ? ?h W G? Z ? ? 5 :  W?? ? [ P/\^] ? ` ? _ ? Y ? X / ? + a  b c ? d ?gf ? ? q ? 5 ?ih|i ? ` ?? ? okj ? ? sl? ? C?nm ? ]?GoT? : ? 7qp t ?  5 ? \ ?5 ? W ? F U ? r ? y 5 ???sut ? ? j k? ? e ? \  ?? a ? v ? ? }~??? ?! ? ? ? ? ?? ??? < W `?W ? ? e : ? `+y ? D{z ? ?? ? ? ? ?#? ? e ? | ? ? ? ? ? ? . y ? %? ? ? e ? h ?%???> ?  ? ? ? wgx a ? ? ? ? ? y ? ? ? ?? o ? ? ?  ? ? \  ? 5 U W  ?W ? t ? ?? F ? \ ? ? ; W : 5'? ? ? ? 8 9? s ? ? ? 8? ? d ? ? ? ? ?8? U ? ? ???z? D{? ? ? ?? ?M? ? W ? ? u W \ ? ?n? : F?  ? ? ? 3 8 ? ? ? ?'5 ? ? ? ? ? O ? ? ? ? ? 8? u ? ? ? ? ? p W ? ? ? ?? ?? ? .)? ! $ ? ? 1 ? R ? h ? $ ? ? ? ? ? y y " ? ? ? a ? ? \ ? : ?D ? ? ? ?2? ?B? ? ? ? ? ? : ?5? ? 8 ? j ? ??? tV? 2? ? ? ? ? ? R : ? ? ? ? ? W ? ? ? ???? ?G? ? \ ?B??? ? 3 5 RY? g? ? ? ? ??? ? ? ?? 2? ? ? ? 8 ? ? >? \?W : ? 6 ? ? ? ? ? ? : ? ? ? ??? ? ? ? ? ? 8 ?h ? ? 8 ?? ? W ? ??? 2? ? ? W  5W ? ? ? ?G? ?? ? W \ ? ? {? E ? ? ? ?a ? 8 5'? ??? ?w??8 ? ? ? ??? ? ? ? ? ? n 8? ? ? ? ? ? ? ? ? ? ? ? 5 < ? ? ? ? ? o ? ? ?M?P? ? ?  ? ? ? ? ? ? ?8 rz? y?? ???V?? ? ?Z? ? b ? ? ?? ? E?? ? ? ? ? D? ? D?^? * \ R ? ? ? ? S ? ?75 < ?? 18? ? ? ? & ? ? ? ? ? ?!? ? ` ? ? t8 ? t ? kg? ? ?2? ?O ? 5~ y ? a " ? ? ? ??? ? ??  W ? , ? ? ? ? ?? ? F5?D O ? 8 ? z ? ?)? ? ? ? ? ? 5 8 ? ? a ? 8? ? ?h ? W { 5 8>? a 8 ? ? W ? ? ?v? ? ? ?$?q?e????B?2?? ??? p W ? ? g t? ?eLM[ a? ? ? 8 U W 5?? ? ?? 5 ? 8 ? ? ? N ?? ? ? ? W ? 2 ? F ? ? ??? ?8? ~ ? ~ ? W ? o q ? ?  ? ??  ????? ??? ? ? 8 ? ?7? t ? ? ~ ? ??? ?  ? t W ? [ k ? W \ ? ? W?~ 8 ?? ? ? g ? ?  ? ? W ? ? ! W05? 6 ?? ?  ??? ? ? : ? ? ? ? ?  ? ?T? \ : ? ? ? ? ?h ~ ? P 8z? ? ? ? ? W ? ? W ? ? ? ? 5 ? ??W ? \ ? ?? ? ?  p ? ? ? ? ?  ? 5 W\ ? ?!W ? ? = ? ?? ? ? ? W u<?h ?? ?{ ? ? ? U ? W5  ?? j &+? ?(`Y? ? O b ? ?!W ?Y? ?  8 ? ? { ?  ? b ? ? ? u ??? : ? ? ? ?? W ? u y ? ? ? ?? ? 8 W ? ? ? ` ? ? ?  ? SP?????{? W \  5 ?? ?Yb ? ?   ?n? r ??? ? * r ? ?!? ??? ??? ?" ?4? ? ? p ? ? ? ? : ? ? ? U ? ?E ? ?#? \ ? F ? ? - ?8??? ? ?  W ? ?+??? 5 ? ? ?i? S ?bo 1 ? ?  ?? ?8W ? ? ? ? 5 ? ? ?? ? ?5 ? `%W ? ? U? ? : \? ? ? ? ? ?M? ? ? ?? ??? ? W ` ? ?  ? ??W ? ? b ? 8??G? { q R @?5 ? ? y W ?E ? W? ? 8 ? ?2? W ?? W ? ? E? 122      !#"%$&' (*)+-,. /01 /2 1 3546 87:9 ;=<>@?BA CD; 4FEG ><IH%J%K5< A > 4L 1NM 2 1=O PQ1#R%4S&UTWVX 4YBZ\[ ^] < _a`b ' c d e 0fg4Uh A H%iBj A 4 " AlkWm $ n1 2 1poq rDsut M%vYwn4@x4 6 2@y ' z|{ 2B} O~  ?D4B???  6 ? 2 %&W4 ' ??K %?%? d ???? ? i??? Hn?B?? 7 ?  H? 1 G ` ? r 1 E <F' ? _ / A  > 4???B4B%4? 2B? O?? 2 [ ' ?   / $ A ? HA _ 2 &4 s 15E%4 ' ? B?l%A ??G 3 O ? r A ???n?n??? ??,? ???B? '? iD? [?? G 1 [ <?A  K? n?  ?Q?4??D,z m#? A `? ?D? 1 4 ??>@4 A m 4 ? 4 6 2B? E'? ? L ? ? ?? 1 K??? 4 E 2 >B4SG@?D?*? ? '? <? 15? ] < ? 1 M%O?|? i A???%? 4B_ s? 4I? 2 > 4  ? K 4 Al> 4 ? 1 M%? ? t:? ja? `? , ] ' ?*? 4?n1 A  @? f < 1 ??? ? }  ?  2 0%1 4 ? @? 4 6 ?\? ? O? { O ? ? ?l? 1N? $ ' ? ? 1 k A  G ? < 1 A  " @?? 4I& G@r %?1 G o ? rYm f 2 ? 4 A\ 4B? ; A5? m C & 2 ?? ????N????D?????????????5? ?  4  < ?  4 A    ? v?1?; A Bm-G 6 4 m ?m?_ 2B? ? M 2 1 ; 4a? 2 ?4 2 m ,  j ? n? v 1 2 ? 1 %4 ? ?,?%?Q s ? ( ? 4  s; <?M 2 > 4 2 [ %< *+ , i [ < ? 1  3 M v '  r } < 6  G  ?  ! 4 DAl4Wm <r 1 2 h O "   %A ? $ 1 ?  #6 m ? < E %G $ O&  'A  ? ) ? / 4 > 7 & E 2 %? 4S15. E - '/  t 4  x 2  - r 1 A  1 4 6n4%4 4 G ? d 89 TB[p,: ? <  =@?.ACBD E Fnk? 1  GIH1J  0 ? ?A 3[ 2 4 ' ? 5 r ; ? - 2 2 > > K  CL ? + BAl1 ? ,MON / @,/  1 C 1pE < 6 <  4A5G  ' P { 2 } ' ? H Q  P 'A R  ' S f Gn%&4& 2 r ? OT f / Al1 ]  $ 4? ? 1? U ,W? V XIY 6 <Gm oO? Z X ?  B4 A r  f 2 <  ge[  1 4 A _- ? ' 2 1 FO hl4 j1 i L 4lk m  1 E < O? ? [  6 [ \5T ? 4 r 1 A /  C ob ?A5< ??2 1 'z  ) m d 4 1?;  4 e j ]_^ `a r G ? ? j c ? ? G a n o   , ? !  tGB  ? 15G ' ? ? %< ul?  Ap? m ? ' 1 E!1#R4 6 %?a' ? %1?4 6 vwCx y < h G J [ {? z ,|1?? + p q 1 fr0 ? 4 ? 4 1 46N_FOs %4 ? C ? ) ] 2 ?? u E%S,O? }!~I<? H $n4? 6B4 < A 2 ? ` { 2 1 ' ?  Z? / 4@A ?l A _ 2  & < ? ???c?%???5?c?7????? ??3?7???? M 2 ? > 7o 8s????? J 2 & / A  i 2 ? 2 [ 7O  G%? 6  A '? 13? ? ? ; 7o [#E ?? [ ,? ? 2 ? ? m < ? M'a ???.? x ? ?  [?O? m sr? s o?? ?????!? ??? ? ?7? ? ??? ?3? ? ?@?<? ?? ? ?? ?@??? ? ? ???? ? ? ? ?p?? ? ??????B?? ?%? ? sc???????c?  ? Aj??? ? ?? ??? C<? ?<A C ?c???e? G 1 4????? ? s  i<?r? ?!?!? ???1? s?? ?1??? ? ?{? ? ????? A?A3A ??2  C ? ? ? M 4 ?????I? 2 f? L A `7 3 M ? ?  AS? 44? ?  A ? G? ? %4?1 ?. ? ??{?B,b }?? ? ,?? ??? 0 ?{? < ,z ? M ? K?C?????O? ? ?? ? ?c? ???? ? ? ? ?? ? ?1??? ?? ?? ? ? ?  ? ? ? ?? ( ? ? ? ? ? ?  ? ?  ? ? ? ?   ? A h { U 0 s g m C  ?f 7 s 2 r ? 2 1 4W ?? ? ? s / ? ? ?!#"%$& ?' s ? ? s ) '#*,+ 2@? `? _ ? A 2 s /? . 1 ? 02 '? f ?3 4n1 A L i j54 $176 8 ? O~   4 K  G ?  ? 1 ; :9<;@m s/= aO ??>A@ BC ? ? ?<D#E ? ? ?F ?GI?H- J J K ? LNMPO ? ?Q ? ??c ? Q ? R ? ? ? ?TS ? U ? ? ? ? U<? ? ?B?? ?WV ? X YZ ?#[ ?P\ ? ? ?s   ? ~ ? s /] ^ ?`_ ?(badcc e s fhgij kml n,oqp s? r Al /c C p ? ? E AduD<?? K K  ? G  ?wv ? ? 7 ? 4B6 t ? s /? x ? ? Bm51 ? ? & 3 Oa   yIOz?, ] }d{6d| \ G  st < ? z ? ? V ? ? ? ? ?#? ?W?b?:? ? ? ? ? ?  ? ? ? ?W? ?N? ? ? ? E ? ?  ? ?????/??? ??? C ??? O  <#} ? < A  [ G t ' ? ?  ? ? ? ??? ? ' z ?? ? ?  ????????  [ {%? j s ? ? K ? ~ s   ??? i C D ? G | N\ ? 2 1 4 :??? ??? s i / ~??  '%?? ?? C?? j ? %? ? + ?  ? ?  6 s x?x ? ? ? 4 f H > ' ? Z ? ? 1#GA 6 4 1F1 A5Gnos r%'t  6 2 ? ? %A oT 1 M f s ? s O?  ? ? ? V??? X ? ? ???!? ? ? ? ??? ? ??? ? ? ? s ?b?/??z ?#??? ? ??? ?'? ? ?   ? ?%? ? ? ? ? ? ?F ??? ?W? ?#? ? s ? ? ?  ? W  7 1 ? ? d 1 ? ? ? N m C  ' s?  ??? H ?? ? H A jns?? ??8??  / ? ?? C?? ? g ? p ?s? ?b? ?%? ? ? " ??? ? ? ?N? 0_ <#? ??? G 6?1 s ?q?N????? ?z n1 % G ? ? + | v P? ????,? ` ? '? H _ K s /? x?? $ ? A  o ? 6 '? ? 1d?????  , ? G ? '? ???#!?? ? ? ?N? ? ???   +"  C ? a ? E <<?< x ? ?B15? ?  r ?10 < ? ? ? A  O ?  / ? ? T1 O   O ]?C O  r ? ? ?  ? ? Y   ? Q ?  ?  ? Q ? E : ? ~ ~ ? ? # ?<% ? N $'&%fW? ? ? l? 1 C)( t ?+*1? ? + ? ? <-, ? G /. C ?Br ??1 M 410 N? MPO 0 4 mN ? ? A & M 6 ? iDs G32 ? ' 4 57s 6/8 9 $ns;:3<>=? @ A BDC ?+E A $FG ?Hb 15E ?JI 0? ? 5A ?s LK ? z ? C  " | ' ';Q? s?? ? " ? RTS , ? ?VU ? T ? ? GUr . ? x ? ? G@A'Z ?[? r 6 ?10 & ? ? H , ?*& <Q? ? ' ? n1 BO \  2 ? A ?/ 6 vm < ^ N ]'_`ab } ' ?   7d GYX ?  O e ? 7 ?+W ? Y o ? 3 q>rtsvu!: ? ( ? ? c;e f  $ 6 ?10 ? $ ' ?  2 _ ' ? G t5'  @ 2 i D~ A H 2 + E C ? ?hg ? ?ji>k Y ?l ?  ? ? ? ? ? s)m 3n ? s / / ? w M>x v)y{z [ | < J m%}-~5G 2 ) ? ? ? v ?^ ? ? H ? ? A 4 G 1 ot  6 2 A } ,???D&#2z?G ?  ? ? A 2 t  %4 1 ; A?? p K?? ?f ? H 1I? <P? ]?? ??? ? ? x x ? s s ?? G>?? ? ? < sT? ? ?? ?#? Y?? ?#[?? ? ? ? ? ?#s  1 ? C /  Q ? ? C ? (!?.3 ? ? ?P?
873 |@word h:1 cu:1 maz:1 mhn:1 c0:1 sex:1 km:4 d2:1 r:3 t_:1 dba:1 q1:2 tr:1 ld:3 n8:1 xiy:1 bc:1 od:1 si:1 dde:1 fn:1 j1:1 gv:1 yr:4 nq:1 rts:3 xk:1 dn:1 ra:1 ry:2 td:1 jm:1 pf:2 kg:1 sut:1 adc:1 elm:1 w8:6 qm:1 dfu:1 yn:3 ak:1 id:3 ap:1 kml:2 au:1 r4:1 co:1 fot:1 vu:1 lf:1 sxt:1 sq:1 w4:1 oqp:2 got:1 tsr:1 ga:1 py:1 paq:1 go:2 y8:1 gur:1 mq:1 dw:1 hd:1 fx:1 qh:1 gm:1 jk:1 ipo:1 cy:2 wj:1 mpn:1 oe:2 mz:1 yk:1 rq:1 kmi:1 ycx:1 ov:1 hjl:1 uh:1 mh:1 mht:1 wkr:1 kp:1 sc:2 mon:1 wg:1 gi:3 g1:1 ip:2 fr:1 tu:2 oz:2 az:1 qr:1 e_a:1 tk:2 oo:1 n_:1 qt:1 ex:5 rko:1 f4:1 vc:1 kb:1 ja:1 pfb:1 dne:1 ao:1 f1:1 mxy:1 cb:1 lm:1 m0:1 al1:1 vz:1 wl:1 qv:1 r_:1 pn:1 a3a:1 ihk:1 lj:1 fhg:2 iu:1 l3e:2 rop:1 yu:3 np:4 dg:1 m4:1 n1:5 ab:1 fd:1 a5:2 bpc:1 nl:1 hg:2 xb:1 fu:2 lh:1 bq:1 gn:1 ar:1 poq:1 tg:3 kq:1 acb:3 sv:1 st:5 ie:1 l5:1 yl:1 vm:1 nm:5 hn:1 f5:1 de:1 skl:1 jc:1 mp:1 ixk:1 zha:1 s15:1 b6:1 scm:2 om:1 who:3 vp:1 zy:1 fo:1 mwt:1 npo:4 ut:1 ea:2 ok:2 jw:6 yb:1 ox:1 dey:1 xzy:1 ei:1 fehg:1 yf:1 aj:1 b3:3 wp:1 jkd:1 i2:6 gw:2 ehj:1 oc:2 o3:1 ay:1 vo:2 tn:1 edgf:1 ilc:1 jlknm:1 qp:4 ji:4 qht:2 b4:2 s8:2 he:1 rkb:1 jey:1 ocm:1 hjk:1 nhk:1 l3:1 ikjml:1 gj:2 rkd:1 irq:1 wv:6 yi:7 tvm:2 ced:2 ey:3 eo:1 w3bn:1 xf:2 seh:1 a1:1 jns:1 ko:1 bz:1 c1:4 ot:8 sr:2 ycp:1 hb:2 w3:2 mpq:1 cn:1 br:2 qj:1 gb:1 ltd:1 k9:1 ul:2 f:1 e3:2 npe:1 vy:1 rb:1 hji:2 g32:1 iz:4 ibj:1 qhy:2 ht:1 kkq:1 hpm:1 ylz:1 v12:1 p3:2 ob:1 dy:1 k8:1 hy:1 qb:1 gmi:1 bdc:2 tv:1 kd:4 tw:1 n4:2 hl:1 mpo:2 feg:1 xo:1 kdi:1 ihe:1 hh:1 ge:1 b_:1 k5:1 rz:1 qsr:1 xw:2 yx:3 dfe:2 uj:1 rt:1 prq:1 hq:1 cw:2 oa:2 vd:1 s_:1 fy:1 ru:1 gji:1 ql:1 fe:2 xqx:4 ba:2 zt:1 sm:1 t:4 dc:1 lb:1 wyx:2 kxw:1 kl:2 z1:2 fv:1 hih:1 qa:1 wy:1 mnm:1 yc:3 fp:2 tb:2 rf:1 wz:2 brb:1 hot:1 eh:3 zr:4 mn:3 yq:1 ne:2 lk:2 hm:1 gf:5 kj:3 xq:4 kf:1 zh:1 rpit:1 otd:3 h2:4 cd:1 lo:1 fg:1 xn:2 kz:1 t5:1 bm:1 lz:1 qx:1 opp:1 lir:1 ml:3 b1:2 ehe:1 z6:1 jhm:1 sk:1 ku:1 f_:1 cl:1 uwv:2 t8:2 sp:1 pk:1 lc:3 n:1 sf:1 pe:4 hw:1 rk:4 zu:1 jt:2 badc:1 r2:2 ih:7 kr:1 o9:1 sx:2 mf:7 xcm:2 lt:2 fnk:1 g2:2 bo:2 a8:1 fw:1 zb:1 egf:2 la:1 ofi:1 d1:1 wcx:1
7,106
874
How to Choose an Activation Function H. N. Mhaskar Department of Mathematics California State University Los Angeles, CA 90032 hmhaska@calstatela.edu c. A. Micchelli IBM Watson Research Center P. O. Box 218 Yorktown Heights, NY 10598 cam@watson.ibm.com Abstract We study the complexity problem in artificial feedforward neural networks designed to approximate real valued functions of several real variables; i.e., we estimate the number of neurons in a network required to ensure a given degree of approximation to every function in a given function class. We indicate how to construct networks with the indicated number of neurons evaluating standard activation functions. Our general theorem shows that the smoother the activation function, the better the rate of approximation. 1 INTRODUCTION The approximation capabilities of feedforward neural networks with a single hidden layer has been studied by many authors, e.g., [1, 2, 5]. In [10], we have shown that such a network using practically any nonlinear activation function can approximate any continuous function of any number of real variables on any compact set to any desired degree of accuracy. A central question in this theory is the following. If one needs to approximate a function from a known class of functions to a prescribed accuracy, how many neurons will be necessary to accomplish this approximation for all functions in the class? For example, Barron shows in [1] that it is possible to approximate any function satisfying certain conditions on its Fourier transform within an L2 error of O(1/n) using a feedforward neural network with one hidden layer comprising of n 2 neurons, each with a sigmoidal activation function. On the contrary, if one is interested in a class of functions of s variables with a bounded gradient on [-1, I]S , 319 320 Mhaskar and Micchelli then in order to accomplish this order of approximation, it is necessary to use at least 0(11$) number of neurons, regardless of the activation function (cf. [3]). In this paper, our main interest is to consider the problem of approximating a function which is known only to have a certain number of smooth derivatives. We investigate the question of deciding which activation function will require how many neurons to achieve a given order of approximation for all such functions. We will describe a very general theorem and explain how to construct networks with various activation functions, such as the Gaussian and other radial basis functions advocated by Girosi and Poggio [13] as well as the classical squashing function and other sigmoidal functions. In the next section, we develop some notation and briefly review some known facts about approximation order with a sigmoidal type activation function. In Section 3, we discuss our general theorem. This theorem is applied in Section 4 to yield the approximation bounds for various special functions which are commonly in use. In Section 5, we briefly describe certain dimension independent bounds similar to those due to Barron [1], but applicable with a general activation function. Section 6 summarizes our results. 2 SIGMOIDAL-TYPE ACTIVATION FUNCTIONS In this section, we develop some notation and review certain known facts. For the sake of concreteness, we consider only uniform approximation, but our results are valid also for other LP -norms with minor modifications, if any. Let s 2: 1 be the number of input variables. The class of all continuous functions on [-1, IP will be denoted by C$. The class of all 27r- periodic continuous functions will be denoted by C$*. The uniform norm in either case will be denoted by II . II. Let IIn,I,$,u denote the set of all possible outputs of feedforward neural networks consisting of n neurons arranged in I hidden layers and each neuron evaluating an activation function (j where the inputs to the network are from R$. It is customary to assume more a priori knowledge about the target function than the fact that it belongs to C$ or cn. For example, one may assume that it has continuous derivatives of order r 2: 1 and the sum of the norms of all the partial derivatives up to (and including) order r is bounded. Since we are interested mainly in the relative error in approximation, we may assume that the target function is normalized so that this sum of the norms is bounded above by 1. The class of all the functions satisfying (or if the functions are periodic). In this this condition will be denoted by paper, we are interested in the universal approximation of the classes W: (and their periodic versions). Specifically, we are interested in estimating the quantity W: (2.1) W:'" sup En,l,$,u(f) JEW: where (2.2) En,l,$,u(f) := p E Anf n,l,s,1T III - PII? The quantity En,l,s ,u(l) measures the theoretically possible best order of approximation of an individual function I by networks with 11 neurons. We are interested How to Choose an Activation Function in determining the order that such a network can possibly achieve for all functions in the given class. An equivalent dual formulation is to estimate (2.3) En,l,s,O'(W:) := min{m E Z : sup Em,l,s,O'(f) ~ lin}. fEW: This quantity measures the minimum number of neurons required to obtain accuracy of lin for all functions in the class W:. An analogous definition is assumed for W:* in place of W: . Let IH~ denote the class of all s-variable trigonometric polynomials of order at most n and for a continuous function f, 27r-periodic in each of its s variables, E~(f):= min (2.4) PEIH~ Ilf - PII? We observe that IH~ can be thought of as a subclass of all outputs of networks with a single hidden layer comprising of at most (2n + 1)" neurons, each evaluating the activation function sin X. It is then well known that (2.5) Here and in the sequel, c, Cl, ... will denote positive constants independent of the functions and the number of neurons involved, but generally dependent on the other parameters of the problem such as r, sand (j. Moreover, several constructions for the approximating trigonometric polynomials involved in (2.5) are also well known. In the dual formulation, (2.5) states that if (j(x) := sinx then (2.6) It can be proved [3] that any "reasonable" approximation process that aims to approximate all functions in W:'" up to an order of accuracy lin must necessarily depend upon at least O(ns/r) parameters. Thus, the activation function sin x provides optimal convergence rates for the class W:*. The problem of approximating an r times continuously differentiable function f R s --+ R on [-1, I]S can be reduced to that of approximating another function from the corresponding periodic class as follows. We take an infinitely many times differentiable function 1f; which is equal to 1 on [-2,2]S and 0 outside of [-7r, 7rp. The function f1f; can then be extended as a 27r-periodic function. This function is r times continuously differentiable and its derivatives can be bounded by the derivatives of f using the Leibnitz formula. A function that approximates this 27r-periodic function also approximates f on [-I,I]S with the same order of approximation. In contrast, it is not customary to choose the activation function to be periodic. In [10] we introduced the notion of a higher order sigmoidal function as follows. Let k > O. We say that a function (j : R --+ R is sigmoidal of order k if ( 2.7) lim (j( x) - 1 x-+oo xk -, lim (j(x) - 0 x-+-oo xk - and (2.8) xE R. , 321 322 Mhaskar and Micchelli A sigmoidal function of order 0 is thus the customary bounded sigmoidal function. We proved in [10] that for any integer r ~ 1 and a sigmoidal function r - 1, we have (j of order if s = 1, if s > 2. (2.9) Subsequently, Mhaskar showed in [6] that if (j is a sigmoidal function of order k and r ~ 1 then, with I = O(log r/ log k)), >2 (2.10) Thus, an optimal network can be constructed using a sigmoidal function of higher order. During the course of the proofs in [10] and [6], we actually constructed the networks explicitly. The various features of these constructions from the connectionist point of view are discussed in [7, 8, 9]. In this paper, we take a different viewpoint. We wish to determine which activation function leads to what approximation order. As remarked above, for the approximation of periodic functions, the periodic activation function sin x provides an optimal network. Therefore, we will investigate the degree of approximation by neural net.works first in terms of a general periodic activation function and then apply these results to the case when the activation function is not periodic. 3 A GENERAL THEOREM In this section, we discuss the degree of approximation of periodic functions using periodic activation functions. It is our objective to include the case of radial basis functions as well as the usual "first. order" neural networks in our discussion. To encompass both of these cases, we discuss the following general formuation. Let s ~ d 2: 1 be integers and ?J E Cd?. We will consider the approximation of functions in ca. by linear combinat.ions of quantities of the form ?J(Ax + t) where A is a d x s matrix and t E Rd. (In general, both A and t are parameters ofthe network.) When d = s, A is the identity matrix and ?J is a radial function, then a linear combination of n such quantities represents the output of a radial basis function network with n neurons. When d = 1 then we have the usual neural network with one hidden layer and periodic activation function ?J. We define the Fourier coefficients of ?J by the formula (3.1) , 1 ?J(m) := (2 )d 7r 1 . ?J(t)e- zm .t dt, [-lI',lI']d Let (3.2) and assume that there is a set J co Itaining d x s matrices with integer entries such that (3.3) How to Choose an Activation Function where AT denotes the transpose of A. If d = 1 and ?(l) #- 0 (the neural network case) then we may choose S4> = {I} and J to be Z8 (considered as row vectors). If d = sand ?J is a function with none of its Fourier coefficients equal to zero (the radial basis case) then we may choose S4> zs and J {Is x s}. For m E Z8, we let k m be the multi-integer with minimum magnitude such that m = ATk m for some A = Am E J. Our estimates will need the quantities = (3.4) mn := = min{I?(km)1 : -2n::; m::; 2n} and (3.5) N n := max{lkml : -2n::; m < 2n} where Ikml is the maximum absolute value of the components of km. In the neural network case, we have mn = 1?(1)1 and N n = 1. In the radial basis case, N n = 2n. Our main theorem can be formulated as follows. THEOREM 3.1. Let s ~ d ~ 1, n ~ 1 and N ~ N n be integers, f E Cn , ?J E C d*. It is possible to construct a network (3.6) such that (3.7) In (3.6), the sum contains at most O( n S Nd) terms, Aj E J, tj E R d, and dj are linear functionals of f, depending upon n, N, <p. The estimate (3.7) relates the degree of approximation of f by neural networks explicitly in terms of the degree of approximation of f and ?J by trigonometric polynomials. Well known estimates from approximation theory, such as (2.5), provide close connections between the smoothness of the functions involved and their degree of trigonometric polynomial approximation. In particular, (3.7) indicates that the smoother the function ?J the better will be the degree of approximation. In [11], we have given explicit constructions of the operator Gn, N,4>. The formulas in [11] show that the network can be trained in a very simple manner, given the Fourier coefficients of the target function. The weights and thresholds (or the centers in the case of the radial basis networks) are determined universally for all functions being approximated . Only the coefficients at the output layer depend upon the function . Even these are given explicitly as linear combinations of the Fourier coefficients of the target function. The explicit formulas in [11] show that in the radial basis case, the operator G n ,N,4> actually contains only O( n + N)S summands. 4 APPLICATIONS In Section 3, we had assumed that the activation function ?J is periodic. If the activation function (J is not periodic, but satisfies certain decay conditions near 323 324 Mhaskar and Micchelli it is still possible to construct a periodic function for which Theorem 3.1 can be applied. Suppose that there exists a function 1j; in the linear span of Au,J := {(T( Ax + t) A E J, t E R d}, which is integrable on R d and satisfies the condition that 00, for some (4.1) T> d. Under this assumption, the function ( 4.2) 1j;0 (x):= L 1j;(x - 27rk) kEZ d is a 27r-periodic function integrable on [-7r, 7r]s. We can then apply Theorem 3.1 with 1j;0 instead of ?. In Gn,N,tjJo, we next replace 1j;0 by a function obtained by judiciously truncating the infinite sum in (4.2). The error made in this replacement can be estimated using (4.1). Knowing the number of evaluations of (T in the expression for '1/) as a finite linear combination of elements of Au,J, we then have an estimate on the degree of approximation of I in terms of the number of evaluations of (T. This process was applied on a number of functions (T. The results are summarized in Table 1. 5 DIMENSION INDEPENDENT BOUNDS In this section, we describe certain estimates on the L2 degree of approximation that are independent of the dimension of the input space. In this section, II? II denotes th(' L2 norm on [-1, I]S (respectively [-7r, 7r]S) and we approximate functions in the class S Fs defined by (5.1 ) SFtI := {I E C H : II/l1sF,s:= L li(m)l::; I}. mEZ' Analogous to the degree of approximation from IH~, we define the n-th degree of approximation of a function I E CS* by the formula En s(f) (5.2) := , inf ACZ' ,IAI~n III - L i(m)eimOxlI mEA where we require the norm involved to be the L2 norm. In (5.2), there is no need to assume that n is an integer. Let ? be a square integrable 27r-periodic function of one variable. We define the L2 degree of approximation by networks with a single hidden layer by the formula E~~~)f) := PEj~~l"'~ III - PII (.5.3) where m is the largest integer not exceeding n. Our main theorem in this connection is the following THEOREM 5.1. Let s 2: 1, integers n, N (5.4) 2: 1 be an integer, IE SFs , ? E Li and J(1) f:. O. Then, for How to Choose an Activation Function Table 1: Order of magnitude of En,l,s,o-(W:) for different O"S --- Function En Iso- 0' Remarks Sigmoidal, order r - 1 n 1/ r Sigmoidal, order r - 1 n lJ / r +(s+2r)/r 2 s ~ 2, d n IJ / r+ (2r+s )/2r k k ~ 2, s ~ 2, d xk, if x ~ 0, 0, if x (1 < O. + e-x)-l nlJ/r(log n)2 s=d=I,/=1 = 1, I = 1 = 1, I = 1 s~2,d=I,/=1 = Sigmoidal, order k n lJ / r k ~ 2, s ~ 1, d 1, I (log r/ log k)) exp( -lxl 2/2) n 2s / r s=d>2/=1 , Ixlk(log Ixl)6 n( IJ /r)(2+(3s+2r)/ k) =o = d > 2, k > 0, k + seven, 6 = 0 if s odd, 1 if s even, I = 1 S where {6 n } is a sequence of positive numbers, 0 ::; 6n ::; 2, depending upon f such that 6n --- 0 as n --- 00. Moreover, the coefficients in the network that yields (5.1,) are bounded, independent of nand N. We may apply Theorem 5.1 in the same way as Theorem 3.1. For the squashing activation fUllction, this gives an order of approximation O(n-l/2) with a network consisting of n(lo~ n)2 neurons arranged in one hidden layer. With the truncated power function x + (cf. Table 1, entry 3) as the activation function, the same order of approximation is obtained with a network with a single hidden layer and O(n1+1/(2k?) neurons. 6 CONCLUSIONS. We have obtained estimates on the number of neurons necessary for a network with a single hidden layer to provide a gi ven accuracy of all functions under the only a priori assumption that the derivatives of the function up to a certain order should exist. We have proved a general theorem which enables us to estimate this number 325 326 Mhaskar and Micchelli in terms of the growth and smoothness of the activation function. We have explicitly constructed networks which provide the desired accuracy with the indicated number of neurons. Acknowledgements The research of H. N. Mhaskar was supported in part by AFOSR grant 2-26 113. References 1. BARRON, A. R., Universal approximation bounds for superposition of a sigmoidal function, IEEE Trans. on Information Theory, 39. 2. CYBENKO, G., Approximation by superposition of sigmoidal functions, Mathematics of Control, Signals and Systems, 2, # 4 (1989), 303-314. 3. DEVORE, R., HOWARD , R. AND MICCHELLI, C.A., Optimal nonlinear approximation, Manuscripta Mathematica, 63 (1989), 469-478. 4. HECHT-NILESEN, R., Thoery of the backpropogation neural network, IEEE International Conference on Neural Networks, 1 (1988), 593-605. 5. HORNIK, K., STINCHCOMBE, M. AND WHITE, H ., Multilayer feedforward networks are universal approximators, Neural Networks, 2 (1989),359-366. 6. MHASKAR, H. N., Approximation properties of a multilayered feedforward artificial neural network, Advances in Computational Mathematics 1 (1993), 61-80. 7. MHASKAR, H. N., Neural networks for localized approximation of real functions, in "Neural Networks for Signal Processing, III", (Kamm, Huhn, Yoon, Chellappa and Kung Eds.), IEEE New York, 1993, pp. 190-196. 8. MHASKAR, H. N., Approximation of real functions using neural networks, in Proc. of Int. Conf. on Advances in Comput. Math., New Delhi, India, 1993, World Sci. Publ., H. P. Dikshit, C. A. Micchelli eds., 1994. 9. MHASKAR, H. N., Noniterative training algorithms for neural networks, Manuscript, 1993. 10. MHASKAR, H. N. AND MICCHELLI, C. A. , Approximation by superposition of a sigmoidal function and radial basis functions, Advances in Applied Mathematics, 13 (1992),350-373 . 11. MHASKAR, H. N. AND MICCHELLI, C. A., Degree of approximation by superpositions of a fixed function, in preparation. 12. MHASKAR, H. N. AND MICCHELLI, C. A., Dimension independent bounds on the degree of approximation by neural networks, Manuscript, 1993. 13. POGGIO, T. AND GIROSI, F., Regularization algorithms for learning that are equivalent to multilayer networks, Science, 247 (1990), 978-982.
874 |@word version:1 briefly:2 polynomial:4 norm:7 nd:1 km:2 contains:2 com:1 activation:29 must:1 girosi:2 enables:1 designed:1 xk:3 iso:1 provides:2 math:1 sigmoidal:17 height:1 constructed:3 manner:1 theoretically:1 multi:1 kamm:1 estimating:1 bounded:6 notation:2 moreover:2 what:1 z:1 every:1 subclass:1 growth:1 control:1 grant:1 positive:2 au:2 studied:1 co:1 universal:3 thought:1 radial:9 close:1 operator:2 mea:1 ilf:1 equivalent:2 center:2 regardless:1 truncating:1 notion:1 analogous:2 target:4 construction:3 suppose:1 element:1 satisfying:2 approximated:1 yoon:1 complexity:1 cam:1 trained:1 depend:2 upon:4 basis:8 various:3 thoery:1 describe:3 chellappa:1 artificial:2 outside:1 valued:1 say:1 gi:1 transform:1 ip:1 sequence:1 differentiable:3 net:1 zm:1 trigonometric:4 achieve:2 los:1 convergence:1 oo:2 develop:2 depending:2 ij:2 odd:1 minor:1 advocated:1 c:1 indicate:1 pii:3 subsequently:1 atk:1 sand:2 require:2 cybenko:1 practically:1 considered:1 deciding:1 exp:1 proc:1 applicable:1 superposition:4 largest:1 gaussian:1 aim:1 ax:2 indicates:1 mainly:1 contrast:1 am:1 dependent:1 lj:2 nand:1 hidden:9 comprising:2 interested:5 dual:2 denoted:4 priori:2 special:1 equal:2 construct:4 represents:1 ven:1 connectionist:1 few:1 individual:1 consisting:2 replacement:1 sinx:1 n1:1 interest:1 investigate:2 evaluation:2 tj:1 partial:1 necessary:3 poggio:2 desired:2 gn:2 entry:2 uniform:2 periodic:20 accomplish:2 international:1 ie:1 sequel:1 continuously:2 central:1 choose:7 possibly:1 conf:1 derivative:6 li:4 summarized:1 coefficient:6 int:1 explicitly:4 view:1 sup:2 capability:1 square:1 accuracy:6 yield:2 ofthe:1 none:1 itaining:1 pej:1 explain:1 ed:2 definition:1 mathematica:1 involved:4 remarked:1 pp:1 proof:1 proved:3 knowledge:1 lim:2 actually:2 manuscript:2 higher:2 dt:1 devore:1 iai:1 arranged:2 formulation:2 box:1 mez:1 nonlinear:2 aj:1 indicated:2 normalized:1 regularization:1 white:1 sin:3 during:1 yorktown:1 iin:1 discussed:1 approximates:2 backpropogation:1 smoothness:2 rd:1 mathematics:4 dj:1 had:1 summands:1 showed:1 belongs:1 inf:1 certain:7 watson:2 xe:1 approximators:1 integrable:3 minimum:2 determine:1 signal:2 ii:5 relates:1 mhaskar:14 encompass:1 smoother:2 smooth:1 lin:3 hecht:1 multilayer:2 ion:1 noniterative:1 contrary:1 integer:9 near:1 feedforward:6 iii:4 cn:2 knowing:1 judiciously:1 angeles:1 expression:1 fullction:1 f:1 york:1 remark:1 generally:1 sfs:1 s4:2 reduced:1 exist:1 estimated:1 threshold:1 concreteness:1 sum:4 place:1 reasonable:1 summarizes:1 bound:5 layer:10 sake:1 fourier:5 prescribed:1 min:3 span:1 jew:1 department:1 combination:3 em:1 lp:1 modification:1 discus:3 apply:3 observe:1 barron:3 customary:3 rp:1 denotes:2 ensure:1 cf:2 include:1 approximating:4 classical:1 micchelli:10 objective:1 question:2 quantity:6 usual:2 gradient:1 sci:1 seven:1 publ:1 neuron:17 howard:1 finite:1 truncated:1 extended:1 introduced:1 required:2 connection:2 nlj:1 california:1 delhi:1 trans:1 including:1 max:1 stinchcombe:1 power:1 mn:2 review:2 l2:5 acknowledgement:1 determining:1 relative:1 afosr:1 ixl:1 localized:1 degree:15 viewpoint:1 cd:1 ibm:2 squashing:2 row:1 course:1 lo:1 supported:1 transpose:1 india:1 absolute:1 dimension:4 evaluating:3 valid:1 z8:2 world:1 author:1 commonly:1 made:1 universally:1 functionals:1 approximate:6 compact:1 assumed:2 continuous:5 table:3 ca:2 hornik:1 cl:1 necessarily:1 main:3 multilayered:1 en:7 ny:1 n:1 wish:1 explicit:2 exceeding:1 comput:1 theorem:14 formula:6 rk:1 decay:1 exists:1 ih:3 f1f:1 magnitude:2 lxl:1 infinitely:1 satisfies:2 identity:1 formulated:1 replace:1 specifically:1 determined:1 infinite:1 kung:1 preparation:1
7,107
875
Efficient Computation of Complex Distance Metrics Using Hierarchical Filtering Patrice Y. Simard AT&T Bell Laboratories Holmdel, NJ 07733 Abstract By their very nature, memory based algorithms such as KNN or Parzen windows require a computationally expensive search of a large database of prototypes. In this paper we optimize the searching process for tangent distance (Simard, LeCun and Denker, 1993) to improve speed performance. The closest prototypes are found by recursively searching included subset.s of the database using distances of increasing complexit.y. This is done by using a hierarchy of tangent distances (increasing the Humber of tangent. vectors from o to its maximum) and multiresolution (using wavelets). At each stage, a confidence level of the classification is computed. If the confidence is high enough, the c.omputation of more complex distances is avoided. The resulting algorithm applied to character recognition is close to t.hree orders of magnitude faster than computing the full tangent dist.ance on every prot.ot.ypes . 1 INTRODUCTION Memory based algorithms such as KNN or Parzen windows have been extensively used in pattern recognition. (See (Dasal'athy, 1991) for a survey.) Unfortunately, these algorithms often rely 011 simple distances (such a<; Euclidean distance, Hamming distance, etc.). As a result, t.hey suffer from high sensitivity to simple transformations of the input patterns that should leave the classification unchanged (e.g. translation or scaling for 2D images). To make the problem worse, these algorithms 168 Efficient Computation of Complex Distance Metrics Using Hierarchical Filtering are further limited by extensive computational requirements due to the large number of distance computations. (If no optimization technique is used, the computational cost is given in equation 1.) computational cost ~ number of prototypes x dist.ance complexity (1) Recently, the problem of transformation sensitivity has been addressed by the introduction of a locally transformation-invariant metric, the tangent distance (Simard, LeCun and Denker, 1993). The basic idea is that instead of measuring the distance d(A, B) between two patterns A and B, their respective sets of transformations TA and TB are approximated to the first order, and the distance between these two approximated sets is computed. Unfortunately, the tangent distance becomes computationally more expensive as more transformations are taken into consideration, which results in even stronger speed requirements. The good news is that memory based algorithms are well suited for optimization using hierarchies of prototypes, and that this is even more true when the distance complexity is high. In this paper, we applied these ideas to tangent distance in two ways: 1) Finding the closest prototype can be done by recursively searching included subsets of the database using distances of increasing complexity. This is done by using a hierarchy of tangent distances (increasing the number of tangent vectors from 0 to its maximum) and l11ultiresolution (using wavelets). 2) A confidence level can be computed fm each distance. If the confidence in the classification is above a threshold early on, there is no need to compute the more expensive distances. The two methods are described in the next section. Their application on a real world problem will be shown in the result section. 2 FILTERING USING A HIERARCHY OF DISTANCES Our goal is to compute the distance from one unknown pattern to every prototype in a large database in order to determine which one is the closest. It is fairly obvious that some patterns are so different from each other that a very crude approximation of our distance can tell us so. There is a wide range of variation in computation time (and performance) depending on the choice of the distance. For instance, computing the Euclidean distance on n-pixel images is a factor 11/ k of the computation of computing it on k-pixels images. ? Similarly, at a given resolution, computing the tangent distance with 111 tangent vectors is (m + 1)2 times as expensive as computing the Euclidean distance (m = tangent vectors). This observations provided us wit.h a hierarchy of about a dozen different distances ranging in computation time from 4 multiply/adds (Euclidean distance on a 2 x 2 averaged image) to 20,000 multiply /adds (tangent distance, 7 tangent vectors, 16 x 16 pixel images). The resulting filtering algorithm is very straightforward and is exemplified in Figure 1. The general idea is to store the database of prototypes several times at different resolutions and with different tangent. vectors. Each of these resolutions and groups of tangent vectors defines a distance di . These distances are ordered in increasing 169 170 Simard Unknown Pattern ( Prototypes ~ 10~OOO Euc. Dist 2x2 Cost: 4 ~ Confidence Euc. Dist ~ {soc 4x4 Cost: 16 ~ Confidence ~ soo Tang.Dist Category 14 vectors t----t~ 16x16 Confidence Figure 1: Pattern recognition using a hierarchy of distance. The filter proceed from left (starting with the whole database) to right (where only a few prototypes remain). At each stage distances between prototypes and the unknown pattern are computed, sorted and the best candidate prototypes are selected for the next stage. As the complexity of the distance increases, the number of prototypes decreases, making computation feasible. At each stage a classification is attempted and a confidence score is computed. If the confidence score is high enough, the remaining stages are skipped . accuracy and complexity. The first distance dl is computed on all (1\0) prototypes of the database. The closest J\ 1 pat.terns are then selected and identified to the next stage. This process is repeated for each of the distances; i.e. at each stage i, the distance d i is computed on each J\i-l patterns selected by the previous stage. Of course, the idea is that as the complexity of the distance increases, the number of patterns on which this distance must be computed decreases. At the last stage, the most complex and accurate distance is computed on all remaining patterns to determine the classificat.ion. The only difficult part is to det.ermine the minimum I<i patterns selected at each stage for which the filtering does not decrease t.he overall performance. Note that if the last distance used is the most accurat.e distance, setting all J\j to the number of patterns in the database will give optimal performance (at the most expensive cost). Increasing I<i always improves the performance in the sense that it allows to find patterns that are closer for the next distance measure d j + 1 . The simplest way to determine I<i is by selecting a validation set and plotting t.he performance on this validation set as a function of !\j. The opt.imal !\?i is then determined graphically. An automatic way of computing each 1\; is currently being developed. This method is very useful when the performance is not degraded by choosing small J{j. In this case, the dist.ance evaluation is done using distance metrics which are relatively inexpensive to compute. The computation cost becomes: Efficient Computation of Complex Distance Metrics Using Hierarchical Filtering computational cost ~ L number of prototypes at stage i X distance complexity at stage i Curves showing the performance as a function of the value of the result section. 3 !{i (2) will be shown in PRUNING THE SEARCH USING CONFIDENCE SCORES If a confidence score is computed at each stage of the distance evaluation, it is possible for certain patterns to avoid completely computing the most expensive distances. In the extreme case, if the Euclidean distance between two patterns is 0, there is really no need to compute the tangent distance. A simple (and crude) way to compute a confidence score at a given stage i, is to find the closest prototype (for distance di ) in each of the possible classes. The distance difference between the closest class and the next closest class gives an approximation of a confidence of this classification. A simple algorithm is then to compare at stage i the confidence score Cip of the current unknown patt.ern p to a threshold ()j, and to stop the classification process for this pattern as soon as Cip > ()j. The classification will then be determined by the closest prototype at this stage. The computation time will therefore be different depending on the pattern to be classified. Easy patterns will be recognized very quickly while difficult. patterns will need to be compared to some of the prototypes using the most complex distance . The total computation cost is therefore: computational cost ~ L number of prototypes at stage i X distance complexity at. stage i X probabili ty to reach stage i (3 ) Note that if all ()j are high, the performance is maximized but so is the cost . We therefore wish to find the smallest value of Oi which does not degrade the performance (increasing (Jj a.lways improves the performance). As in the previous section, the simplest way to determine the optimal ()j is graphically with a validation set.. Example of curves representing the perfornlance as a function of ()j will be given in the result section . 4 4.1 CHOSING A GOOD HIERARCHY, OPTIMIZATION k-d tree Several hierarchies of distance are possible for optimizing the search process. An incremental nearest neighbor search algorithm based on k-d tree (Broder, 1990) was implemented . The k-d tree structure was interesting because it can potentially be used with tangent distance. Indeed, since the separating hyperplanes have n-1 dimension, they can be made parallel to many tangent vectors at the same time. As much as 36 images of 256 pixels ,,,ith each 7 t.angent. vectors can be separat.ed into two group of 18 images by Olle hyperplane which is parallel to all tangent 171 172 Simard vectors. The searching algorithm is taking some knowledge of the transformation invariance into account when it computes on which side of each hyperplane the unknown pattern is. Of course, when a leaf is reached, the full tangent. distance must be computed. The problem with the k-d tree algorithm however is that in high dimensional space, the distance from a point to a hyperplane is almost always smallel' than the distance between any pair of points. As a result, the unknown pattern must be compared to many prototypes to have a reasonable accuracy. The speed up factor was comparable to our multiresolution approach in the case of Euclidean distance (about 10), but we have not been able to obtain both good performance and high speedup with the k-d tree algorithm applied to tangent distance. This algorithm was not used in our final experiments. 4.2 Wavelets One of the main advantages of the multiresolution approach is that it is easily implemented with wavelet transforms (i\'1allat, 1989), and that in the wavelet space, the tangent distance is conserved (with orthonormal wavelet bases). Furthermore, the multiresolution decomposition is completely orthogonal to the tangent distance decomposition. In our experiment.s, the Haar transform was used. 4.3 Hierarchy of tangent distance Many increasingly accurate approximations can be made for the tangent distance at a given resolution. For instance, the tangent distance can be computed by an iterative process of alternative projections onto the tangent hyperplanes. A hierarchy of distances results, derived from the number of projections performed. This hierarchy is not very good because the initial projection is already fairly expensive. It is more desirable to have a better efficiency in the first stages since only few patterns will be left for the latter stages. Our most successful hierarchy consisted in adding tangent vectors one by one, on both sides. Even though this implies solving a new linear system at each stage, the computational cost is mainly dominated by computing dot products between tangent vectors. These dot-products are then reused in the subsequent stages to create larger linear systems (invol ving more tangent vectors). This hierarchy has the advantage that the first stage is only twice as expensive, yet much more accurate, than the Euclidean distance . Each subsequent stage brings a lot of accuracy at a reasonable cost. (The cost inCl'eases quicker toward the lat.er stages since solving the linear system grows with the cube of the number of tangent vector .) In addition, the last stage is exactly the full tangent distance. As we will see in section 5 the cost in the final stages is negligible. Obviously, the tangent vectors can be added in different order. \Ve did not try to find the optimal order. For character recognition application adding translations first, followed by hyperbolic deformations, the scalings, the thickness deformations and the rotations yielded good performance . Efficient Computation of Complex Distance Metrics Using Hierarchical Filtering z # of T.V. 0 1 2 3 4 5 6 7 0 0 0 1 2 4 6 8 10 12 14 9 10 8 Reso 4 16 64 64 256 256 256 256 256 256 256 # of prot.o (Ki) # of prod 9709 3500 500 125 50 45 25 15 10 5 5 1 1 1 2 5 7 9 11 13 15 17 Probab 1.00 1.00 1.00 0.90 0.60 0.40 0.20 0.10 0.10 0.05 0.0.5 # of mul/add 40,000 56,000 32,000 14,000 40,000 32,000 11,000 4,000 3,000 1,000 1,000 Table 1: Summary computation for the classification of 1 pattern: The first column is the distance index, the second column indicates the number of tangent vector (0 for the Euclidean distance), and the third column indicates the resolution in pixels, the fourth is J{j or the number of prototypes on which the distance di must be computed, the fifth column indicat.es the number of additional dot products which must be computed to evaluate distance di, the sixth column indicates the probability to not skip that stage after the confidence score has been used, and the last column indicates the total average number of multiply-adds which must be performed (product of column 3 to 6) at each stage. 4.4 Selecting the k closests out of N prototypes in O(N) In the multiresolution filter, at the early stages we must select the k closest prototypes from a large number of protot.ypes. This is problematic because the prototypes cannot be sorted since O( N ZagN) is expensive compared to computing N distances at very low resolution (like 4 pixels). A simple solution consist.s in using a variation of "quicksort" or "finding the k-t.h element" (Aho, Hopcroft and Ullman, 1983), which can select the k closests out of N prototypes in O(N). The generic idea is to compute the mean of the distances (an approximation is actually sufficient) and then to split the distances into two halves (of different sizes) according to whether they are smaller or larger than the mean distance. If they are more dist.ances smaller than the mean than k, the process is reiterat.ed on the upper half, ot.herwise it is reiterated on the lowel' half. The process is recursively executed until there is only one distance in each half. (k is then reached and all the k prototypes in the lower halves are closer to the unknown pattei'll than all the N - ~~ prototypes in the upper halves.) Note that. the elements are not sorted and t.hat only t.he expected t.ime is O(N), but this is sufficient for our problem. 5 RESULTS A simple task of pattern classification was used to test the filtering. The prototype set and the test set consisted l'especti vely of 9709 and 2007 labeled images (16 by 16 pixels) of handwritten digit.s. The prot.otypes were also averaged t.o lower 173 174 Simard 5 8 71 t Error in % Error in % 4 / Resolution 16 pIXels 6 ResolutIOn 64 pixels 5 Reso lutlO n 64 pixels 1 tangent veC10r Resol ution 16 pixels 3 4 3 2~~~~~~ o K (in 1000) 10 20 30 40 __~__~~~~ 50 60 70 80 90 100 % of pat. kept. Figure 2: Left: Raw error performance as a function of Kl and 1\2. The final chosen values were J{ 1 = 3500 and [\'2 = 500. Right: Raw error as a function of the percentage of pattern which have not exceeded the confidence threshold Oi. A 100% means all the pattern were passed to the next stage. resolutions (2 by 2, 4 by 4 and 8 by 8) and copied to separate databases . The 1 by 1 resolution was not useful for anything. Therefore the fastest distance was the Euclidean distance on 2 by 2 images, while the slowest distance was the full tangent distance with 7 tangent vectors for both the prototype and the unknown pattern (Simard, LeCun and Denker, 1993). Table 1 summarizes the results. Several observations can be made. First, simple distance metrics are very useful to eliminate large proportions of Pl"Ototypes at no cost in performances. Indeed the Euclidean distance computed on 2 by 2 images can remove 2 third of the prototypes. Figure 2, left, shows the performance as a function of J{l and 1\2 (2 .5 % raw error was considered optimal performance). It can be noticed that for J{j above a certain value, the performance is optimal alld c.onstant. The most complex distances (6 and 7 tangent vectors on each side) need only be computed for 5% of the prototypes. The second observation is that the use of a confidence score can greatly reduce the number of distance evaluations in later stages. For instance the dominant phases of the computation would be with 2, 4 and 6 tangent vectors at resolution 256 if there were not reduced to 60%, 40% and 20% respectively using the confidence sc.ores. Figure 2, right, shows the raw error performance as a function of the percentage of rejection (confidence lower than OJ) at stage i. It can be noticed that above a certain threshold, the performance are optimal and constant . Less than 10% of the unknown patterns need the most. complex distances (5, 6 and 7 tangent vectors on each side), to be comput.ed. Efficient Computation of Complex Distance Metrics Using Hierarchical Filtering 6 DISCUSSION Even though our method is by no way optimal (the order of the tangent vector can be changed, intermediate resolution can be used, etc ... ), the overall speed up we achieved was about 3 orders of magnitude (compared with computing the full tangent distance on all the patterns). There was no significant decrease in performances. This classification speed is comparable with neural network method, but the performance are better with tangent distance (2.5% versus 3%). Furthermore the above methods require no learning period which makes them very attractive for application were the distribution of the patterns to be classified is changing rapidly. The hierarchical filtering can also be combined with learning the prototypes using algorithms such as learning vector quantization (LVQ). References Aho, A. V., Hopcroft, J. E., and Ullman, J. D. (1983). Data Structure and Algorithms. Addison- \V'esley. Broder, A. J. (1990). Strategies for Efficient Incremental Nearest Neighbor Search. Pattern Recognition, 23: 171-178. Dasarathy, B. V. (1991). Nearest Neighbor (NN) Norms: NN Pattern classification Techniques. IEEE Computer Society Press, Los Alamitos, California. Mallat, S. G. (1989). A Theory for I\,I ultiresolution Signal Decomposition: The Wavelet Representation. IEEE Transactions 011 Pattern Analysis and Machine Intelligence, 11, No. 7:674-693. Simard, P. Y., LeCun, Y., and Denker, J. (1993). Efficient Pattern Recognition Using a New Transformation Distance. In Neural Information Processing Systems, volume 4, pages 50-58, Sa.n Mateo, CA. 175
875 |@word norm:1 stronger:1 proportion:1 reused:1 decomposition:3 recursively:3 initial:1 complexit:1 score:8 selecting:2 current:1 yet:1 must:7 subsequent:2 remove:1 half:6 selected:4 leaf:1 intelligence:1 ith:1 hyperplanes:2 expected:1 indeed:2 dist:7 window:2 increasing:7 becomes:2 provided:1 developed:1 finding:2 transformation:7 nj:1 every:2 exactly:1 prot:3 negligible:1 dasarathy:1 twice:1 mateo:1 fastest:1 limited:1 range:1 averaged:2 lecun:4 ance:3 euc:2 digit:1 bell:1 hyperbolic:1 ving:1 projection:3 confidence:19 onto:1 close:1 cannot:1 optimize:1 straightforward:1 graphically:2 starting:1 survey:1 resolution:12 wit:1 orthonormal:1 searching:4 variation:2 hierarchy:13 mallat:1 element:2 expensive:9 recognition:6 approximated:2 database:9 labeled:1 quicker:1 probabili:1 news:1 aho:2 decrease:4 complexity:8 solving:2 efficiency:1 completely:2 easily:1 hopcroft:2 sc:1 tell:1 choosing:1 larger:2 knn:2 transform:1 patrice:1 final:3 obviously:1 advantage:2 product:4 rapidly:1 multiresolution:5 los:1 requirement:2 incremental:2 leave:1 depending:2 nearest:3 sa:1 soc:1 implemented:2 skip:1 implies:1 filter:2 require:2 really:1 opt:1 pl:1 considered:1 early:2 smallest:1 currently:1 create:1 always:2 avoid:1 derived:1 indicates:4 mainly:1 slowest:1 greatly:1 skipped:1 sense:1 nn:2 eliminate:1 pixel:11 overall:2 classification:11 fairly:2 cube:1 x4:1 few:2 ime:1 ve:1 phase:1 cip:2 multiply:3 evaluation:3 extreme:1 separat:1 accurate:3 closer:2 respective:1 orthogonal:1 vely:1 tree:5 euclidean:10 deformation:2 instance:3 column:7 measuring:1 cost:15 subset:2 successful:1 thickness:1 combined:1 broder:2 sensitivity:2 eas:1 parzen:2 quickly:1 worse:1 simard:8 ullman:2 account:1 performed:2 try:1 lot:1 later:1 reached:2 parallel:2 alld:1 oi:2 accuracy:3 degraded:1 maximized:1 handwritten:1 raw:4 classified:2 reach:1 ed:3 sixth:1 inexpensive:1 ty:1 obvious:1 di:4 hamming:1 stop:1 knowledge:1 improves:2 actually:1 exceeded:1 ta:1 ooo:1 done:4 though:2 furthermore:2 stage:34 until:1 defines:1 brings:1 grows:1 consisted:2 true:1 laboratory:1 attractive:1 ll:1 anything:1 image:10 ranging:1 consideration:1 recently:1 rotation:1 volume:1 he:3 significant:1 automatic:1 similarly:1 dot:3 indicat:1 etc:2 add:4 base:1 dominant:1 closest:9 optimizing:1 store:1 certain:3 conserved:1 minimum:1 additional:1 recognized:1 determine:4 period:1 signal:1 full:5 desirable:1 faster:1 basic:1 metric:8 achieved:1 ion:1 addition:1 ore:1 addressed:1 ot:2 intermediate:1 split:1 enough:2 easy:1 ution:1 fm:1 identified:1 reduce:1 idea:5 prototype:31 det:1 whether:1 passed:1 suffer:1 proceed:1 jj:1 useful:3 transforms:1 extensively:1 locally:1 category:1 simplest:2 reduced:1 percentage:2 problematic:1 patt:1 group:2 olle:1 threshold:4 changing:1 kept:1 fourth:1 almost:1 reasonable:2 summarizes:1 holmdel:1 scaling:2 comparable:2 ki:1 followed:1 copied:1 yielded:1 x2:1 dominated:1 speed:5 relatively:1 ern:1 speedup:1 according:1 remain:1 smaller:2 increasingly:1 character:2 making:1 ypes:2 invariant:1 taken:1 computationally:2 equation:1 allat:1 addison:1 denker:4 hierarchical:6 generic:1 alternative:1 hat:1 hree:1 remaining:2 lat:1 reiterated:1 society:1 unchanged:1 noticed:2 already:1 added:1 alamitos:1 strategy:1 lowel:1 distance:93 separate:1 separating:1 degrade:1 toward:1 index:1 difficult:2 unfortunately:2 executed:1 potentially:1 reso:2 unknown:9 upper:2 observation:3 pat:2 pair:1 kl:1 extensive:1 california:1 able:1 pattern:35 exemplified:1 tb:1 oj:1 memory:3 soo:1 rely:1 haar:1 imal:1 representing:1 improve:1 probab:1 tangent:44 interesting:1 filtering:10 versus:1 validation:3 sufficient:2 plotting:1 translation:2 course:2 summary:1 changed:1 last:4 soon:1 side:4 wide:1 neighbor:3 taking:1 fifth:1 curve:2 dimension:1 world:1 computes:1 made:3 avoided:1 transaction:1 pruning:1 ances:1 quicksort:1 ermine:1 search:5 iterative:1 prod:1 table:2 nature:1 ca:1 complex:10 did:1 main:1 whole:1 repeated:1 mul:1 chosing:1 x16:1 wish:1 comput:1 candidate:1 crude:2 third:2 wavelet:7 tang:1 dozen:1 showing:1 er:1 dl:1 consist:1 quantization:1 adding:2 magnitude:2 rejection:1 suited:1 ordered:1 hey:1 goal:1 sorted:3 classificat:1 lvq:1 feasible:1 included:2 determined:2 hyperplane:3 total:2 invariance:1 e:1 attempted:1 select:2 incl:1 tern:1 latter:1 evaluate:1
7,108
876
Illumination-Invariant Face Recognition with a Contrast Sensitive Silicon Retina Joachim M. Buhmann Rheinische Friedrich-Wilhelms-U niversitiit Institut fUr Informatik II, RomerstraBe 164 0-53117 Bonn, Germany Martin Lades Ruhr-Universitiit Bochum Institut fiir Neuroinformatik 0-44780 Bochum, Germany Frank Eeckman Lawrence Livermore National Laboratory ISCR, P.D.Box 808, L-426 Livermore, CA 94551 Abstract Changes in lighting conditions strongly effect the performance and reliability of computer vision systems. We report face recognition results under drastically changing lighting conditions for a computer vision system which concurrently uses a contrast sensitive silicon retina and a conventional, gain controlled CCO camera. For both input devices the face recognition system employs an elastic matching algorithm with wavelet based features to classify unknown faces. To assess the effect of analog on -chip preprocessing by the silicon retina the CCO images have been "digitally preprocessed" with a bandpass filter to adjust the power spectrum. The silicon retina with its ability to adjust sensitivity increases the recognition rate up to 50 percent. These comparative experiments demonstrate that preprocessing with an analog VLSI silicon retina generates image data enriched with object-constant features. 1 Introdnction Neural computation as an information processing paradigm promises to enhance artificial pattern recognition systems with the learning capabilities of the cerebral cortex and with the 769 770 Buhmann, Lades, and Eeckman adaptivity of biological sensors. Rebuilding sensory organs in silicon seems to be particularly promising since their neurophysiology and neuroanatomy, including the connections to cortex, are known in great detail. This knowledge might serve as a blueprint for the design of artificial sensors which mimic biological perception. Analog VLSI retinas and cochleas, as designed by Carver Mead (Mead, 1989; Mahowald, Mead, 1991) and his collaborators in a seminal research program, will ultimately be integrated in vision and communication systems for autonomous robots and other intelligent information processing systems. The study reported here explores the influence of analog retinal preprocessing on the recognition performance of a face recognition system. Face recognition is a challenging classification task where object inherent distortions, like facial expressions and perspective changes, have to be separated from other image variations like changing lighting conditions. Preprocessing with a silicon retina is expected to yield an increased recognition rate since the first layers of the retina adjust their local contrast sensitivity and thereby achieve invariance to variations in lighting conditions. Our face recognizer is equipped with a silicon retina as an adaptive camera. For comparison purposes all images are registered simultaneously by a conventional CCD camera with automatic gain control. Galleries with images of 109 different test persons each are taken under three different lighting conditions and two different viewing directions (see Fig. 1). These different galleries provide separate statistics to measure the sensitivity of the system to variations in light levels or contrast and image changes due to perspective distortions. Naturally, the performance of an object recognition system depends critically on the classification strategy pursued to identify unknown objects in an image with the models stored in a database. The matching algorithm selected to measure the performance enhancing effect of retinal preprocessing deforms prototype faces in an elastic fashion (Buhmann et aI., 1989; Buhmann et al., 1990; Lades et al., 1993). Elastic matching has been shown to perform well on the face classification task recognizing up to 80 different faces reliably (Lades et al., 1993) and in a translation, size and rotation invariant fashion (Buhmann et aI., 1990). The face recognition algorithm was initially suggested as a simplified version of the Dynamic Link A rchitecture (von der Malsburg, 1981), an innovative neural classification strategy with fast changes in the neural connectivity during recognition stage. Our recognition results and conclusions are expected to be qualitatively typical for a whole range of face/object recognition systems (Turk, Pentland, 1991; Yuille, 1991; Brunelli, Poggio, 1993), since any image preprocessing with emphasis on object constant features facilitates the search for the correct prototype. 2 The Silicon Retina The silicon retina used in the recognition experiments models the interactions between receptors and horizontal cells taking place in the outer plexiform layer of the vertebrate retina. All cells and their interconnections are explicitly represented in the chip so that the following description simultaneously refers to both biological wetware and silicon hardware. Receptors and horizontal cells are electrically coupled to their neighbors. The weak electrical coupling between the receptors smoothes the image and reduces the influence of voltage offsets between adjacent receptors. The horizontal cells have a strong lateral electrical coupling and compute a local background average. There are reciprocal excitatory-inhibitory synapses between the receptors and the horizontal cells. The horizontal cells use shunting inhibition to adjust the membrane conductance of the receptors and Illumination-Invariant Face Recognition with a Contrast Sensitive Silicon Retina thereby adjust their sensitivity locally. This feedback interaction produces an antagonistic center/surround organization of receptive fields at the output The center is represented by the weakly coupled excitatory receptors and the surround by the more strongly coupled inhibitory horizontal cells. The center/surround organization removes the average intensity and expands the dynamic range without response compression. Furthennore, it enhances edges. In contrast to this architecture, a conventional CCD camera can be viewed as a very primitive retina with only one layer of non-interacting detectors. There is no DC background removal, causing potential over- and underexposure in parts of the image which reduces the useful dynamic range. A mechanical iris has to be provided to adjust the mean luminance level to the appropriate setting. Since cameras are designed for faithful image registration rather than vision, on-chip pixel processing, if provided at all, is used to improve the camera resolution and signal-to-noise ratio. Three adjustable parameters allow us to fine tune the retina chip for an object recognition experiment: (i) the diffusivity of the cones (ii) the diffusivity ofthe horizontal cells (iii) the leak in the horizontal cell membrane. Changes in the diffusivities affect the shape of the receptive fields, e.g., a large diffusivity between cones smoothes out edges and produces a blurred image. The other extreme of large diffusivity between horizontal cells pronounces edges and enhances the contrast gain . The retina chip has a resolution of 90 x 92 pixels, it was designed by (Boahen, Andreou, 1992) and fabricated in 2flm n-well technology by MOSIS. 3 Elastic Matching Algorithm for Face Recognition Elastic matching is a pattern classification strategy which explicitly accounts for local distortions. A prototype template is elastically deformed to measure local deviations from a new, unknown pattern. The amount of deformation and the similarity oflocal image features provide us with a decision criterion for pattern classification. The rubbersheet-like behavior of the prototype transformation makes elastic matching a particularly attractive method for face recognition where ubiquitous local distortions are caused for example by perspective changes and different facial expressions. Originally, the technique was developed for handwritten character recognition (Burr, 1981). The version of elastic matching employed for our face recognition experiments is based on attributed graph matching. A detailed description with a plausible interpretation in neural networks terms is published in (Lades et al., (993). Each prototype face is encoded as a planar graph with feature vectors attached to the vertices of the graph and metric information attached to the edges. The feature vectors extract local image information at pixel Xi in a multiscale fashion, i.e., they are functions of wavelet coefficients. Each feature vector establishes a correspondence between a vertex i of a prototype graph and a pixel Xi in the image. The components of a feature vector are defined as the magnitudes of the convolution of an image with a set of two-dimensional, DC free Gaussian kernels centered at pixel Xi. The kernels with the form 1/!'k (X) fl exp = (72 (flx2 - ) - exp (-(72/2) 1 - 2(72 ) [exp (ikX (I) are parameterized by the wave vector k defining their orientations and their sizes. To construct a self-similar set of filter functions we select eight different orientations and five 771 772 Buhmann, Lades, and Eeckman different scales according to k(v, tt) = ~ Tv/2 (cos( itt), sin( itt)) (2) with v E {O, ... ,4};tt E {O, ... , 7}. The multi-resolution data format represents local distortions in a robust way, i.e., only feature vectors in the vicinity x of an image distortion are altered by the changes. The edge labels encode metric information, in particular we choose the difference vectors AXij == Xi - Xj as edge labels. To generate a new prototype graph for the database, the center of a new face is determined by matching a generic face template to it. A 7 x 10 rectangular grid with 10 pixel spacing between vertices and edges between adjacent vertices is then centered at that point. The saliency of image points is taken into account by deforming that generic grid so that each vertex is moved to the nearest pixel with a local maximum in feature vector length. The classification of an unknown face as one of the models in the database or its rejection as an unclassified object is achieved by computing matching costs and distortion costs. The matching costs are designed to maximize the similarity between feature vector J;M of vertex i in the model graph (M) and feature vector Jl (Xi) associated with pixel Xi in the new image (I). The cosine of the angle between both feature vectors -[...., S(JI(x) jM) = '" -M J (Xi) . Ji (3) M Ilf (Xi)IIIIJ; II 1 is suited as a similarity function for elastic matching since global contrast changes in images only scale feature vectors but do not rotate them. Besides maximizing the similarity between feature vectors the elastic matching algorithm penalizes large distortions. The distortion cost term is weighted by a factor ,\ which can be interpreted as a prior for expected distortions. The combined matching cost function which is used in the face recognition system compromises between feature similarity and distortion, i.e, it minimizes the cost function (4) for the model M in the face database with respectto the correspondence points {xf}. (i, j) in Eq. (4) denotes that index j runs over the neighborhood of vertex i and index i runs in the new image over all vertices. By minimizing Eq. (4) the algorithm assigns pixel I to vertex i in the prototype graph M. Numerous classification experiments revealed that a steepest descent algorithm is sufficient to minimize cost function (4) although it is nonconvex and local minima may cause non-optimal correspondences with reduced recognition rates. x; During a recognition experiment all prototype graphs in the database are matched to the new image. A new face is classified as prototype A if H A is minimal and if the significance criterion (5) is fulfilled. The average costs (Ji) and their standard deviation LH are calculated excluding match A. This heuristic is based on the assumption that a new face image strongly Illumination-Invariant Face Recognition with a Contrast Sensitive Silicon Retina l> gr.tl rr.m.1 l>gr.rr~ l>~"~ l>~.tr. .llo'" ~1Ib.ka-:2to,.. > Workstation Datacube ~~ . .. Figure I: Laboratory setup of the face recognition experiments. correlates with the correct prototype but the matching costs to all the other prototype faces is approximately Gaussian distributed with mean (1l) and standard deviation I.H. The threshold parameter 0 is used to limit the rate of false positive matches, i.e., to exclude significant matches to wrong prototypes. 4 Face Recognition Results To measure the recognition rate of the face recognition system using a silicon retina or a CCD camera as input devices, pictures of 109 different persons are taken under 3 different lighting conditions and 2 different viewing directions. This setup allows us to quantify the influence of changes in lighting conditions on the recognition performance separate from the influence of perspective distortions. Figure 2 shows face images of one person taken under two different lighting setups. The images in Figs. 2a,c with both lights on are used as the prototype images for the respective input devices. To test the influence of changing lighting conditions the left light is switched off. The faces are now strongly illuminated from the right side. The CCD camera images (Figs. 2a,b) document the drastic changes of the light settings. The corresponding responses of the silicon retina shown in Figs. 2c,d clearly demonstrate that the local adaptivity of the silicon retina enables the recognition system to extract object structure from the bright and the dark side of the face. For control purposes all recognition experiments have been repeated with filtered CCD camera images. The filter was adjusted such that the power spectra of the retina chip images and the filtered CCD images are identical. The images (e,f) are filtered versions of the images (a,b). It is evident that information in the dark part of image (b) has been erased due to saturation effects of the CCD camera and cannot be recovered by any local filtering procedure. We first measure the performance of the silicon retina under uniform lighting conditions, n3 b ~ .... .. ~ ~ It ,. ...'. - '\. .. ... C .... .- ?... ~ ?. . ~ .~ ..... ? 1? Figure 2: (a) Conventional CCD camera images (a,b) and silicon retina image (c,d) under different lighting conditions. The images (e,O are filtered CCD camera images with a power spectrum adjusted to the images in (c,d). The images (a,c) are used to generate the Illumination-Invariant Face Recognition with a Contrast Sensitive Silicon Retina Table 1: (a) Face recognition results in a well illuminated environment and (b) in an environment with drastic changes in lighting conditions. a b f. p. rate 100% 100/0 50/0 10/0 1000/0 10% 50/0 10/0 silicon retina 83.5 81.7 76.2 71.6 96.3 96.3 96.3 93.6 cony. CCD 86.2 83.5 82.6 79.8 80.7 76.2 72.5 64.2 filt. CCD 85.3 84.4 80.7 75.2 78.0 75.2 72.5 62.4 i.e., both lamps are on and the person looks 20-30 degrees to the right. The recognition system has to deal with perspective distortions only. A gallery of 109 faces is matched to a face database of the same 109 persons. Table la shows that the recognition rate reaches values between 80 and 90 percent if we accept the best match without checking its significance. Such a decision criterion is unpractical for many applications since it corresponds to a false positive rate (f. p. rate) of 100 percent. If we increase the threshold E> to limit false positive matches to less than 1 percent the face recognizer is able to identify three out of four unknown faces. Filtering the CCD imagery does not hurt the recognition performance as the third column in Table 1a demonstrates. All necessary information for recognition is preserved in the filtered CCD images. The situation changes dramatically when we switch off the lamp on the left side of the test person. We compare a test gallery of persons looking straight ahead, but illuminated only from the right side, to our model gallery. Table 1b summarizes the recognition results for different false positive rates. The advantage of using a silicon retina are 20 to 45 percent higher recognition rates than for a system with a CCD camera. For a false positive rate below one percent a silicon retina based recognition system identifies two third more persons than a conventional system. Filtering does not improve the recognition rate of a system that uses a CCD camera as can be seen in the third column. Our comparative face recognition experiment clearly demonstrates that a face recognizer with a retina chip is performing substantially better than conventional CCD camera based systems in environments with uncontrolled, substantially changing lighting conditions. Retina-like preprocessing yields increased recognition rates and increased significance levels. We expect even larger discrepancies in recognition rates if object without a bilateral symmetry have to be classified. In this sense the face recognition task does not optimally explore the potential of adaptive preprocessing by a silicon retina. Imagine an object recognition task where the most significant features for discrimination are hardly visible or highly ambiguous due to poor illumination. High error rates and very low significance levels are an inevitable consequence of such lighting conditions. The limited resolution and poor signal-to-noise ratio of silicon retina chips are expected to be improved by a new generation of chips fabricated in 0.7 /lm CMOS technology with a 775 776 Buhmann, Lades, and Eeckman potential resolution of256 x 256 pixels. Lighting conditions as simulated in ourrecognition experiment are ubiquitous in natural environments. Autonomous robots and vehicles or surveillance systems are expected to benefit from the silicon retina technology by gaining robustness and reliability. Silicon retinas and more elaborate analog VLSI chips for low level vision are expected to be an important component of an Adaptive Vision System. Acknowledgement: It is a pleasure to thank K. A. Boahen for providing us with the retina chips. We acknowledge stimulating discussions with C. von der Malsburg and C. Mead. This work was supported by the German Ministry of Science and Technology (lTR-8800-H 1) and by the Lawrence Livermore National Laboratory (W-7405-Eng-48). References Boahen, K., Andreou, A. 1992. A Contrast Sensitive Silicon Retina with Reciprocal Synapses. Pages 764-772 of: NIPS91 Proceedings. IEEE. Brunelli, R., Poggio, T. (1993). Face Recognition: Features versus Templates. IEEE Trans. on Pattern Analysis Machine Intelligence, 15, 1042-1052. Buhmann, J., Lange, J., von der Malsburg, C. 1989. Distortion Invariant Object Recognition by Matching Hierarchically Labeled Graphs. Pages I 155-159 of' Proc. llCNN, Washington. IEEE. Buhmann, J., Lades, M., von der Malsburg, C. 1990. Size and Distortion Invariant Object Recognition by Hierarchical Graph Matching. Pages II 411-416 of' Proc. llCNN, SanDiego. IEEE. Burr, D. J. (1981). Elastic Matching of Line Drawings. IEEE Trans. on Pat. An. Mach. Intel., 3, 708-713. Lades, M., Vorbriiggen, J.C., Buhmann, J., Lange, J., von der Malsburg, C., Wurtz, R.P., Konen, W. (1993). Distortion Invariant Object Recognition in the Dynamic Link Architecture. IEEE Transactions on Computers, 42, 300-311. Mahowald, M., Mead, C. (1991). The Silicon Retina. Scientific American, 264(5), 76. Mead, C. (1989). Analog VLSI and Neural Systems. New York: Addison Wesley. Turk, M., Pentland, A. (1991). Eigenfaces for Recognition. J. Cog. Sci., 3, 71-86. von der Malsburg, Christoph. 1981. The Correlation Theory of Brain Function. Internal Report. Max-Planck-Institut, Biophys. Chern., Gottingen, Germany. Yuille, A. (1991). Deformable Templates for Face Recognition. J. Cog. Sci., 3, 60-70.
876 |@word neurophysiology:1 deformed:1 version:3 compression:1 seems:1 cco:2 ruhr:1 llo:1 eng:1 thereby:2 tr:1 document:1 ka:1 recovered:1 visible:1 shape:1 enables:1 remove:1 designed:4 discrimination:1 pursued:1 selected:1 device:3 intelligence:1 lamp:2 reciprocal:2 steepest:1 filtered:5 five:1 burr:2 expected:6 behavior:1 multi:1 brain:1 axij:1 equipped:1 jm:1 vertebrate:1 provided:2 matched:2 interpreted:1 minimizes:1 substantially:2 developed:1 transformation:1 fabricated:2 gottingen:1 expands:1 wrong:1 demonstrates:2 control:2 planck:1 positive:5 local:11 limit:2 consequence:1 receptor:7 elastically:1 mach:1 mead:6 approximately:1 might:1 emphasis:1 challenging:1 christoph:1 co:1 limited:1 range:3 faithful:1 camera:15 procedure:1 deforms:1 oflocal:1 matching:18 refers:1 bochum:2 cannot:1 influence:5 seminal:1 ilf:1 conventional:6 center:4 blueprint:1 maximizing:1 primitive:1 rectangular:1 resolution:5 assigns:1 his:1 autonomous:2 variation:3 antagonistic:1 hurt:1 imagine:1 wetware:1 us:2 recognition:53 particularly:2 database:6 labeled:1 electrical:2 digitally:1 boahen:3 environment:4 leak:1 dynamic:4 ultimately:1 weakly:1 compromise:1 serve:1 yuille:2 respectto:1 chip:11 represented:2 separated:1 fast:1 artificial:2 neighborhood:1 rebuilding:1 encoded:1 heuristic:1 plausible:1 larger:1 distortion:16 drawing:1 interconnection:1 furthennore:1 ability:1 statistic:1 advantage:1 rr:2 interaction:2 causing:1 achieve:1 deformable:1 description:2 moved:1 produce:2 comparative:2 cmos:1 object:14 coupling:2 nearest:1 eq:2 strong:1 quantify:1 direction:2 correct:2 filter:3 centered:2 viewing:2 collaborator:1 biological:3 konen:1 adjusted:2 exp:3 great:1 lawrence:2 lm:1 purpose:2 recognizer:3 proc:2 label:2 sensitive:6 organ:1 establishes:1 weighted:1 concurrently:1 sensor:2 gaussian:2 clearly:2 rather:1 surveillance:1 voltage:1 encode:1 joachim:1 rheinische:1 fur:1 contrast:11 sense:1 integrated:1 accept:1 initially:1 vlsi:4 germany:3 pixel:10 classification:8 orientation:2 field:2 construct:1 llcnn:2 washington:1 identical:1 represents:1 look:1 inevitable:1 mimic:1 discrepancy:1 report:2 intelligent:1 inherent:1 employ:1 retina:36 simultaneously:2 national:2 conductance:1 organization:2 highly:1 adjust:6 extreme:1 light:4 edge:7 necessary:1 poggio:2 lh:1 facial:2 institut:3 respective:1 carver:1 penalizes:1 deformation:1 minimal:1 increased:3 classify:1 column:2 mahowald:2 cost:9 deviation:3 vertex:9 uniform:1 recognizing:1 gr:2 optimally:1 reported:1 stored:1 combined:1 person:8 explores:1 sensitivity:4 unpractical:1 off:2 enhance:1 connectivity:1 von:6 imagery:1 choose:1 american:1 account:2 potential:3 exclude:1 retinal:2 coefficient:1 blurred:1 explicitly:2 caused:1 depends:1 bilateral:1 vehicle:1 wave:1 capability:1 ass:1 minimize:1 bright:1 wilhelms:1 yield:2 identify:2 ofthe:1 saliency:1 weak:1 handwritten:1 critically:1 informatik:1 lighting:15 published:1 straight:1 classified:2 detector:1 synapsis:2 reach:1 turk:2 naturally:1 associated:1 attributed:1 workstation:1 gain:3 knowledge:1 ubiquitous:2 wesley:1 originally:1 higher:1 vorbriiggen:1 planar:1 response:2 improved:1 box:1 strongly:4 stage:1 correlation:1 horizontal:9 multiscale:1 scientific:1 effect:4 nips91:1 lades:9 vicinity:1 laboratory:3 brunelli:2 deal:1 attractive:1 adjacent:2 sin:1 during:2 self:1 ambiguous:1 cosine:1 iris:1 criterion:3 evident:1 tt:2 demonstrate:2 percent:6 image:40 rotation:1 ji:3 attached:2 cerebral:1 jl:1 analog:6 interpretation:1 silicon:28 significant:2 surround:3 eeckman:4 ai:2 automatic:1 grid:2 reliability:2 robot:2 cortex:2 similarity:5 inhibition:1 perspective:5 nonconvex:1 der:6 seen:1 minimum:1 ministry:1 employed:1 neuroanatomy:1 paradigm:1 maximize:1 signal:2 ii:4 reduces:2 xf:1 match:5 shunting:1 controlled:1 vision:6 enhancing:1 metric:2 wurtz:1 cochlea:1 kernel:2 achieved:1 cell:10 chern:1 preserved:1 background:2 fine:1 spacing:1 plexiform:1 facilitates:1 revealed:1 iii:1 ikx:1 affect:1 xj:1 switch:1 architecture:2 lange:2 prototype:14 expression:2 filt:1 york:1 cause:1 hardly:1 dramatically:1 useful:1 detailed:1 tune:1 amount:1 dark:2 locally:1 cony:1 hardware:1 reduced:1 generate:2 datacube:1 inhibitory:2 fulfilled:1 promise:1 four:1 threshold:2 changing:4 preprocessed:1 registration:1 luminance:1 mosis:1 graph:10 cone:2 run:2 unclassified:1 parameterized:1 angle:1 place:1 smoothes:2 decision:2 summarizes:1 illuminated:3 layer:3 fl:1 uncontrolled:1 correspondence:3 ahead:1 n3:1 bonn:1 generates:1 innovative:1 performing:1 martin:1 format:1 tv:1 according:1 poor:2 electrically:1 membrane:2 character:1 invariant:8 taken:4 german:1 addison:1 drastic:2 eight:1 hierarchical:1 appropriate:1 generic:2 robustness:1 denotes:1 ccd:16 malsburg:6 strategy:3 receptive:2 enhances:2 separate:2 link:2 lateral:1 simulated:1 pleasure:1 outer:1 thank:1 sci:2 gallery:5 length:1 besides:1 index:2 ratio:2 minimizing:1 providing:1 setup:3 frank:1 design:1 reliably:1 unknown:5 perform:1 adjustable:1 convolution:1 acknowledge:1 descent:1 pentland:2 pat:1 defining:1 situation:1 communication:1 excluding:1 looking:1 dc:2 interacting:1 intensity:1 neuroinformatik:1 mechanical:1 livermore:3 friedrich:1 connection:1 andreou:2 registered:1 flm:1 trans:2 able:1 suggested:1 below:1 pattern:5 perception:1 program:1 saturation:1 including:1 gaining:1 max:1 power:3 natural:1 buhmann:10 improve:2 altered:1 technology:4 numerous:1 picture:1 identifies:1 coupled:3 extract:2 prior:1 acknowledgement:1 removal:1 checking:1 expect:1 adaptivity:2 generation:1 filtering:3 versus:1 switched:1 degree:1 ltr:1 sufficient:1 translation:1 excitatory:2 supported:1 free:1 drastically:1 side:4 allow:1 neighbor:1 template:4 face:43 taking:1 eigenfaces:1 distributed:1 benefit:1 feedback:1 calculated:1 sensory:1 qualitatively:1 adaptive:3 preprocessing:8 simplified:1 correlate:1 transaction:1 fiir:1 global:1 xi:8 spectrum:3 search:1 table:4 promising:1 itt:2 robust:1 ca:1 elastic:10 symmetry:1 significance:4 hierarchically:1 whole:1 noise:2 repeated:1 enriched:1 fig:4 intel:1 tl:1 elaborate:1 fashion:3 diffusivity:4 bandpass:1 ib:1 third:3 wavelet:2 cog:2 offset:1 false:5 magnitude:1 illumination:5 biophys:1 rejection:1 suited:1 explore:1 corresponds:1 stimulating:1 viewed:1 erased:1 universitiit:1 change:12 typical:1 determined:1 invariance:1 la:1 deforming:1 select:1 internal:1 romerstrabe:1 rotate:1
7,109
877
The "Softmax" Nonlinearity: Derivation Using Statistical Mechanics and Useful Properties as a Multiterminal Analog Circuit Element I. M. Elfadel Research Laboratory of Electronics Massachusetts Institute of Technology Cambridge, MA 02139 J. L. Wyatt, Jr. Research Laboratory of Electronics Massachusetts Institute of Technology Cambridge, MA 02139 Abstract We use mean-field theory methods from Statistical Mechanics to derive the "softmax" nonlinearity from the discontinuous winnertake-all (WTA) mapping. We give two simple ways of implementing "soft max" as a multiterminal network element. One of these has a number of important network-theoretic properties. It is a reciprocal, passive, incrementally passive, nonlinear, resistive multiterminal element with a content function having the form of informationtheoretic entropy. These properties should enable one to use this element in nonlinear RC networks with such other reciprocal elements as resistive fuses and constraint boxes to implement very high speed analog optimization algorithms using a minimum of hardware. 1 Introduction In order to efficiently implement nonlinear optimization algorithms in analog VLSI hardware, maximum use should be made of the natural properties of the silicon medium. Reciprocal circuit elements facilitate such an implementation since they 882 The "Softmax" Nonlinearity can be combined with other reciprocal elements to form an analog network having Lyapunov-like functions: the network content or co-content. In this paper, we show a reciprocal implementation of the "softmax" nonlinearity that is usually used to enforce local competition between neurons [Peterson, 1989]. We show that the circuit is passive and incrementally passive, and we explicitly compute its content and co-content functions. This circuit adds a new element to the library of the analog circuit designer that can be combined with reciprocal constraint boxes [Harris, 1988] and nonlinear resistive fuses [Harris, 1989] to form fast, analog VLSI optimization networks. 2 Derivation of the Softmax Nonlinearity To a vector y E ~n of distinct real numbers, the discrete winner-take-all (WTA) mapping W assigns a vector of binary numbers by giving the value 1 to the component of y corresponding to maxl<i<n Yi and the value 0 to the remaining components. Formally, W is defined as - - W(y) = (Wl(y), ... Wn(y?T I where for every 1 ~ j ~ n, 1 if YJ' > Yi, V 1 ~ i ~ n Wj (y) = { 0 otherwise (1) Following [Geiger, 1991], we assign to the vector y the "energy" function n Ey(z) = - L ZkYk = _zT y, z E Vnl (2) k=l = where Vn is the set of vertices of the unit simplex Sn {x E ~n, Zi > 0, 1 < i < n and E~=l Zk = 1}. Every vertex in the simplex encodes one possible winner. It is then easy to show that W(y) is the solution to the linear programming problem n max LZkYk. ZE'V .. k=l Moreover, we can assign to the energy Ey(z) the Gibbs distribution e-Ey(Z)/T Py(z) = Py(Zl' ... , zn) = ZT where T is the temperature of the heat bath and ZT is a normalizing constant. Then one can show that the mean of Zj considered as a random variable is given by [Geiger, 1991] A _ Fj(y/T) ey;/T ey;/T = = -z = Ei=l n/T? T eY ' Zj The mapping F : ~n -+ ~n whose components are the Fj's, 1 ~ j < n, is the generalized sigmoid mapping [Peterson, 1989] or "soft max" . It plays, in WTA networks, a role similar to that of the sigmoidal function in Hopfield and backpropagation 883 884 Elfadel and Wyatt v Figure 1: A circuit implementation of softmax with 5 inputs and 5 outputs. This circuit is operated in subthreshold mode, takes the gates voltages as inputs and gives the drain currents as outputs. This circuit is not a reciprocal multiterminal element. networks [Hopfield, 1984, Rumelhart, 1986] and is usually used for enforcing competitive behavior among the neurons of a single cluster in networks of interacting clusters [Peterson, 1989, Waugh, 1993]. A For y E ~n, we denote by FT(y) = F(y IT). The softmax mapping satisfies the following properties: 1. The mapping FT converges pointwise to W over ~n as T -+ 0 and to the center of mass of Sn, *e = *(1,1, . . . , 1)T E ~n, as T -+ +00. 2. The Jacobian DF of the softmax mapping is a symmetric n x n matrix that satisfies DF(y) = diag (F,,(y? - F(y)F(y)T. (3) It is always singular with the vector e being the only eigenvector corresponding to the zero eigenvalue. Moreover, all its eigenvalues are upper-bounded by maxlS"Sn F,,(y) < 1. 3. The soft max mapping is a gradient map, i.e, there exists a "potential" function 'P : ~n -+ ~ such that F = V'P. Moreover 'P is convex. The symbol 'P was chosen to indicate that it is a potential function. It should be noted that if F is the gradient map of 'P then FT is the gradient map of T'PT where 'PT(Y) 'P(yIT). In a related paper [Elfadel, 1993], we have found that the convexity of'P is essential in the study of the global dynamics of analog WTA networks. Another instance where the convexity of 'P was found important is the one reported in [Kosowsky, 1991] where a mean-field algorithm was proposed to solve the linear assignment problem. = The "Softmax" Nonlinearity v, Vz v" Va Ta Figure 2: Modified circuit implementation of softmax. In this circuit all the transistors are diode-connected, and all the drain currents are well in saturation region. Note that for every transistor, both the voltage input and the current output are on the same wire - the drain. This circuit is a reciprocal multiterminal element. 3 Circuit Implementations and Properties Now we propose two simple CMOS circuit implementations of the generalized sigmoid mapping. See Figures 1 and 2. When the transistors are operated in the subthreshold region the drain currents i l , .. . ,in are the outputs of a softmax mapping whose inputs are the gate voltages Vl, ?.. , V n . The explicit v - i characteristics are given by (4) where K, is a process-dependent parameter and Vo is the thermal voltage ([Mead, 1989],p. 36). These circuits have the interesting properties of being unclocked and parallel. Moreover, the competition constraint is imposed naturally through the KCL equation and the control current source. From a complexity point of view, this circuit is most striking since it computes n exponentials, n ratios, and n - 1 sums in one time constant! A derivation similar to the above was independently carried out in [Waugh, 1993] for the circuit of Figure 1. Although the first circuit implements softmax, it has two shortcomings. The first is practical: the separation between inputs and outputs implies additional wiring. The second is theoretical: this circuit is not a reciprocal multiterminal element, and therefore it can't be combined with other reciprocal elements like resistive fuses or constraint boxes to design analog, reciprocal optimization networks. Therefore, we only consider the circuit of Figure 2 and let v and i be the ndimensional vectors representing the input voltages and the output currents, respectively. 1 The softmax mapping i = F(v) represents a voltage-controlled, nonlinear, lCompare with Lazarro et. al.'s WTA circuit [Lazzaro, 1989] whose inputs are currents and outputs are voltages. 885 886 Elfadel and Wyatt resistive multiterminal element. The main result of our paper is the following: 2 Theorem 1 The softmax multiterminal element F is reciprocal, passive, locally passive and has a co-content function given by 1 n ~(v) K, Ie Vo In exp(K,vm/Vo) L = (5) m=l and a content function given by ..w..*(O) _ IeVo ~ im 1nim I - - - L.J -. 'If K, m=l Ie (6) Ie Thus, with this reciprocal, locally passive implementation of the softmax mapping, we have added a new circuit element to the library of the circuit designer. Note that this circuit element implements in an analog way the constraint L:~=1 y" = 1 defining the unit simplex Sn. Therefore, it can be considered a nonlinear constraint box [Harris, 1988] that can be used in reciprocal networks to implement analog optimization algorithms. The expression of ~* is a strong reminder of the information-theoretic definition of entropy. We suggest the name "entropic resistor" for the circuit of Figure 2. 4 Conclusions In this paper, we have discussed another instance of convergence between the statistical physics paradigm of Gibbs distributions and analog circuit implementation in the context of the winner-take-all function. The problem of using the simple, reciprocal circuit implementation of softmax to design analog networks for finding near optimal solutions of the linear assignment problem [Kosowsky, 1991] or the quadratic assignment problem [Simic, 1991] is still open and should prove a challenging task for analog circuit designers. Acknowledgements I. M. Elfadel would like to thank Alan Yuille for many helpful discussions and Fred Waugh for helpful discussions and for communicating the preprint of [Waugh, 1993]. This work was supported by the National Science Foundation under Grant No. MIP-91-17724. References [Peterson, 1989] C. Peterson and B. Soderberg. A new method for mapping optimization problems onto neural networks. International Journal of Neural Systems, 1(1):3 - 22, 1989. 2The concepts of reciprocity, passivity, content, and co-content are fundamental to nonlinear circuit theory. They are carefully developed in [Wyatt, 1992]. The "Softmax" Nonlinearity J. G. Harris. Solving early vision problems with VLSI constraint networks. In Neural Architectures for Computer Vision Workshop, AAAI-88, Minneapolis, MN, 1988. [Harris,1989] J. G. Harris, C. Koch, J. Luo, and J. Wyatt. Resistive fuses: Analog hardware for detecting discontinuities in early vision. In C. Mead and M. Ismail, editors, Analog VLSIImplemenation of Neural Systems. Kluwer Academic Publishers, 1989. [Geiger, 1991] D. Geiger and A. Yuille. A common framework for image segmentation. Int. J. Computer Vision, 6:227 - 253, 1991. [Hopfield, 1984] J. J. Hopfield. Neurons with graded responses have collective computational properties like those of two-state neurons. Proc. Nat'l Acad. Sci., USA, 81:3088-3092, 1984. [Rumelhart, 1986] D. E. Rumelhart et. al. Parallel Distributed Processing, volume 1. MIT Press, 1986. [Waugh, 1993] F. R. Waugh and R. M. Westervelt. Analog neural networks with local competition. I. dynamics and stability. Physical Review E, 1993. in press. [Elfadel, 1993] I. M. Elfadel. Global dynamics of winner-take-all networks. In [Harris, 1988] SPIE Proceedings, Stochastic and Neural Methods in Image Processing, volume 2032, pages 127 - 137, San Diego, CA, 1993. [Kosowsky, 1991] J. J. Kosowsky and A. L. Yuille. The invisible hand algorithm: Solving the assignment problem with statistical physics. TR # 91-1, Harvard Robotics Laboratory, 1991. [Mead, 1989] Carver Mead. Analog VLSI and Neural Systems. Addison-Wesley, 1989. [Lazzaro,1989] J. Lazarro, S. Ryckebush, M. Mahowald, and C. Mead. Winnertake-all circuits of O(n) complexity. In D. S. Touretsky, editor, Advances in Neural Information Processing Systems I, pages 703 - 711. Morgan Kaufman, 1989. [Wyatt, 1992] J. L. Wyatt. Lectures on Nonlinear Circuit Theory. MIT VLSI memo # 92-685,1992. [Simic, 1991] P. D. Simic. Constrained nets for graph matching and other quadratic assignment problems. Neural Computation, 3:169 - 281, 1991. 887
877 |@word graded:1 diode:1 concept:1 indicate:1 implies:1 lyapunov:1 added:1 open:1 discontinuous:1 symmetric:1 laboratory:3 stochastic:1 wiring:1 enable:1 gradient:3 implementing:1 tr:1 noted:1 simic:3 thank:1 sci:1 assign:2 electronics:2 generalized:2 theoretic:2 vo:3 im:1 invisible:1 enforcing:1 temperature:1 passive:7 current:7 fj:2 koch:1 luo:1 considered:2 image:2 exp:1 touretsky:1 reciprocity:1 mapping:13 sigmoid:2 common:1 ratio:1 physical:1 entropic:1 early:2 winner:4 memo:1 volume:2 analog:17 proc:1 discussed:1 kluwer:1 implementation:9 elfadel:7 zt:3 silicon:1 design:2 collective:1 cambridge:2 gibbs:2 vz:1 reciprocal:15 wl:1 upper:1 neuron:4 wire:1 thermal:1 defining:1 detecting:1 mit:2 nonlinearity:7 always:1 winnertake:2 interacting:1 sigmoidal:1 modified:1 rc:1 voltage:7 add:1 passivity:1 prove:1 resistive:6 kosowsky:4 binary:1 behavior:1 helpful:2 mechanic:2 waugh:6 dependent:1 yi:2 morgan:1 minimum:1 vl:1 additional:1 discontinuity:1 usually:2 ey:6 vlsi:5 paradigm:1 saturation:1 max:4 moreover:4 bounded:1 circuit:30 medium:1 mass:1 among:1 facilitate:1 alan:1 kaufman:1 natural:1 academic:1 eigenvector:1 constrained:1 developed:1 softmax:17 ndimensional:1 field:2 finding:1 vnl:1 having:2 mn:1 representing:1 technology:2 va:1 represents:1 every:3 controlled:1 library:2 carried:1 vision:4 df:2 simplex:3 sn:4 zl:1 unit:2 control:1 grant:1 robotics:1 review:1 acknowledgement:1 drain:4 national:1 local:2 lecture:1 singular:1 source:1 acad:1 publisher:1 interesting:1 mead:5 foundation:1 soderberg:1 name:1 usa:1 challenging:1 co:4 editor:2 near:1 operated:2 pointwise:1 minneapolis:1 easy:1 wn:1 practical:1 supported:1 zi:1 yj:1 architecture:1 implement:5 backpropagation:1 institute:2 peterson:5 carver:1 expression:1 mip:1 distributed:1 theoretical:1 matching:1 maxl:1 fred:1 instance:2 computes:1 soft:3 suggest:1 wyatt:7 made:1 onto:1 san:1 zn:1 assignment:5 lazzaro:2 context:1 mahowald:1 useful:1 vertex:2 py:2 informationtheoretic:1 map:3 imposed:1 center:1 global:2 locally:2 independently:1 convex:1 reported:1 hardware:3 assigns:1 zj:2 communicating:1 combined:3 international:1 fundamental:1 designer:3 ie:3 stability:1 vm:1 physic:2 discrete:1 zk:1 ca:1 pt:2 play:1 diego:1 multiterminal:8 aaai:1 programming:1 yit:1 diag:1 harvard:1 element:16 rumelhart:3 ze:1 main:1 fuse:4 graph:1 sum:1 potential:2 role:1 ft:3 preprint:1 striking:1 int:1 wj:1 region:2 connected:1 explicitly:1 vn:1 separation:1 geiger:4 view:1 explicit:1 resistor:1 exponential:1 convexity:2 complexity:2 competitive:1 jacobian:1 parallel:2 nim:1 dynamic:3 quadratic:2 theorem:1 solving:2 constraint:7 yuille:3 characteristic:1 efficiently:1 westervelt:1 subthreshold:2 encodes:1 symbol:1 normalizing:1 exists:1 essential:1 hopfield:4 speed:1 workshop:1 derivation:3 nat:1 distinct:1 fast:1 kcl:1 shortcoming:1 heat:1 entropy:2 jr:1 whose:3 definition:1 solve:1 energy:2 wta:5 otherwise:1 naturally:1 spie:1 satisfies:2 massachusetts:2 harris:7 equation:1 eigenvalue:2 transistor:3 net:1 reminder:1 ma:2 propose:1 segmentation:1 addison:1 carefully:1 wesley:1 content:9 ta:1 bath:1 response:1 ismail:1 box:4 enforce:1 competition:3 convergence:1 cluster:2 hand:1 gate:2 formally:1 ei:1 cmos:1 converges:1 nonlinear:8 incrementally:2 remaining:1 derive:1 mode:1 giving:1 strong:1
7,110
878
Learning Stochastic Perceptrons Under k-Blocking Distributions Mario Marchand Ottawa-Carleton Institute for Physics University of Ottawa Ottawa, Ont., Canada KIN 6N5 mario@physics.uottawa.ca Saeed Hadjifaradji Ottawa-Carleton Institute for Physics University of Ottawa Ottawa, Ont., Canada KIN 6N5 saeed@physics.uottawa.ca Abstract We present a statistical method that PAC learns the class of stochastic perceptrons with arbitrary monotonic activation function and weights Wi E {-I, 0, +I} when the probability distribution that generates the input examples is member of a family that we call k-blocking distributions. Such distributions represent an important step beyond the case where each input variable is statistically independent since the 2k-blocking family contains all the Markov distributions of order k. By stochastic percept ron we mean a perceptron which, upon presentation of input vector x, outputs 1 with probability fCLJi WiXi - B). Because the same algorithm works for any monotonic (nondecreasing or nonincreasing) activation function f on Boolean domain, it handles the well studied cases of sigmolds and the "usual" radial basis functions. 1 INTRODUCTION Within recent years, the field of computational learning theory has emerged to provide a rigorous framework for the design and analysis of learning algorithms. A central notion in this framework, known as the "Probably Approximatively Correct" (PAC) learning criterion (Valiant, 1984), has recently been extended (Hassler, 1992) to analyze the learn ability of probabilistic concepts (Kearns and Schapire, 1994; Schapire, 1992). Such concepts, which are stochastic rules that give the probability that input example x is classified as being positive, are natural probabilistic 280 Mario Marchand, Saeed Hadjifaradji extensions of the deterministic concepts originally studied by Valiant (1984). Motivated by the stochastic nature of many "real-world" learning problems and by the indisputable fact that biological neurons are probabilistic devices, some preliminary studies about the PAC learnability of simple probabilistic neural concepts have been reported recently (Golea and Marchand, 1993; Golea and Marchand, 1994). However, the probabilistic behaviors considered in these studies are quite specific and clearly need to be extended. Indeed, only classification noise superimposed on a deterministic signum function was considered in Golea and Marchand (1993). The probabilistic network, analyzed in Golea and Marchand (1994), consists of a linear superposition of signum functions and is thus solvable as a (simple) case of linear regression. What is clearly needed is the extension to the non-linear cases of sigmolds and radial basis functions. Another criticism about Golea and Marchand (1993 , 1994) is the fact that their learn ability results was established only for distributions where each input variable is statistically independent from all the others (sometimes called product distributions). In fact, very few positive learning results for non-trivial p-concepts classes are known to hold for larger classes of distributions. Therefore, in an effort to find algorithms that will work in practice, we introduce in this paper a new family of distributions that we call k-blocking. As we will argue, this family has the dual advantage of avoiding malicious and unnatural distributions that are prone to render simple concept classes unlearnable (Lin and Vitter, 1991) and of being likely to contain several distributions found in practice. Our main contribution is to present a simple statistical method that PAC learns (in polynomial time) the class of stochastic perceptrons with monotonic (but otherwise arbitrary) activation functions and weights Wi E { -1,0, +1} when the input examples are generated according to any distribution member of the k-blocking family. Due to space constraints, only a sketch of the proofs is presented here. 2 DEFINITIONS The instance (input) space, In, is the Boolean domain {-1, +1}n. The set of all input variables is denoted by X. Each input example x is generated according to some unknown distribution D on In. We will often use PD(X), or simply p(x), to denote the probability of observing the vector value x under distribution D. If U and V are two disjoint subsets of X, Xu and Xv will denote the restriction (or projection) of x over the variables of U and V respectively and p D (xu Ixv) will denote the probability, under distribution D, of observing the vector value Xu (for the variables in U) given that the variables in V are set to the vector value xv. Following Kearns and Schapire (1994), a probabilistic concept (p-concept) is a map c : In ~ [0, 1] for which c(x) represents the probability that example X is classified as positive. More precisely, upon presentation of input x, an output of a = 1 is generated (by an unknown target p-concept) with probability c(x) and an output of a = 0 is generated with probability 1 - c(x). A stochastic perceptron is a p-concept parameterized by a vector of n weights Wi and a activation function fO such that, the probability that input example X is Learning Stochastic Perceptrons under k-Blocking Distributions 281 classified as positive is given by (1) We consider the case of a non-linear function fO since the linear case can be solved by a standard least square approximation like the one performed by Kearns in Schapire (1994) for linear sums of basis functions. We restrict ourselves to the case where fO is monotonic i.e. either nondecreasing or nonincreasing. But since any nonincreasing f (.) combined with a weight vector w can always be represented by a nondecreasing f(?) combined with a weight vector -w, we can assume without loss of generality that the target stochastic perceptron has a nondecreasing f (. ). Hence, we allow any sigmoid-type of activation function (with arbitrary threshold). Also, since our instance space zn is on an-sphere, eq. 1 also include any nonincreasing radial basis function of the type ?(z2) where z = Ix - wi and w is interpreted as the "center" of ?. The only significant restriction is on the weights where we allow only for Wi E {-I, 0, +1}. As usual, the goal of the learner is to return an hypothesis h which is a good approximation of the target p-concept c. But, in contrast with decision rule learning which attempts to "filter out" the noisy behavior by returning a deterministic hypothesis, the learner will attempt the harder (and more useful) task of modeling the target p-concept by returning a p-concept hypothesis. As a measure of error between the target and the hypothesis p-concepts we adopt the variation distance dv (?,?) defined as: err(h,c) = dv(h,c) ~f LPD(X) Ih(x) - c(x)1 x (2) Where the summation is over all the 2n possible values of x. Hence, the same D is used for both training and testing. The following formulation of the PAC criterion (Valiant, 1984; Hassler, 1992) will be sufficient for our purpose. Definition 1 Algorithm A is said to PAC learn the class C of p-concepts by using the hypothesis class H (of p-concepts) under a family V of distributions on instance space In, iff for any c E C, any D E V, any 0 < t,8 < 1, algorithm A returns in a time polynomial in (l/t, 1/8, n), an hypothesis h E H such that with probability at least 1 - 8, err(h, c) < t. 3 K-BLOCKING DISTRIBUTIONS To learn the class of stochastic perceptrons, the algorithm will try to discover each weight Wi that connects to input variable Xi by estimating how the probability of observing a positive output (0" = 1) is affected by "hard-wiring" variable Xi to some fixed value. This should clearly give some information about Wi when Xi is statistically independent from all the other variables as was the case for Golea and Marchand (1993) and Schapire (1992). However, if the input variables are correlated, then the process of fixing variable Xi will carry over neighboring variables which in turn will affect other variables until all the variables are perturbed (even in the simplest case of a first order Markov chain). The information about Wi will 282 Mario Marchand, Saeed HadjiJaradji then be smeared by all the other weights. Therefore, to obtain information only on Wi, we need to break this "chain reaction" by fixing some other variables. The notion of blocking sets serves this purpose. Loosely speaking, a set of variables is said to be a blocking set 1 for variable Xi if the distribution on all the remaining variables is unaffected by the setting of Xi whenever all the variables of the blocking set are set to a fixed value. More precisely, we have: Definition 2 Let B be a subset of X and let U = X - (B U {Xi}). Let XB and Xu be the restriction of X on Band U respectively and let b be an assignment for XB. Then B is said to be a blocking set for variable Xi (with respect to D), iff: PD(xulxB = b,Xi = +1) = PD(xulxB = b,Xi = -1) for all b and Xu In addition, if B is not anymore a blocking set when we remove anyone of its variables, we then say that B is a minimal blocking set for variable Xi. We thus adopt the following definition for the k-blocking family. Definition 3 Distribution D on rn is said to be k-blocking iff IBil < k 1,2?? . n when each Bi is a minimal blocking set for variable Xi. for i = The k-blocking family is quite a large class of distributions. In fact we have the following property: Property 1 All Markov distributions of kth order are members of the 2k-blocking family. Proof: By kth order Markov distributions, we mean distributions which can be exactly written as a Chow(k) expansion (see Hoeffgen, 1993) for some permutation of the variables. We prove it here (by using standard techniques such as in Abend et. al, 1965) for first order Markov distributions, the generalization for k > 1 is straightforward. Recall that for Markov chain distributions we have: p(XjIXj-b??? xI) = p(XjIXj_l) for 1 < j ~ n. Hence: P(XI ... Xj-2, Xj+2? .. XnlXj-b Xj, Xj+!) = p(Xl)p(X2Ixl)??? p(Xj IXj-l)p(Xj+llxj)??? P(XnIXn-l)!p(Xj-b Xj, Xj+!) = p(xI)p(x2Ixd??? p(xj-llxj-2)P(Xj+2Ixj+l)? .. P(XnIXn-l)!p(Xj-I) = P(Xl?? ?Xj-2,Xj+2? ??XnIXj-bXj,Xj+!) where Xj denotes the negation of Xj. Thus, we see that Markov chain distributions are a special case of 2-blocking distributions: the blocking set of each variable consisting only of the two first-neighbor variables. D. The proposed algorithm for learning stochastic perceptrons needs to be provided with a blocking set (of at most k variables) for each input variable. Hoeffgen (1993) has recently proven that Chow(l) and Chow(k > 1) expansions are efficiently learnable; the latter under some restricted conditions. We can thus use these algorithms IThe wording "blocking set" was also used by Hancock & Mansour (Proc. of COLT'91 , 179-183, Morgan Kaufmann Publ.) to denote a property of the target concept. In contrast, our definition of blocking set denotes a property of the input distribution only. Learning Stochastic Perceptrons under k-Blocking Distributions 283 to discover the blocking sets for such distributions. However , the efficient learnability of unrestricted Chow(k > 1) expansions and larger classes of distributions, such as the k-blocking family, is still unknown. In fact, from the hardness results of Hoeffgen (1993), we can see that it is definitely very hard (perhaps NP-complete) to find the blocking sets if the learner has no information available other than the fact that the distribution is k-blocking. On the other hand, we can argue that the "natural" ordering of the variables present in many "real-world" situations is such that the blocking set of any given variable is among the neighboring variables. In vision for example, we expect that the setting of a pixel will directly affect only those located in it's neighborhood; the other pixels being affected only through this neighborhood. In such cases, the neighborhood of a variable "naturally" provides its blocking set. 4 LEARNING STOCHASTIC PERCEPTRONS We first establish (the intuitive fact) that, without making much error, we can always consider that the target p-concept is defined only over the variables which are not almost always set to the same value. Lemma 1 Let V be a set of v variables Xi for which Pr(xi = ai) > 1 - a. Let c be a p-concept and let c' be the same p-concept as c except that the reading of each variable Xi E V is replaced by the reading of the constant value ai. Then err( c' , c) < v . a. Proof: Let a be the vector obtained from the concatenation of all ais and let Xv be the vector obtained from X by keeping only the components Xi which are in V. Then err(c', c) ~ Pr(xv =I- a) ~ L:iEVPr(Xi =I- ai). D. For a given set of blocking sets {Bi }f=l ' the algorithm will try to discover each weight Wi by estimating the blocked influence of Xi defined as: Binf(xilhi) ~f Pr(O' = 11xBi = hi , Xi = +1) - Pr(O' = 11xBi = hi, Xi = -1) where XB i denotes the restriction of x on the blocking set Bi for variable Xi and hi is an assignment for XB i ? The following lemma ensures the learner that Binf(xilhi) contains enough information about Wi. Lemma 2 Let the taryet p-concept be a stochastic perceptron on In having a nondecreasing activation function and weights taken from {-1, 0, +1}. Then, for any assignment hi for the variables in the blocking set Bi of variable Xi, we have: Binf(xilhi) { ~ 0 if Wi = +1 if Wi = 0 0 if Wi = -1 = 0 ~ (3) Proof sketch: Let U = X - (Bi U {Xi}), s = L:jEUWjXj and ( = L:kEBi wkbk? Let pes) denote the probability of observing s (under D). Then Binf(xilhi) = L:sp(s) [f(s + (+ Wi) - f(s + (- Wi)]; from which we find the desired result for a nondecreasing f(?) . D. 284 Mario Marchand, Saeed Hadjifaradji In principle, lemma 2 enables the learner to discover Wi from Binf(xilbi). The learner, however, has only access to its empirical estimate Bfnf(xilbi) from a finite sample. Hence, we will use Hoeffding's inequality (Hoeffding, 1963) to find the number of examples needed for a probability p to be close to its empirical estimate ft with high probability. Lemma 3 (Hoeffding, 1963) Let YI. ... , Y m be a sequence of m independent Bernoulli trials, each succeeding with probability p. Let ft = L:l ~/m. Then: Pr (1ft - pi > E) ~ 2 exp (-2mE2) Hence, by writing Binf(xilbi) in terms of (unconditional) probabilities that can be estimated from all the training examples, we find from lemma 3 that the number mO(E,8,n) of examples needed to have IBfnf(xilbi) - Binf(xilbi)I < E with probability at least 1 - 8 is given by: mO(E,8,n) ~ ~ (:E) 2 ln (~) where /'i, = a k +1 is the lowest permissible value for PD(b i , Xi) (see lemma 1). So, if the minimal nonzero value for IBinf(xilbi)1 is (3, then the number of examples needed to find, with confidence at least 1 - 8, the exact value for Wi among { -1,0, + I} is such that we need to have: Pr(IBfnf(xilbi) - Binf(xilbi)1 < (3/2) > 1 - 8. Thus, whenever (3 is oH2(e- n ), we will need of O(e2n ) examples to find (with prob > 1-8) the value for Wi. So, in order to be able to PAC learn from a polynomial sample, we must arrange ourselves so that we do not need to worry about such low values for IBinf(xilbi)l. We therefore consider the maximum blocked influence defined as: Binf(xi) ~f Binf(xilbn where b; is the vector value for which IBinf(xilbi)1 is the largest. We now show that the learner can ignore all variables Xi for which IBinf(xi)1 is too small (without making much error). Lemma 4 Let c be a stochastic perceptron with nondecreasing activation function f (.) and weights taken from { -1, 0, + 1}. Let V c X and let cv be the same stochastic perceptron as c except that Wi = 0 for all Xi E V and its activation function is changed to f ( . + e). Then, there always exists a value for e such that: err(cv, c) :::; 2: IBinf(xi)I iEV Proof sketch: By induction on IVI. To first verify the lemma for V = {Xl}, let b be a vector of values for the setting of all Xi E Bl and let Xu be a vector of values for the setting of all Xj E U = X - (Bl U {Xl}). Let s = LjEU WjXj and ( = LjEB l WjXj, then for e = WI, we have: err(cv, c) = 2: 2:PD(xul b )PD(blxl = -l)PD(XI = -1) Xu b x If(s + ( + WI) - f(s + ( - wl)1 ~ IBinf(xdl Learning Stochastic Perceptrons under k-Blocking Distributions 285 We now assume that the lemma holds for V = {Xl. X2?? . Xk} and prove it for W = V U {Xk+1}. Let S = {Xk+1} and let f(? + Ow), f(? + Ov) and f(? + Os) denote respectively the activation function for cw, Cv and cs. By inspecting the expressions for err(cv, c) and err(cw, cs), we can see that there always exist a value for Ow E {Ov + Wk+1,OV - Wk+l} and Os E {Wk+l. -wk+d such that err( cw, cs) ::; err( cv, c). And since dv (-, .) satisfies the triangle inequality, err(cw,c)::; err(cv,c) + IBinf(xk+1)I. D. After discovering the weights, the hypothesis p-concept h returned by the learner will simply be the table look-up of the estimated probabilities of observing a positive classification given that ~~= 1 Wi Xi = s for all s values that are observed with sufficient probability (the hypothesis can output any value for the values of s that are observed very rarely). We thus have the following learning algorithm for stochastic perceptrons. Algorithm LearnSP(n, ?, 6, {Bi}i=l) 1. Call m = 128 e: ) 2kHIn e~n) training examples (where k = maXi IBi I). = +1) for each variable Xi. Neglect Xi whenever we have Pr(xi = +1) < ?/(4n) or Pr(xi = +1) > 1 - ?/(4n). 2. Compute Pr(xi 3. For each variable Xi and for each of its blocking vector value hi, compute Bfnf(xilhi). Let h; be the value of hi for which IBfnf(xilhi)1 is the largest. Let Bfnf(xi) = Bfnf(xilh;). 4. For each variable Xi: (a) Let Wi = +1 whenever Bfnf(xi) > ?/(4n). (b) Let Wi = -1 whenever Bfnf(xi) < ?/(4n). (c) Otherwise let Wi = 0 5. Compute Pr(~~=l Wi Xi = s) for s = -n, . .. + n. 6. Return the hypothesis p-concept h formed by the table look-up: h(x) = h'(s) = Pr (0' = 1 t WiXi = s) ~=l for all s for which Pr(~~=l WiXi = s) > ?/(8n let h'(s) = 0 (or any other value). + 8). For the other s values, Theorem 1 Algorithm LearnSP PAC learns the class of stochastic perceptrons on In with monotonic activation functions and weights Wi E {-1, 0, +1} under any k-blocking distribution (when a blocking set for each variable is known). The number of examples required is m = 128 (2:) 2kHIn (l~n) (and the time needed is O(n x m)) for the returned hypothesis to make error at most ? with confidence at least 1 - 6. Proof sketch: From Hoeffding's inequality (lemma 3) we can show that this sample size is sufficient to ensure that: 286 Mario Marchand, Saeed Hadjifaradji ? IPr(Xi = +1) - Pr(xi = +1)1 < ? IBfnf(xi) - Binf(xi)1 < ? IPr(I:~=l WiXi = s) - ~/(4n) with confidence at least 1 - ~/(4n) with confidence at least 1 - Pr(I:~=l WiXi = s)1 < 6/(4n) 6/(4n) ~2/[64(n + 1)] with confidence at least 1- 6/(4n + 4) ? IPr(O" = llI:~=l WiXi = s) - Pr(O" = dence at least 1 - 6/4 llI:~=l WiXi = s)1 < ~/4 with confi- From this and from lemma 1,2 and 4, it follows that returned hypothesis will make error at most ~ with confidence at least 1 - 6. D. Acknowledgments We thank Mostefa Golea, Klaus-U. Hoeffgen and Stefan Poelt for useful comments and discussions about technical points. M. Marchand is supported by NSERC grant OGP0122405. Saeed Hadjifaradji is supported by the MCHE of Iran. References Abend K., Hartley T.J. & Kanal L.N. (1965) "Classification of Binary Random Patterns", IEEE Trans. Inform. Theory vol. IT-II, 538-544. Golea, M. & Marchand M. (1993) "On Learning Perceptrons with Binary Weights", Neural Computation vol. 5, 765-782. Golea, M. & Marchand M. (1994) "On Learning Simple Deterministic and Probabilistic Neural Concepts", in Shawe-Talor J. , Anthony M. (eds.), Computational Learning Theory: EuroCOLT'93, Oxford University Press, pp. 47-60. Haussler D. (1992) "Decision Theoritic Generalizations of the PAC Model for Neural Net and Other Learning Applications", Information and Computation vol. 100,78150. Hoeffgen K.U. (1993) "On Learning and Robust Learning of Product Distributions", Proceedings of the 6th ACM Conference on Computational Learning Theory, ACM Press, 77-83. Hoeffding W. (1963) "Probability inequalities for sums of bounded random variabIes", Journal of the American Statistical Association, vol. 58(301), 13-30. Kearns M.J. and Schapire R.E. (1994) "Efficient Distribution-free Learning ofProbabilistic Concepts", Journal of Computer and System Sciences, Vol. 48, pp. 464-497. Lin J.H. & Vitter J.S. (1991) "Complexity Results on Learning by Neural Nets", Machine Learning, Vol. 6, 211-230. Schapire R.E. (1992) The Design and Analysis of Efficient Learning Algorithms, Cambridge MA: MIT Press. Valiant L.G. (1984) "A Theory of the Learnable", Comm. ACM, Vol. 27, 11341142.
878 |@word trial:1 polynomial:3 harder:1 carry:1 contains:2 reaction:1 err:12 z2:1 ixj:2 activation:10 written:1 must:1 enables:1 remove:1 succeeding:1 discovering:1 device:1 xk:4 provides:1 ron:1 consists:1 prove:2 vitter:2 introduce:1 hardness:1 indeed:1 behavior:2 eurocolt:1 ont:2 provided:1 discover:4 estimating:2 bounded:1 lowest:1 what:1 interpreted:1 mostefa:1 exactly:1 returning:2 unlearnable:1 grant:1 positive:6 xv:4 oxford:1 ipr:3 ibi:1 studied:2 xbi:2 bi:6 statistically:3 acknowledgment:1 testing:1 practice:2 empirical:2 projection:1 confidence:6 radial:3 close:1 influence:2 writing:1 restriction:4 deterministic:4 map:1 center:1 straightforward:1 rule:2 haussler:1 handle:1 notion:2 variation:1 target:7 exact:1 hypothesis:11 located:1 blocking:37 observed:2 ft:3 solved:1 ensures:1 ordering:1 abend:2 pd:7 comm:1 complexity:1 wording:1 ov:3 ithe:1 upon:2 xul:1 learner:8 basis:4 triangle:1 represented:1 uottawa:2 hancock:1 klaus:1 neighborhood:3 quite:2 emerged:1 larger:2 say:1 otherwise:2 ability:2 nondecreasing:7 noisy:1 advantage:1 sequence:1 net:2 product:2 neighboring:2 iff:3 intuitive:1 fixing:2 eq:1 c:3 hartley:1 correct:1 filter:1 stochastic:19 golea:9 generalization:2 preliminary:1 biological:1 summation:1 inspecting:1 extension:2 hold:2 considered:2 exp:1 mo:2 arrange:1 adopt:2 purpose:2 proc:1 e2n:1 superposition:1 largest:2 wl:1 smeared:1 stefan:1 mit:1 clearly:3 always:5 signum:2 bernoulli:1 superimposed:1 contrast:2 rigorous:1 criticism:1 chow:4 pixel:2 classification:3 dual:1 colt:1 denoted:1 among:2 special:1 field:1 having:1 represents:1 look:2 others:1 hassler:2 np:1 few:1 replaced:1 ourselves:2 connects:1 consisting:1 saeed:7 negation:1 attempt:2 analyzed:1 unconditional:1 nonincreasing:4 chain:4 xb:4 iev:1 loosely:1 desired:1 minimal:3 instance:3 modeling:1 boolean:2 zn:1 assignment:3 ottawa:6 subset:2 too:1 learnability:2 reported:1 perturbed:1 combined:2 definitely:1 probabilistic:8 physic:4 central:1 hoeffding:5 american:1 return:3 wk:4 performed:1 try:2 break:1 mario:6 analyze:1 observing:5 contribution:1 square:1 formed:1 kaufmann:1 percept:1 efficiently:1 oh2:1 lli:2 unaffected:1 classified:3 fo:3 inform:1 whenever:5 ed:1 definition:6 pp:2 naturally:1 proof:6 wixi:7 recall:1 worry:1 originally:1 formulation:1 generality:1 until:1 sketch:4 hand:1 o:2 perhaps:1 concept:25 contain:1 hoeffgen:5 verify:1 hence:5 nonzero:1 bxj:1 wiring:1 criterion:2 complete:1 wjxj:2 binf:11 recently:3 sigmoid:1 association:1 significant:1 blocked:2 cambridge:1 ai:4 cv:7 shawe:1 access:1 recent:1 inequality:4 binary:2 yi:1 morgan:1 unrestricted:1 ii:1 technical:1 sphere:1 lin:2 regression:1 n5:2 vision:1 represent:1 sometimes:1 addition:1 xdl:1 malicious:1 carleton:2 permissible:1 ivi:1 probably:1 comment:1 member:3 call:3 enough:1 affect:2 xj:18 restrict:1 motivated:1 expression:1 unnatural:1 effort:1 lpd:1 render:1 returned:3 speaking:1 useful:2 iran:1 band:1 simplest:1 schapire:7 exist:1 estimated:2 disjoint:1 vol:7 affected:2 threshold:1 year:1 sum:2 prob:1 parameterized:1 family:10 almost:1 decision:2 hi:6 marchand:14 precisely:2 constraint:1 x2:1 dence:1 generates:1 anyone:1 according:2 wi:28 making:2 dv:3 restricted:1 pr:15 taken:2 ln:1 turn:1 needed:5 serf:1 available:1 anymore:1 denotes:3 remaining:1 include:1 ensure:1 neglect:1 establish:1 bl:2 usual:2 said:4 ow:2 kth:2 distance:1 cw:4 thank:1 concatenation:1 argue:2 trivial:1 induction:1 design:2 publ:1 unknown:3 neuron:1 theoritic:1 markov:7 finite:1 situation:1 extended:2 rn:1 mansour:1 arbitrary:3 canada:2 required:1 established:1 trans:1 beyond:1 able:1 pattern:1 reading:2 natural:2 solvable:1 loss:1 expect:1 permutation:1 proven:1 sufficient:3 principle:1 pi:1 prone:1 changed:1 supported:2 keeping:1 free:1 allow:2 perceptron:6 institute:2 neighbor:1 world:2 ignore:1 indisputable:1 xi:50 table:2 learn:5 nature:1 robust:1 ca:2 kanal:1 expansion:3 anthony:1 domain:2 sp:1 main:1 noise:1 xu:7 xl:5 pe:1 learns:3 ix:1 kin:2 theorem:1 specific:1 pac:9 learnable:2 maxi:1 exists:1 ih:1 valiant:4 simply:2 likely:1 nserc:1 approximatively:1 monotonic:5 satisfies:1 acm:3 ma:1 goal:1 presentation:2 hard:2 except:2 kearns:4 lemma:12 called:1 perceptrons:12 rarely:1 latter:1 avoiding:1 correlated:1
7,111
879
Nonlinear Image Interpolation using Manifold Learning Christoph Bregler Computer Science Division University of California Berkeley, CA 94720 bregler@cs.berkeley.edu Stephen M. Omohundro'" Int . Computer Science Institute 1947 Center Street Suite 600 Berkeley, CA 94704 om@research.nj .nec.com Abstract The problem of interpolating between specified images in an image sequence is a simple, but important task in model-based vision . We describe an approach based on the abstract task of "manifold learning" and present results on both synthetic and real image sequences. This problem arose in the development of a combined lip-reading and speech recognition system. 1 Introduction Perception may be viewed as the task of combining impoverished sensory input with stored world knowledge to predict aspects of the state of the world which are not directly sensed. In this paper we consider the task of image interpolation by which we mean hypothesizing the structure of images which occurred between given images in a temporal sequence. This task arose during the development of a combined lipreading and speech recognition system [3], because the time windows for auditory and visual information are different (30 frames per second for the camera vs. 100 feature vectors per second for the acoustic information). It is an excellent visual test domain in general, however, because it is easy to generate large amounts of test and training data and the performance measure is largely "theory independent" . The test consists of simply presenting two frames from a movie and comparing the "'New address: NEe Research Institute, Inc., 4 Independence Way, Princeton, NJ 08540 974 Christoph Bregler. Stephen M. Omohundro Figure 1: Linear interpolated lips. Figure 2: Desired interpolation. hypothesized intermediate frames to the actual ones. It is easy to use footage of a particular visual domain as training data in the same way. Most current approaches to model-based vision require hand-constructed CADlike models. We are developing an alternative approach in which the vision system builds up visual models automatically by learning from examples. One of the central components of this kind of learning is the abstract problem of inducing a smooth nonlinear constraint manifold from a set of examples from the manifold. We call this "manifold learning" and have developed several approaches closely related to neural networks for doing it [2]. In this paper we apply manifold learning to the image interpolation problem and numerically compare the results of this "nonlinear" process with simple linear interpolation. We find that the approach works well when the underlying model space is low-dimensional. In more complex examples, manifold learning cannot be directly applied to images but still is a central component in a more complex system (not discussed here). We present several approaches to using manifold learning for this task. We compare the performance of these approaches to that of simple linear interpolation. Figure 1 shows the results of linear interpolation of lip images from the lip-reading system. Even in the short period of 33 milliseconds linear interpolation can produce an unnatural lip image. The problem is that linear interpolation of two images just averages the two pictures. The interpolated image in Fig. 1 has two lower lip parts instead of just one. The desired interpolated image is shown in Fig. 2, and consists of a single lower lip positioned at a location between the lower lip positions in the two input pictures. Our interpolation technique is nonlinear, and is constrained to produce only images from an abstract manifold in "lip space" induced by learning. Section 2 describes the procedure, Section 4 introduces the interpolation technique based on the induced manifold, and Sections 5 and 6 describe our experiments on artificial and natural images. 2 Manifold Learning Each n * m gray level image may be thought of as a point in an n * m-dimensional space. A sequence of lip-images produced by a speaker uttering a sentence lie on a Nonlinear Image Interpolation Using Manifold Learning - 975 \ Graylevel Dimensions ------ (l6x16 pixel =256 dim. space) Figure 3: Linear vs nonlinear interpolation. 1-dimensional trajectory in this space (figure 3). If the speaker were to move her lips in all possible ways, the images would define a low-dimensional submanifold (or nonlinear surface) embedded in the high-dimensional space of all possible graylevel images. If we could compute this nonlinear manifold, we could limit any interpolation algorithm to generate only images contained in it. Images not on the manifold cannot be generated by the speaker under normal circumstances. Figure 3 compares a curve of interpolated images lying on this manifold to straight line interpolation which generally leaves the manifold and enters the domain of images which violate the integrity of the model. To represent this kind of nonlinear manifold embedded in a high-dimensional feature space, we use a mixture model of local linear patches. Any smooth nonlinear manifold can be approximated arbitrarily well in each local neighborhood by a linear "patch" . In our representation, local linear patches are "glued" together with smooth "gating" functions to form a globally defined nonlinear manifold [2]. We use the "nearest-point-query" to define the manifold. Given an arbitrary point near the manifold, this returns the closest point on the manifold. We answer such queries with a weighted sum of the linear projections of the point to each local patch. The weights are defined by an "influence function" associated with each linear patch which we usually define by a Gaussian kernel. The weight for each patch is the value of its influence function at the point divided by the sum of all influence functions ("partition of unity"). Figure 4 illustrates the nearest-point-query. Because Gaussian kernels die off quickly, the effect of distant patches may be ignored, improving computational performance. The linear projections themselves consist of a dot product and so are computationally inexpensive. For learning, we must fit such a mixture of local patches to the training data. An initial estimate of the patch centers is obtained from k-means clustering. We fit a patch to each local cluster using a local principal components analysis. Fine tuning 976 Christoph Bregler. Stephen M. Omohundro Influence Function PI P2 \ Linear Patch l:,Gi(X) ' Pi(x) P(x) = . .:. .'-=--l:,Gi(X) , Figure 4: Local linear patches glued together to a nonlinear manifold. of the model is done using the EM (expectation-maximization) procedure. This approach is related to the mixture of expert architecture [4], and to the manifold representation in [6]. Our EM implementation is related to [5] , which uses a hierarchical gating function and local experts that compute linear mappings from one space to another space. In contrast, our approach uses a "one-level" gating function and local patches that project a space into itself. 3 Linear Preprocessing Dealing with very high-dimensional domains (e.g. 256 * 256 gray level images) requires large memory and computational resources. Much of this computation is not relevant to the task, however. Even if the space of images is nonlinear, the nonlinearity does not necessarily appear in all of the dimensions. Earlier experiments in the lip domain [3] have shown that images projected onto a lO-dimensional linear subspace still accurately represents all possible lip configurations. We therefore first project the high-dimensional images into such a linear subspace and then induce the nonlinear manifold within this lower dimensional linear subspace. This preprocessing is similar to purely linear techniques [7, 10, 9] . 4 Constraint Interpolation Geometrically, linear interpolation between two points in n-space may be thought of as moving along the straight line joining the two points. In our non-linear approach to interpolation, the point moves along a curve joining the two points which lies in the manifold of legal images. We have studied several algorithms for estimating the shortest manifold trajectory connecting two given points. For the performance results, we studied the point which is halfway along the shortest trajectory. Nonlinear Image Interpolation Using Manifold Learning 4.1 977 "Free-Fall" The computationally simplest approach is to simply project the linearly interpolated point onto the nonlinear manifold. The projection is accurate when the point is close to the manifold. In cases where the linearly interpolated point is far away (i.e. no weight of the partition of unity dominates all the other weights) the closest-pointquery does not result in a good interpolant. For a worst case, consider a point in the middle of a circle or sphere. All local patches have same weight and the weighted sum of all projections is the center point itself, which is not a manifold point. Furthermore, near such "singular" points, the final result is sensitive to small perturbations in the initial position. 4.2 "Manifold-Walk" A better approach is to "walk" along the manifold itself rather than relying on the linear interpolant. Each step of the walk is linear and in the direction of the target point but the result is immediately projected onto the manifold. This new point is then moved toward the target point and projected onto the manifold, etc. When the target is finally reached, the arc length of the curve is approximated by the accumulated lengths of the individual steps. The point half way along the curve is chosen as the interpolant. This algorithm is far more robust than the first one, because it only uses local projections, even when the two input points are far from each other. Figure 5b illustrates this algorithm. 4.3 "Manifold-Snake" This approach combines aspects of the first two algorithms. It begins with the linearly interpolated points and iteratively moves the points toward the manifold. The Manifold-Snake is a sequence of n points preferentially distributed along a smooth curve with equal distances between them. An energy function is defined on such sequences of points so that the energy minimum tries to satisfy these constraints (smoothness, equidistance, and nearness to the manifold): (1) E has value 0 if all Vi are evenly distributed on a straight line and also lie on the manifold. In general E can never be 0 if the manifold is nonlinear, but a minimum for E represents an optimizing solution. We begin with a straight line between the two input points and perform gradient descent in E to find this optimizing solution. 5 Synthetic Examples To quantify the performance of these approaches to interpolation, we generated a database of 16 * 16 pixel images consisting of rotated bars. The bars were rotated for each image by a specific angle. The images lie on a one-dimensional nonlinear manifold embedded in a 256 dimensional image space. A nonlinear manifold represented by 16 local linear patches was induced from the 256 images. Figure 6a shows 978 Christoph Bregler, Stephen M. Omohundro a) "Free Fall" b) "Surface Walk" c) "Surface Snake" Figure 5: Proposed interpolation algorithms. -~/ Figure 6: a) Linear interpolation, b) nonlinear interpolation. two bars and their linear interpolation. Figure 6b shows the nonlinear interpolation using the Manifold- Walk algorithm. Figure 7 shows the average pixel mean squared error of linear and nonlinear interpolated bars. The x-axis represents the relative angle between the two input points. Figure 8 shows some iterations of a Manifold-Snake interpolating 7 points along a 1 dimensional manifold embedded in a 2 dimensional space . .... pi --- ......... /' ./ ~/ .t. ,__ / VV --- --- -- --- --- - Figure 7: Average pixel mean squared error of linear and nonlinear interpolated bars. Nonlinear Image Interpolation Using Manifold Learning 979 QQOOOO oIter. I Iter. 2 Iter. 5 Iter. to Iter. 30 Iter. Figure 8: Manifold-Snake iterations on an induced 1 dimensional manifold embedded in 2 dimensions. Figure 9: 16x16 images. Top row: linear interpolation. Bottom row : nonlinear "manifold-walk" interpolation. 6 Natural Lip Images We experimented with two databases of natural lip images taken from two different subjects. Figure 9 shows a case of linear interpolated and nonlinear interpolated 16 * 16 pixel lip images using the Manifold- Walk algorithm . The manifold consists of 16 4-dimensional local linear patches. It was induced from a training set of 1931 lip images recorded with a 30 frames per second camera from a subject uttering various sentences. The nonlinear interpolated image is much closer to a realistic lip configuration than the linear interpolated image. Figure 10 shows a case of linear interpolated and nonlinear interpolated 45 * 72 pixel lip images using the Manifold-Snake algorithm. The images were recorded with a high-speed 100 frames per second camera l . Because of the much higher dimensionality of the images, we projected the images into a 16 dimensional linear subspace. Embedded in this subspace we induced a nonlinear manifold consisting of 16 4-dimensionallocallinear patches, using a training set of 2560 images. The linearly interpolated lip image shows upper and lower teeth, but with smaller contrast , because it is the average image of the open mouth and closed mouth. The nonlinearly interpolated lip images show only the upper teeth and the lips half way closed, which is closer to the real lip configuration. 7 Discussion We have shown how induced nonlinear manifolds can be used to constrain the interpolation of gray level images. Several interpolation algorithms were proposed IThe images were recorded in the UCSD Perceptual Science Lab by Michael Cohen Christoph Bregler. Stephen M. Omohundro 980 Figure 10: 45x72 images projected into a 16 dimensional subspace. Top row: linear interpolation. Bottom row: nonlinear "manifold-snake" interpolation. and experimental studies have shown that constrained nonlinear interpolation works well both in artificial domains and natural lip images. Among various other nonlinear image interpolation techniques, the work of [1] using a Gaussian Radial Basis Function network is most closely related to our approach. Their approach is based on feature locations found by pixelwise correspondence, where our approach directly interpolates graylevel images. Another related approach is presented in [8]. Their images are also first projected into a linear subspace and then modelled by a nonlinear surface but they require their training examples to lie on a grid in parameter space so that they can use spline methods. References [1] D. Beymer, A. Shahsua, and T. Poggio Example Based Image Analysis and Synthesis M.I.T. A.1. Memo No. 1431, Nov. 1993. [2] C. Bregler and S. Omohundro, Surface Learning with Applications to Lip-Reading, in Advances in Neural Information Processing Systems 6, Morgan Kaufmann, 1994. [3] C. Bregler and Y. Konig, "Eigenlips" for Robust Speech Recognition in Proc. ofIEEE Int. Conf. on Acoustics, Speech, and Signal Processing, Adelaide, Australia, 1994. [4] R.A. Jacobs, M.1. Jordan, S.J. Nowlan, and G.E. Hinton, Adaptive mixtures of local experts in Neural Compuation, 3, 79-87. [5] M.1. Jordan and R. A. Jacobs, Hierarchical Mixtures of Experts and the EM Algorithm Neural Computation, Vol. 6, Issue 2, March 1994. [6] N. Kambhatla and T.K. Leen, Fast Non-Linear Dimension Reduction in Advances in Neural Information Processing Systems 6, Morgan Kaufmann, 1994. [7] M. Kirby, F. Weisser, and G. DangeImayr, A Model Problem in Represetation of Digital Image Sequences, in Pattern Recgonition, Vol 26, No.1, 1993. [8] H. Murase, and S. K. Nayar Learning and Recognition of 3-D Objects from Brightness Images Proc. AAAI, Washington D.C., 1993. [9] P. Simard, Y. Le Cun, J. Denker Efficient Pattern Recognition Using a New Transformation Distance Advances in Neural Information Processing Systems 5, Morgan Kaufman, 1993. [10] M. Turk and A. Pentland, Eigenfaces for Recognition Journal of Cognitive Neuroscience, Volume 3, Number 1, MIT 1991.
879 |@word middle:1 open:1 sensed:1 jacob:2 brightness:1 reduction:1 initial:2 configuration:3 current:1 com:1 comparing:1 nowlan:1 must:1 realistic:1 distant:1 partition:2 v:2 half:2 leaf:1 short:1 nearness:1 location:2 along:7 constructed:1 consists:3 combine:1 themselves:1 footage:1 globally:1 relying:1 automatically:1 actual:1 window:1 project:3 estimating:1 underlying:1 begin:2 kind:2 kaufman:1 developed:1 transformation:1 nj:2 suite:1 temporal:1 berkeley:3 appear:1 local:15 limit:1 joining:2 interpolation:34 glued:2 studied:2 christoph:5 camera:3 procedure:2 thought:2 projection:5 induce:1 radial:1 cannot:2 onto:4 close:1 influence:4 center:3 immediately:1 graylevel:3 target:3 us:3 recognition:6 approximated:2 database:2 bottom:2 enters:1 worst:1 interpolant:3 ithe:1 purely:1 division:1 basis:1 represented:1 various:2 fast:1 describe:2 artificial:2 query:3 neighborhood:1 gi:2 itself:3 final:1 sequence:7 product:1 relevant:1 combining:1 inducing:1 moved:1 konig:1 cluster:1 produce:2 rotated:2 object:1 nearest:2 p2:1 c:1 murase:1 quantify:1 direction:1 closely:2 australia:1 require:2 bregler:8 lying:1 normal:1 mapping:1 predict:1 kambhatla:1 proc:2 sensitive:1 eigenlips:1 weighted:2 mit:1 gaussian:3 rather:1 arose:2 contrast:2 dim:1 accumulated:1 snake:7 her:1 pixel:6 issue:1 among:1 development:2 constrained:2 equal:1 never:1 washington:1 represents:3 hypothesizing:1 spline:1 individual:1 consisting:2 introduces:1 mixture:5 accurate:1 closer:2 poggio:1 walk:7 desired:2 circle:1 earlier:1 maximization:1 submanifold:1 pixelwise:1 stored:1 answer:1 synthetic:2 combined:2 off:1 michael:1 together:2 quickly:1 connecting:1 synthesis:1 squared:2 central:2 recorded:3 aaai:1 conf:1 cognitive:1 expert:4 simard:1 return:1 int:2 inc:1 satisfy:1 vi:1 try:1 closed:2 lab:1 doing:1 reached:1 om:1 kaufmann:2 largely:1 modelled:1 accurately:1 produced:1 trajectory:3 straight:4 inexpensive:1 energy:2 turk:1 associated:1 auditory:1 knowledge:1 dimensionality:1 positioned:1 impoverished:1 higher:1 leen:1 done:1 furthermore:1 just:2 hand:1 nonlinear:34 gray:3 effect:1 hypothesized:1 iteratively:1 during:1 speaker:3 die:1 presenting:1 omohundro:6 image:62 cohen:1 volume:1 discussed:1 occurred:1 numerically:1 smoothness:1 tuning:1 grid:1 nonlinearity:1 dot:1 moving:1 surface:5 etc:1 integrity:1 closest:2 optimizing:2 arbitrarily:1 lipreading:1 morgan:3 minimum:2 shortest:2 period:1 signal:1 stephen:5 violate:1 smooth:4 sphere:1 divided:1 vision:3 circumstance:1 expectation:1 iteration:2 represent:1 kernel:2 fine:1 singular:1 induced:7 subject:2 jordan:2 call:1 near:2 intermediate:1 easy:2 independence:1 fit:2 architecture:1 unnatural:1 speech:4 interpolates:1 ignored:1 generally:1 amount:1 simplest:1 generate:2 millisecond:1 neuroscience:1 per:4 vol:2 iter:5 geometrically:1 halfway:1 sum:3 angle:2 patch:17 correspondence:1 constraint:3 constrain:1 interpolated:17 aspect:2 speed:1 developing:1 march:1 describes:1 smaller:1 em:3 unity:2 kirby:1 cun:1 taken:1 computationally:2 resource:1 legal:1 apply:1 denker:1 hierarchical:2 away:1 alternative:1 top:2 clustering:1 build:1 move:3 gradient:1 subspace:7 distance:2 street:1 evenly:1 manifold:56 toward:2 length:2 preferentially:1 memo:1 implementation:1 perform:1 upper:2 arc:1 descent:1 pentland:1 hinton:1 frame:5 perturbation:1 ucsd:1 arbitrary:1 nonlinearly:1 specified:1 sentence:2 california:1 acoustic:2 nee:1 address:1 bar:5 usually:1 perception:1 pattern:2 reading:3 memory:1 mouth:2 natural:4 movie:1 picture:2 axis:1 relative:1 embedded:6 digital:1 teeth:2 pi:3 lo:1 row:4 free:2 vv:1 institute:2 fall:2 eigenfaces:1 distributed:2 curve:5 dimension:4 world:2 sensory:1 uttering:2 adaptive:1 preprocessing:2 projected:6 far:3 nov:1 dealing:1 lip:25 robust:2 ca:2 improving:1 excellent:1 interpolating:2 complex:2 necessarily:1 domain:6 linearly:4 fig:2 x16:1 position:2 lie:5 perceptual:1 specific:1 gating:3 experimented:1 dominates:1 consist:1 nec:1 illustrates:2 simply:2 beymer:1 visual:4 contained:1 viewed:1 principal:1 experimental:1 adelaide:1 princeton:1 nayar:1
7,112
88
144 SPEECH RECOGNITION EXPERIMENTS WITH PERCEPTRONS D. J. Burr Bell Communications Research Morristown, NJ 07960 ABSTRACT Artificial neural networks (ANNs) are capable of accurate recognition of simple speech vocabularies such as isolated digits [1]. This paper looks at two more difficult vocabularies, the alphabetic E-set and a set of polysyllabic words. The E-set is difficult because it contains weak discriminants and polysyllables are difficult because of timing variation. Polysyllabic word recognition is aided by a time pre-alignment technique based on dynamic programming and E-set recognition is improved by focusing attention. Recognition accuracies are better than 98% for both vocabularies when implemented with a single layer perceptron. INTRODUCTION Artificial neural networks perform well on simple pattern recognition tasks. On speaker trained spoken digits a layered network performs as accurately as a conventional nearest neighbor classifier trained on the same tokens [1]. Spoken digits are easy to recognize since they are for the most part monosyllabic and are distinguished by strong vowels. It is reasonable to ask whether artificial neural networks can also solve more difficult speech recognition problems. Polysyllabic recognition is difficult because multi-syllable words exhibit large timing variation. Another difficult vocabulary, the alphabetic E-set, consists of the words B, C, D, E, G, P, T, V, and Z. This vocabulary is hard since the distinguishing sounds are short in duration and low in energy. We show that a simple one-layer perceptron [7] can solve both problems very well if a good input representation is used and sufficient examples are given. We examine two spectral representations - a smoothed FFT (fast Fourier transform) and an LPC (linear prediction coefficient) spectrum. A time stabilization technique is described which pre-aligns speech templates based on peaks in the energy contour. Finally, by focusing attention of the artificial neural network to the beginning of the word, recognition accuracy of the E-set can be consistently increased. A layered neural network, a relative of the earlier percept ron [7], can be trained by a simple gradient descent process [8]. Layered networks have been ? American Institute of Physics 1988 145 applied successflJ.lly to speech recognition [1], handwriting recognition [2], and to speech synthesis [11]. A variation of a layered network [3] uses feedback to model causal constraints, which can be useful in learning speech and language. Hidden neurons within a layered network are the building blocks that are used to form solutions to specific problems. The number of hidden units required is related to the problem [1,2]. Though a single hidden layer can form any mapping [12], no more than two layers are needed for disjunctive normal form [4]. The second layer may be useful in providing more stable learning and representation in the presence of noise. Though neural nets have been shown to perform as well as conventional techniques[I,5], neural nets may do better when classes have outliers [5]. PERCEPTRONS A simple perceptron contains one input layer and one output layer of neurons directly connected to each other (no hidden neurons). This is often called a one-layer system, referring to the single layer of weights connecting input to output. Figure 1. shows a one-layer perceptron configured to sense speech patterns on a two-dimensional grid. The input consists of a 64-point spectrum at each of twenty time slices. Each of the 1280 inputs is connected to each of the output neurons, though only a sampling of connections are shown. There is one output neuron corresponding to each pattern class. Neurons have standard linear-weighted inputs with logistic activation. C(1) C(2) C(N-1) C(N) FR:<lBC'V .... 64 units Figure 1. A single layer perceptron sensing a time-frequency array of sample data. Each output neuron CU) (1 <i<N) corresponds to a pattern class and is full connected to the input array (for clarity only a few connections are shown). An input word is fit to the grid region by applying an automatic endpoint detection algorithm. The algorithm is a variation of one proposed by Rabiner and Sambur [9] which employs a double threshold successive approximation 146 method. Endpoints are determined by first detecting threshold crossings of energy and then of zero crossing rate. In practice a level crossing other than zero is used to prevent endpoints from being triggered by background sounds. INPUT REPRESENTATIONS Two different input representations were used in this study. The first is a Fourier representation smoothed in both time and frequency. Speech is sampled at 10 KHz ap.d Hamming windowed at a number of sample points. A 128-point FFT spectrum is computed to produce a template of 64 spectral samples at each of twenty time frames. The template is smoothed twice with a time window of length three and a frequency window of length eight. For comparison purposes an LPC spectrum is computed using a tenth order model on 300-sample Hamming windows. Analysis is performed using the autocorrelation method with Durbin recursion [6]. The resulting spectrum is smoothed over three time frames. Sample spectra for the utterance "neural-nets" is shown in Figure 2. Notice the relative smoothness of the LPC spectrum which directly models spectral peaks. FFT LPC Figure 2. FFT and LPC time-frequency plots for the utterance "neural nets". Time is toward the left, and frequency, toward the right. DYNAMIC TIME ALIGMv1ENT Conventional speech recognition systems often employ a time normalization technique based on dynamic programming [10]. It is used to warp the time scales of two utterances to obtain optimal alignment between their spectral frames. We employ a variation of dynamic programming which aligns energy contours rather than spectra. A reference energy template is chosen for each pattern class, and incoming patterns are warped onto it. Figure 3 shows five utterances of "neural-nets" both before and after time alignment. Notice the improved alignment of energy peaks. 147 ? I ? ? ? I ~ II ~ ~ !I ! >- \b ~ III z W .. .. (a. ) 10 10 ... TIME (b) Figure 3. (a) Superimposed energy plots of five different utterances of "neural nets". (b). Same utterances after dynamic time alignment. POLYSYLLABLE RECOGNITION Twenty polysyllabic words containing three to five syllables were chosen, and five tokens of each were recorded by a single male speaker. A variable number of tokens were used to train a simple perceptron to study the effect of training set size on performance. Two performance measures were used: classification accuracy, and an RMS error measure. Training tokens were permuted to obtain additional experimental data points. Figure 4. Output responses of a perceptron trained with one token per class (left) and four tokens per class (right). 148 Figure 4 shows two representative perspective plots of the output of a perceptron trained on one and four tokens respectively per class. Plots show network response (z-coordinate) as a function of output node (left axis) and test word index (right axis). Note that more training tokens produce a more ideal map - a map should have ones along the diagonal and zeroes everywhere else. Table 1 shows the results of these experiments for three different representations: (1) FFT, (2) LPC and (3) time aligned LPC. This table lists classification accuracy as a function of number of training tokens and input representation. The perceptron learned to classify the unseen patterns perfectly for all cases except the FFT with a single training pattern. Table 1. Polysyllabic Number Training Tokens FFT LPC Time Aligned LPC Permuted Trials Word Recognition 2 1 98.7% 100% 100% 100% 100% 100% 400 300 AccuraclT 3 100% 100% 100% 200 4 100% 100% 100% 100 A different performance measure, the RMS error, evaluates the degree to which the trained network output responses Rjk approximate the ideal targets T jk ? The measure "is evaluated over the N non-trained tokens and M output nodes of the network. Tik equals 1 for J=k and 0 for J=I=k. Figure 5 shows plots of RMS error as a function of input representation and training patterns. Note that the FFT representation produced the highest error, LPC was about 40% less, and time-aligned LPC only marginally better than non-aligned LPC. In a situation where many choices must be made (i.e. vocabularies much larger than 20 words) LPC is the preferred choice, and time alignment could be useful to disambiguate similar words. Increased number of training tokens results in improved performance in all cases. 149 o ci ,-----------------------------~ '"0 .. FFT i "! l- I- ii 0 5 0 g W tJl ~ LPC a: '"0 0 o o TIme Aligned LPC ~--~----~--~----~__~____~ 1.0 2.0 3.0 4.0 Number Traln'ng Tokens Figure 5. RMS error versus number of training tokens for various input representations. E-SET VOCABULARY The E-Set vocabulary consists of the nine E-words of the English alphabet - B, C, D, E, G, P, T, V, Z. Twenty tokens of each of the nine classes were recorded by a single male speaker. To maximize the sizes of training and test sets, half were used for training and the other half for testing. Ten permutations produced a total of 900 separate recognition trials. Figure 6 shows typical LPC templates for the nine classes. Notice the double formant ridge due to the ''E'' sound, which is common to all tokens. Another characteristic feature is the FO ridge - the upward fold on the left of all tokens which characterizes voicing or pitched sound. 150 Figure 6. LP C time-frequency plots for representative tokens of the E-set words. Figure 7. Time-frequency plots of weight values connected to each output neuron ''E'' through "z" in a trained perceptron. 151 Figure 7 shows similar plots illustrating the weights learned by the network when trained on ten tokens of each class. These are plotted like spectra, since one weight is associated with each spectral sample. Note that the patterns have some formant structure. A recognition accuracy of 91.4% included perfect scores for classes C, E, and G. Notice that weights along the FO contour are mostly small and some are slightly negative. This is a response to the voiced ''E" sound common to all classes. The network has learned to discount "voicing" as a discriminator for this vocabulary. Notice also the strong "hilly" terrain near the beginning of most templates. This shows where the network has decided to focus much of its discriminating power. Note in particular the hill-valley pair at the beginning of ''p'' and "T". These are near to formants F2/F3 and could conceivably be formant onset detectors. Note the complicated detector pattern for the ''V'' sound. The classes that are easy to discriminate (C, E, G) produce relatively fiat and uninter~sting weight spaces. A highly convoluted weight space must therefore be correlated with difficulty in discrimination. It makes little sense however that the network should be working hard in the late time C'E" sound) portion of the utterance. Perhaps additional training might reduce this activity, since the network would eventually find little consistent difference there. A second experiment was conducted to help the network to focus attention. The first k frames of each input token were averaged to produce an average spectrum. These average spectra were then used in a simple nearest neighbor recognizer scheme. Recognition accuracy was measured as a function of k. The highest performance was for k=8, indicating that the first 40% of the word contained most of the "action". Figure 8. word. E C P T 0 0 0 0 0 0 100 0 0 0 0 0 0 0 D 0 0 08 0 0 2 0 0 0 E 0 0 0 100 0 0 0 0 0 c 0 0 0 0 100 0 0 0 0 p 0 0 3 0 0 03 4 0 0 T 0 0 0 0 0 0 100 0 0 V 2 0 0 0 0 2 0 0 Z 0 0 0 0 0 0 0 B C B 08 c D V Z 0 08 09 Confusion matrix of the E-set focused on the first 40% of each 152 All words were resampled to concentrate 20 time frames into the first 40% of the word. LPC spectra were recomputed using a 16th order model and the network was trained on the new templates. Performance increased from 91.4% to 98.2%. There were only 16 classification errors out of the 900 recognition tests. The confusion matrix is shown in Figure 8. Learning times for all experiments consisted of about ten passes through the training set. When weights were primed with average spectral values rather than random values, learning time decreased slightly. CONCLUSIONS Artificial neural networks are capable of high performance in pattern recognition applications, matching or exceeding that of conventional classifiers. We have shown that for difficult speech problems such as time alignment and weak discriminability, artificial neural networks perform at high accuracy exceeding 98%. One-layer perceptrons learn these difficult tasks almost effortlessly - not in spite of their simplicity, but because of it. REFERENCES 1. D. J. Burr, "A Neural Network Digit Recognizer", Proceedings of IEEE Conference on Systems, Man, and Cybernetics, Atlanta, GA, October, 1986, pp. 1621-1625. 2. D. J. Burr, "Experiments with a Connectionist Text Reader," IEEE International Conference on Neural Networks, San Diego, CA, June, 1987. 3. M. I. Jordan, "Serial Order: A Parallel Distributed Processing Approach," ICS Report 8604, UCSD Institute for Cognitive Science, La Jolla, CA, May 1986. 4. S. J. Hanson, and D. J. Burr, 'What Connectionist Models Learn: Toward a Theory of Representation in Multi-Layered Neural Networ.ks," submitted for pu blication. 5. W. Y. Huang and R. P. Lippmann, "Comparisons Between Neural Net and Conventional Classifiers," IEEE International Conference on Neural Networks, San Diego, CA, June 21-23, 1987. 6. J. D. Markel and A. H. Gray, Jr., Linear Prediction of Speech, SpringerVerlag, New York, 1976. 7. M. L. Minsky and S. Papert, Perceptrons, MIT Press, Cambridge, Mass., 1969. 153 8. D. E. Rumelhart, G. E. Hinton, and R. J. Williams, ''Learning Internal Representations by Error Propagation," in Parallel Distributed Processing, Vol. 1, D. E. Rumelhart and J. L. McClelland, eds., MIT Press, 1986, pp. 318362. 9. L. R. Rabiner and M. R. Sambur, "An Algorithm for Determining the Endpoints of Isolated Utterances," BSTJ, Vol. 54,297-315, Feb. 1975. 10. H. Sakoe and S. Chiba, "Dynamic Programming Optimization for Spoken Word Recognition," IEEE Trans. Acoust., Speech, Signal Processing, Vol. ASSP-26, No.1, 43-49, Feb. 1978. 11. T. J. Sejnowski and C. R. Rosenberg, "NETtalk: A Parallel Network that Learns to Read Aloud," Technical Report JHU/EECS-86/01, Johns Hopkins University Electrical Engineering and Computer Science, 1986. 12. A. Wieland and R. Leighton, "Geometric Analysis of Neural Network Capabilities," IEEE International Conference on Neural Networks, San Deigo, CA, June 21-24, 1987.
88 |@word trial:2 cu:1 illustrating:1 leighton:1 contains:2 score:1 activation:1 must:2 john:1 plot:8 discrimination:1 half:2 beginning:3 short:1 detecting:1 node:2 ron:1 successive:1 five:4 windowed:1 along:2 consists:3 burr:4 autocorrelation:1 sakoe:1 examine:1 multi:2 formants:1 little:2 window:3 sting:1 mass:1 what:1 spoken:3 acoust:1 nj:1 morristown:1 classifier:3 hilly:1 unit:2 before:1 engineering:1 timing:2 ap:1 might:1 twice:1 discriminability:1 k:1 monosyllabic:1 averaged:1 decided:1 testing:1 practice:1 block:1 digit:4 bell:1 jhu:1 matching:1 word:18 pre:2 spite:1 onto:1 ga:1 layered:6 valley:1 applying:1 conventional:5 map:2 williams:1 attention:3 duration:1 focused:1 simplicity:1 array:2 variation:5 coordinate:1 target:1 diego:2 programming:4 distinguishing:1 us:1 crossing:3 rumelhart:2 recognition:20 jk:1 disjunctive:1 electrical:1 region:1 connected:4 highest:2 dynamic:6 trained:10 networ:1 f2:1 various:1 alphabet:1 train:1 fast:1 sejnowski:1 artificial:6 aloud:1 larger:1 solve:2 formant:3 unseen:1 transform:1 triggered:1 net:7 fr:1 aligned:5 alphabetic:2 convoluted:1 double:2 produce:4 perfect:1 help:1 measured:1 nearest:2 strong:2 implemented:1 concentrate:1 stabilization:1 effortlessly:1 ic:1 normal:1 mapping:1 purpose:1 recognizer:2 tik:1 weighted:1 mit:2 rather:2 primed:1 rosenberg:1 focus:2 june:3 consistently:1 superimposed:1 sense:2 hidden:4 upward:1 classification:3 tjl:1 equal:1 polysyllabic:5 f3:1 ng:1 sampling:1 look:1 connectionist:2 report:2 few:1 employ:3 recognize:1 minsky:1 vowel:1 detection:1 atlanta:1 highly:1 alignment:7 male:2 accurate:1 capable:2 plotted:1 causal:1 isolated:2 increased:3 classify:1 earlier:1 conducted:1 eec:1 referring:1 peak:3 international:3 discriminating:1 physic:1 synthesis:1 connecting:1 hopkins:1 recorded:2 containing:1 huang:1 cognitive:1 american:1 warped:1 coefficient:1 configured:1 onset:1 performed:1 characterizes:1 portion:1 complicated:1 parallel:3 capability:1 voiced:1 accuracy:7 characteristic:1 percept:1 rabiner:2 weak:2 accurately:1 produced:2 marginally:1 cybernetics:1 submitted:1 anns:1 detector:2 fo:2 aligns:2 ed:1 evaluates:1 energy:7 frequency:7 pp:2 associated:1 handwriting:1 hamming:2 sampled:1 pitched:1 ask:1 fiat:1 focusing:2 response:4 improved:3 evaluated:1 though:3 working:1 propagation:1 logistic:1 gray:1 perhaps:1 building:1 effect:1 consisted:1 read:1 nettalk:1 speaker:3 hill:1 ridge:2 confusion:2 performs:1 common:2 discriminants:1 permuted:2 endpoint:4 khz:1 cambridge:1 smoothness:1 automatic:1 grid:2 language:1 stable:1 pu:1 lbc:1 feb:2 perspective:1 jolla:1 additional:2 maximize:1 signal:1 ii:2 full:1 sound:7 technical:1 serial:1 prediction:2 lly:1 normalization:1 background:1 decreased:1 else:1 pass:1 jordan:1 near:2 presence:1 ideal:2 iii:1 easy:2 fft:9 fit:1 perfectly:1 reduce:1 whether:1 rms:4 speech:13 york:1 nine:3 action:1 useful:3 discount:1 ten:3 mcclelland:1 wieland:1 notice:5 per:3 vol:3 recomputed:1 four:2 threshold:2 clarity:1 prevent:1 tenth:1 everywhere:1 almost:1 reasonable:1 reader:1 layer:12 resampled:1 syllable:2 fold:1 durbin:1 activity:1 constraint:1 fourier:2 relatively:1 jr:1 slightly:2 lp:1 conceivably:1 outlier:1 eventually:1 needed:1 eight:1 spectral:6 voicing:2 distinguished:1 diagonal:1 exhibit:1 gradient:1 separate:1 toward:3 rjk:1 length:2 index:1 providing:1 difficult:8 mostly:1 october:1 negative:1 twenty:4 perform:3 neuron:8 descent:1 situation:1 hinton:1 communication:1 assp:1 frame:5 ucsd:1 smoothed:4 pair:1 required:1 connection:2 discriminator:1 hanson:1 learned:3 trans:1 pattern:12 lpc:17 power:1 difficulty:1 recursion:1 scheme:1 axis:2 utterance:8 text:1 geometric:1 determining:1 relative:2 permutation:1 versus:1 degree:1 sufficient:1 consistent:1 token:20 english:1 perceptron:10 institute:2 neighbor:2 template:7 warp:1 distributed:2 slice:1 feedback:1 chiba:1 vocabulary:9 contour:3 made:1 san:3 approximate:1 lippmann:1 preferred:1 incoming:1 spectrum:12 terrain:1 table:3 disambiguate:1 learn:2 ca:4 noise:1 representative:2 papert:1 exceeding:2 late:1 learns:1 specific:1 sensing:1 list:1 ci:1 contained:1 corresponds:1 markel:1 man:1 hard:2 aided:1 included:1 determined:1 except:1 typical:1 springerverlag:1 called:1 total:1 discriminate:1 experimental:1 la:1 perceptrons:4 indicating:1 internal:1 correlated:1
7,113
880
Using a Saliency Map for Active Spatial Selective Attention: Implementation & Initial Results Shumeet Baluja baluja@cs.cmu.edu School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 Dean A. Pomerleau pomerleau@cs.cmu.edu School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 Abstract In many vision based tasks, the ability to focus attention on the important portions of a scene is crucial for good performance on the tasks. In this paper we present a simple method of achieving spatial selective attention through the use of a saliency map. The saliency map indicates which regions of the input retina are important for performing the task. The saliency map is created through predictive auto-encoding. The performance of this method is demonstrated on two simple tasks which have multiple very strong distracting features in the input retina. Architectural extensions and application directions for this model are presented. 1 MOTIVATION Many real world tasks have the property that only a small fraction of the available input is important at any particular time. On some tasks this extra input can easily be ignored. Nonetheless, often the similarity between the important input features and the irrelevant features is great enough to interfere with task performance. Two examples of this phenomena are the famous "cocktail party effect", otherwise known as speech recognition in a noisy environment, and image processing of a cluttered scene. In both cases, the extraneous information in the input signal can be easily confused with the important features, making the task much more difficult. The concrete real world task which motivates this work is vision-based road following. In this domain, the goal is to control a robot vehicle by analyzing the scene ahead, and choosing a direction to travel based on the location of important features like lane marking and road edges. This is a difficult task, since the scene ahead is often cluttered with extraneous features such as other vehicle, pedestrians, trees, guardrails, crosswalks, road signs and many other objects that can appear on or around a roadway. 1 While we have had significant success on the road following task using simple feed-forward neural networks to transform images of the road ahead into steering commands for the vehicle [pomerleau, 1993b], these methods fail when presented with cluttered environments like those encounI . For the gl:neral task of autonomous navigation, these extra features are extremely important, but for restricted task of road following. which is the focus of this paper. these features are merely distractions. Although we are addressing the more general task using the techniques described here in combination with other methods, a description of these efforts is beyond the scope of this paper. 452 Shumeet Baluja, Dean A. Pomerleall tered when driving in heavy traffic, or on city streets. The obvious solution to this difficulty is to focus the attention of the processing system on only the relevant features by masking out the "noise". Because of the high degree of simi1arity between the relevant features and the noise, this filtering is often extremely difficult. Simultaneously learning to perform a task like road following and filtering out clutter in a scene is doubly difficult because of a chicken-and-egg problem. It is hard to learn which features to attend to before knowing how to perform the task, and it is hard to learn how to perform the task before knowing which features to attend to. This paper describes a technique designed to solve this problem. It involves deriving a "saliency map" of the image from a neural network's internal representation which highlights regions of the scene considered to be important. This saliency map is used as feedback to focus the attention of the network's processing on subsequent images. This technique overcomes the chicken-and-egg problem by simultaneously learning to identify which aspects of the scene are important, and how to use them to perform a task. 2 THE SALIENCY MAP A saliency map is designed to indicate which portions of the image are important for completing the required task. The trained network should be able to accomplish two goals with the presentation of each image. The first is to perform the given task using the inputs and the saliency map derived from the previous image, and the second is to predict the salient portions of the next image. 2.1 Implementation The creation of the saliency map is similar to the technique of Input Reconstruction Reliability Estimation (IRRE) by [Pomerleau, 1993]. IRRE attempts to predict the reliability of a network's output. The prediction is made by reconstructing the input image from linear transformations of the activations in the hidden layer, and comparing it with the actual image. IRRE works on the premise that the greater the similarity between the input image and the reconstructed input image, the more the internal representation has captured the important input features, and therefore the more reliable the network's response. A similar method to IRRE can be used to create a saliency map. The saliency map should be determined by the important features in the current image for the task to be performed. Because compressed representations of the important features in the current image are represented in the activations of the hidden units, the saliency map is derived from these, as shown in Figure 1. It should be noted that the hidden units, from which the saliency map is derived, do not necessarily contain information similar to principal components (as is achieved through auto-encoder networks), as the relevant task may only require information on a small portion of the image. In the simple architecture depicted in Figure 1, the internal representation must contain information which can be transformed by a single layer of adjustable weights into a saliency map for the next image. If such a transformation is not possible, separate hidden layers, with input from the task-specific internal representations could be employed to create the saliency map. The saliency map is trained by using the next image, of a time-sequential data set, as the target image for the prediction, and applying standard error backpropagation on the differences between the next image and the predicted next image. The weights from the hidden Using a Saliency Map for Active Spatial Selective Attention (delayed I time step) Predicted Output Units Hidden Units L------! ~':~e ,....--=--., erived aliency p Input Retina x X ----+--- Relevant Portion of Input 453 Figure 1: A simple architecture for using a saliency map. The dashed "chilled line represents connections", i.e. errors from these connections do not propagate back further to impact the activations of the hidden units. This architecture assumes that the target task contains information which will help determine the salient portions of the next frame. units to the saliency map are adjusted using standard backpropagation, but the error terms are not propagated to the weights from the inputs to the hidden units. This ensures that the hidden representation developed by the network is determined only by the target task, and not by the task of prediction. In the implementation used here, the feedback is to the input layer. The saliency map is created to either be the same size as the input layer, or is scaled to the same size, so that it can influence the representation in a straight-forward manner. The saliency map's values are scaled between 0.0 and 1.0, where 1.0 represents the areas in which the prediction matched the next image exactly. The value of 1.0 does not alter the activation of the input unit, a value of 0.0 turns off the activation. The exact construction of the saliency map is described in the next section, with the description of the experiments. The entire network is trained by standard backpropagation; in the experiments presented, no modifications to the standard training algorithm were needed to account for the feedback connections. The training process for prediction is complicated by the potential presence of noise in the next image. The saliency map cannot "reconstruct" the noise in the next image, because it can only construct the portions of the next image which can be derived from the activation of the hidden units, which are task-specific. There/ore, the noise in the next image will not be constructed, and thereby will be de-emphasized in the next time step by the saliency map. The saliency map serves as a filter, which channels the processing to the important portions of the scene [Mozer, 1988]. One of the key differences between the filtering employed in this study, and that used in other focus of attention systems, is that this filtering is based on expectations from multiple frames, rather than on the retinal activations from a single frame. An alternative neural model of visual attention which was explored by [Olshausen et al., 1992] achieved focus of attention in single frames by using control neurons to dynamically modify synaptic strengths. The saliency map may be used in two ways. It can either be used to highlight portions of the input retina or, when the hidden layer is connected in a retinal fashion using weight sharing, as in [LeCun et al., 1990], it can be used to highlight important spatial locations within the hidden layer itself. The difference is between highlighting individual pixels from which the features are developed or highlighting developed features. Discussion of the psychological evidence for both of these types of highlighting (in single-frame retinal activation based context), is given in [pashler and Badgio, 1985]. This network architecture shares several characteristics with a Jordan-style recurrent network [Jordan, 1986], in which the output from the network is used to influence the pro- 454 Shumeet Baluja, Dean A. Pomerleau cessing of subsequent input patterns in a temporal sequence. One important distinction is that the feedback in this architecture is spatially constrained. The saliency map represents the importance of local patches of the input, and can influence only the network's processing of corresponding regions of the input. The second distinction is that the outputs are not general task outputs, rather they are specially constructed to predict the next image. The third distinction is in the form of this influence. Instead of treating the feedback as additional input units, which contribute to the weighted sum for the network's hidden units, this architecture uses the saliency map as a gating mechanism, suppressing or emphasizing the processing of various regions of the layer to which it is fed-back. In some respects, the influence of the saliency map is comparable to the gating network in the mixture of experts architecture [Jacobs et al., 1991]. Instead of gating between the outputs of multiple expert networks, in this architecture the saliency map is used to gate the activations of the input units within the same network. 3 THE EXPERIMENTS In order to study the feasibility of the saliency map without introducing other extraneous factors, we have conducted experiments with two simple tasks described below. Extensions of these ideas to larger problems are discussed in sections 4 & 5. The first experiment is representative of a typical machine vision task, in which the relevant features move very little in consecutive frames. With the method used here, the relevant features are automatically determined and tracked. However, if the relevant features were known a priori, a more traditional vision feature tracker which begins the search for features within the vicinity of the location of the features in the previous frame, could also perform well. The second task is one in which the feature of interest moves in a discontinuous manner. A traditional feature tracker without exact knowledge the feature's transition rules would be unable to track this feature, in the presence of the noise introduced. The transition rules of the feature of interest are learned automatically through the use of the saliency map. In the first task, there is a slightly tilted vertical line of high activation in a 30x32 input unit grid. The width of the line is approximately 9 units, with the activation decaying with distance from the center of the line. The rest of the image does not have any activation. The task is to center a gaussian of activation around the center of the x-intercept of the line, 5 pixels above the top of the image. The output layer contains 50 units. In consecutive images, the line can have a small translational move and/or a small rotational move. Sample training examples are given in Figure 2. This task can be easily learned in the presence of no noise. The task is made much harder when lines, which have the same visual appearance as the real line (in everything except for location and tilt) randomly appear in the image. In this case, it is vital that the network is able to distinguish between the real line and noise line by using information gathered from previous image(s). In the second task, a cross ("+") of size 5x5 appears in a 20x20 grid. There are 16 positions in which the cross can appear, as shown in Figure 2c. The locations in which the cross appears is set according to the transition rules shown in Figure 2c. The object of this problem is to reproduce the cross in a smaller lOx 10 grid, with the edges of the cross extended to the edges of the grid, as shown in Figure 2b. The task is complicated by the presence of randomly appearing noise. The noise is in the form of another cross which appears exactly similar to the cross of interest. Again, in this task, it is vital for the network to be able to distinguish between the real cross, and crosses which appear as noise. As in the first task, this is only possible with knowledge of the previous image(s). Using a Saliency Map for Active Spatial Selective Attention At 455 12 5 9 2 7 1 II 16 10 15 4 \3 3 14 6 8 A2 c Bt B2 Figure 2: (A) The first task, image (AI) with no distractions, image (A2) with one distracting feature. (8) The second task, image (8 I) with no distractions, image (82) with two distractions. (C) Transition rules for the second task. 3.1 Results The two problems described above were attempted with networks trained both with and without noise. Each of the training sessions were also tested with and without the saliency map. Each type of network was trained for the same number of pattern presentations with the same training examples. The results are shown in Table 1. The results reported in Table I represent the error accrued over 10,000 testing examples. For task 1, errors are reported in terms of the absolute difference between the peak's of the Gaussians produced in the output layer, summed for all of the testing examples (the max error per image is 49). In task 2, the errors are the sum of the absolute difference between the network's output and the target output, summed across all of the outputs and all of the testing examples. When noise was added to the examples, it was added in the following manner (for both training and testing sets): In task 1, '1 noise' guarantees a noise element, similarly, '2 noise' guarantees two noise elements. However, in task 2, '1 noise' means that there is a 50% chance of a noise element occurring in the image, '2 noise' means that there is a 50% chance of another noise element occurring, independently of the appearance of the first noise element. The positions of the noise elements are determined randomly. The best performance. in task 1, came from the cases in which there was no noise in testing or training. and no saliency map was used. This is expected, as this task is not difficult when no noise is present. Surprisingly, in task 2, the best case was found with the saliency map, when training with noise and testing without noise. This performed even better than training without noise. Investigation into this result is currently underway. In task 1, when training and testing without noise. the saliency map can hurt performance. If the predictions made by the saliency map are not correct, the inputs appear slightly distorted; therefore, the task to be learned by the network becomes more difficult. Nevertheless. the benefit of using a saliency map is apparent when the test set contains noise. In task 2. the network without the saliency map, trained with noise, and tested without noise cannot perform well; the performance further suffers when noise is introduced into the testing set. The noise in the training prevents accurate learning. This is not the case when the saliency map is used (Table 1, task 2). When the training set contains noise, the network with the saliency map works better when tested with and without noise. 456 Sizumeet Baluja. Dean A. Pomerleau Table 1: Summed Error of 10,000 Testing Examples Testing Set Training Set Task 1 o Noise 1 Noise Task 2 2 Noise o Noise I Noise 2 Noise o Noise (Saliency) 12672 60926 82282 7174 94333 178883 o Noise (No Saliency) 10241 91812 103605 7104 133496 216884 1 Noise (Saliency) 18696 26178 52668 4843 10427 94422 1 Noise (No Saliency) 14336 80668 97433 31673 150078 227650 When the noise increased beyond the level of training, to 2 noise elements per image, the performances of networks trained both with and without the saliency map declined. It is suspected that training the networks with increased noise will improve performance in the network trained with the saliency map. Nonetheless, due to the amount of noise compared to the small size of the input layer, improvements in results may not be dramatic. In Figure 3, a typical test run of the second task is shown. In the figure, the inputs, the predicted and actual outputs, and the predicted and actual saliency maps, are shown. The actual saliency map is just a smaller version of the unfiltered next input image. The input size is 20x20, the outputs are 10xlO, and the saliency map is 10xlO. The saliency map is scaled to 20x20 when it is applied to the next inputs. Note that in the inputs to the network, one cross appears much brighter than the other; this is due to the suppression of the distracting cross by the saliency map. The particular use of the saliency map which is employed in this study, proceeded as follows: the difference between the saliency map (derived from input imagei) and the input imagei+l was calculated. This difference image was scaled to the range of 0.0 to 1.0. Each pixel was then passed through a sigmoid; alternatively, a hard-step function could have been used. This is the saliency map. The saliency map was multiplied by input imagei+ 1; this was used as the input into the network. If the sigmoid is not used, features, such as incomplete crosses, sometimes appear in the input image. This happens because different portions of the cross may have slightly different saliency values associated with them -due to errors in prediction coupled with the scaling of the saliency map. Although the sigmoid helps to alleviate the problem, it does not eliminate it. This explains why training with no noise with a saliency map sometimes does not perform as well as training without a saliency map. 4 ALTERNATIVE IMPLEMENTATIONS An alternative method of implementing the saliency map is with standard additive connections. However, these connection types have several drawbacks in comparison with the multiplicative ones use in this study. First, the additive connections can drastically change the meaning of the hidden unit's activation by changing the sign of the activation. The saliency map is designed to indicate which regions are important for accomplishing the task based upon the features in the hidden representations; as little alteration of the important features as possible is desired. Second, if the saliency map is incorrect, and suggests an area of which is not actually important, the additive connections will cause 'ghost' images to appear. These are activations which are caused only by the influence of the addi- Using a Saliency Map for Active Spatial Selective Attention 457 predicted saliency target saliency predicted output target output inputs 1 4 5 6 7 Figure 3: A typical sequence of inputs and outputs in the second task. Note that when two crosses are in the inputs. one is much brighter than the other. The "noise" cross is de-emphasized. tive saliency map. The multiplicative saliency map, as is implemented here, does not have either of these problems. A second alternative, which is more closely related to a standard recurrent network [Jordan, 1986], is to use the saliency map as extra inputs into the network. The extra inputs serve to indicate the regions which are expected to be important. Rather than hand-coding the method to represent the importance of the regions to the network, as was done in this study, the network learns to use the extra inputs when necessary. Further, the saliency map serves as the predicted next input. This is especially useful when the features of interest may have momentarily been partially obscured or have entirely disappeared from the image. This implementation is currently being explored by the authors for use in a autonomous road lane-tracking system in which the lane markers are not always present in the input image. 5 CONCLUSIONS & FUTURE DIRECTIONS These experiments have demonstrated that an artificial neural network can learn to both identify the portions of a visual scene which are important, and to use these important features to perform a task. The selective attention architecture we have develop uses two simple mechanisms, predictive auto-encoding to form a saliency map, and a constrained form of feedback to allow this saliency map to focus processing in the subsequent image. There are at least four broad directions in which this research should be extended. The first is, as described here, related to using the saliency map as a method for automatically actively focusing attention to important portions of the scene. Because of the spatial dependence of the task described in this paper, with the appropriate transformations, the output units could be directly fed back to the input layer to indicate saliency. Although this does not weaken the result, in terms of the benefit of using a saliency map, future work should also focus on problems which do not have this property to determine how easily a saliency map can be constructed. Will the use of weight sharing be enough to develop the necessary spatially oriented feature detectors? Harder problems are those with target tasks which does not explicitly contain spatial saliency information. An implementation problem which needs to be resolved is in networks which contain more than a single hidden layer: from which layer should the saliency map be con- 458 Shumeet Baluja, Dean A. Pomerleau structed? The trade-off is that at the higher layers, the information contained is more task specific. However, the higher layers may effectively extract the information required to perform the task, without maintaining the information required for saliency map creation. The opposite case is true in the lower layers; these may contain all of the information required, but may not provide enough discrimination to narrow the focus effectively. The third area for research is an alternative use for the saliency map. ANNs have often been criticized for their uninterpretability, and lack of mechanism to explain performance. The saliency map provides a method for understanding, at a high level, what portions of the inputs the ANN finds the most important. Finally, the fourth direction for research is the incorporation of higher level, or symbolic knowledge. The saliency map provides a very intuitive and direct method for focusing the network's attention to specific portions of the image. The saliency map may prove to be a useful mechanism to allow other processes, including human users, to simply "point at" the portion of the image to which the network should be paying attention. The next step in our research is to test the effectiveness of this technique on the main task of interest, autonomous road following. Fortunately, the first demonstration task employed in this paper shares several characteristics with road following. Both tasks require the network to track features which move over time in a cluttered image. Both tasks also require the network to produce an output that depends on the positions of the important features in the image. Because of these shared characteristics, we believe that similar performance improvements should be possible in the autonomous driving domain. Acknowledgments Shumeet Baluja is supported by a National Science Foundation Graduate Fellowship. Support has come from "Perception for Outdoor Navigation" (contract number DACA76-89-C-0014, monitored by the US Army Topographic Engineering Center) and "Unmanned Ground Vehicle System" (contract number DAAE07-90-C-R059, monitored by TACOM). Support has also come from the National Highway Traffic Safety Administration under contract number DTNH22-93-C-07023. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing official policies, either expressed or implied, of the National Science Foundation, ARPA, or the U.S. Government. References Cottrell, G.W. & Munro, P. (1988) Principal Component Analysis ofImages via back-propagation. Proc Soc. of Photo-Opticallnstr. Eng., Cambridge, MA. Jordan, M.I., (1989). Serial Order: A Parallel, Distributed Processing Approach. In Advances in Connectionist Theory: Speech, eds. J.L. Elman and D.E. Rumerlhart. Hillsdale: Erlbaum. Jacobs, R.A., Jordan, M.I., Nowlan, S.J. & Hinton, G.E. (1991). Adaptive Mixtures of Local Experts. Neural Computation, 3:1. LeCun, Y., Boser, B., Denker, J.S., Henderson, D. Howard, R.E., Hurnmand W., and Jackel, L.D. (1989) Backpropagation Applied to Handwritten Zip Code Recognition. Neural Computation 1,541-551. MIT, 1989. Mozer, M.C. (1988) A Connectionist Model of Selective Attention in Visual Perception. Technical Report, University of Toronto, CRG-TR-88-4. Pashler, H. & Badgio, P. (1985). Visual Attention and Stimulus Identification. Journal of Experimental Psychology: Human Perception and Performance, II 105-121. Pomerleau, D.A. (1993) Input Reconstruction Reliability Estimation. In Giles, C.L. Hanson, S.J. and Cowan, J.D. (eds). Advances in Neurallnfol71Ultion Processing Systems 5, CA: Morgan Kaufmann Publishers. Pomerleau, D.A. (1993b) Neural Network Perception for Mobile Robot Guidance, Kluwer Academic Publishing. Olshausen, B., Anderson, C., & Van Essen; D. (1992) A Neural Model of Visual Attention and Invariant Pattern Recognition. California Institute of Technology, CNS Program, memo-18.
880 |@word proceeded:1 version:1 propagate:1 jacob:2 eng:1 dramatic:1 thereby:1 tr:1 harder:2 initial:1 contains:4 document:1 suppressing:1 current:2 comparing:1 nowlan:1 activation:16 must:1 tilted:1 subsequent:3 additive:3 cottrell:1 designed:3 treating:1 discrimination:1 provides:2 contribute:1 location:5 toronto:1 constructed:3 direct:1 incorrect:1 prove:1 doubly:1 manner:3 expected:2 elman:1 automatically:3 actual:4 little:2 becomes:1 confused:1 begin:1 matched:1 what:1 interpreted:1 developed:3 transformation:3 guarantee:2 temporal:1 exactly:2 scaled:4 control:2 unit:17 appear:7 before:2 safety:1 attend:2 shumeet:5 modify:1 local:2 engineering:1 encoding:2 analyzing:1 approximately:1 dynamically:1 suggests:1 range:1 graduate:1 acknowledgment:1 lecun:2 testing:10 backpropagation:4 area:3 road:10 symbolic:1 cannot:2 context:1 influence:6 applying:1 intercept:1 pashler:2 map:76 dean:5 demonstrated:2 center:4 attention:18 cluttered:4 independently:1 x32:1 rule:4 deriving:1 autonomous:4 hurt:1 target:7 construction:1 user:1 exact:2 us:2 pa:2 element:7 recognition:3 region:7 ensures:1 connected:1 momentarily:1 trade:1 mozer:2 environment:2 trained:8 predictive:2 creation:2 upon:1 serve:1 easily:4 resolved:1 represented:1 various:1 artificial:1 choosing:1 apparent:1 larger:1 solve:1 otherwise:1 compressed:1 encoder:1 ability:1 reconstruct:1 addi:1 topographic:1 transform:1 noisy:1 itself:1 sequence:2 reconstruction:2 relevant:7 description:2 intuitive:1 produce:1 disappeared:1 object:2 help:2 recurrent:2 develop:2 school:2 paying:1 strong:1 soc:1 implemented:1 c:2 involves:1 indicate:4 predicted:7 come:2 direction:5 drawback:1 discontinuous:1 correct:1 filter:1 closely:1 human:2 everything:1 implementing:1 explains:1 require:3 premise:1 government:1 hillsdale:1 investigation:1 alleviate:1 daca76:1 adjusted:1 extension:2 crg:1 around:2 considered:1 tracker:2 ground:1 great:1 scope:1 predict:3 driving:2 consecutive:2 a2:2 estimation:2 proc:1 travel:1 currently:2 jackel:1 highway:1 create:2 city:1 weighted:1 mit:1 gaussian:1 always:1 rather:3 mobile:1 command:1 derived:5 focus:9 improvement:2 indicates:1 suppression:1 entire:1 bt:1 eliminate:1 hidden:16 selective:7 transformed:1 reproduce:1 pixel:3 translational:1 extraneous:3 priori:1 spatial:8 constrained:2 summed:3 construct:1 represents:3 broad:1 alter:1 future:2 connectionist:2 report:1 stimulus:1 retina:4 randomly:3 oriented:1 simultaneously:2 national:3 individual:1 delayed:1 cns:1 attempt:1 interest:5 essen:1 henderson:1 navigation:2 mixture:2 accurate:1 edge:3 necessary:2 tree:1 incomplete:1 desired:1 obscured:1 guidance:1 weaken:1 psychological:1 increased:2 criticized:1 arpa:1 giles:1 introducing:1 addressing:1 conducted:1 erlbaum:1 reported:2 accomplish:1 accrued:1 peak:1 contract:3 off:2 concrete:1 again:1 expert:3 style:1 actively:1 account:1 potential:1 de:2 retinal:3 alteration:1 b2:1 coding:1 pedestrian:1 caused:1 explicitly:1 depends:1 vehicle:4 performed:2 multiplicative:2 view:1 traffic:2 portion:15 decaying:1 complicated:2 parallel:1 masking:1 irre:4 accomplishing:1 kaufmann:1 characteristic:3 gathered:1 saliency:85 identify:2 famous:1 handwritten:1 identification:1 produced:1 straight:1 anns:1 detector:1 explain:1 suffers:1 sharing:2 synaptic:1 ed:2 nonetheless:2 obvious:1 associated:1 monitored:2 con:1 tered:1 propagated:1 knowledge:3 actually:1 back:4 appears:4 feed:1 focusing:2 higher:3 response:1 done:1 anderson:1 just:1 hand:1 marker:1 lack:1 propagation:1 interfere:1 believe:1 olshausen:2 effect:1 contain:5 true:1 vicinity:1 spatially:2 x5:1 width:1 noted:1 distracting:3 pro:1 image:50 meaning:1 sigmoid:3 tracked:1 tilt:1 discussed:1 kluwer:1 mellon:2 significant:1 cambridge:1 ai:1 cessing:1 grid:4 session:1 similarly:1 had:1 reliability:3 robot:2 similarity:2 irrelevant:1 success:1 came:1 captured:1 morgan:1 greater:1 additional:1 steering:1 fortunately:1 employed:4 zip:1 determine:2 signal:1 dashed:1 ii:2 multiple:3 ofimages:1 technical:1 academic:1 cross:15 serial:1 feasibility:1 impact:1 prediction:7 vision:4 cmu:2 expectation:1 represent:2 sometimes:2 achieved:2 chicken:2 fellowship:1 ore:1 crucial:1 publisher:1 extra:5 rest:1 specially:1 cowan:1 effectiveness:1 jordan:5 presence:4 vital:2 enough:3 psychology:1 brighter:2 architecture:9 opposite:1 idea:1 knowing:2 administration:1 munro:1 passed:1 effort:1 speech:2 cause:1 cocktail:1 ignored:1 useful:2 amount:1 clutter:1 sign:2 track:2 per:2 carnegie:2 key:1 salient:2 four:1 nevertheless:1 achieving:1 changing:1 merely:1 fraction:1 sum:2 run:1 fourth:1 distorted:1 architectural:1 patch:1 scaling:1 comparable:1 entirely:1 layer:17 completing:1 distinguish:2 strength:1 ahead:3 incorporation:1 scene:10 lane:3 aspect:1 extremely:2 performing:1 marking:1 according:1 combination:1 neral:1 describes:1 slightly:3 reconstructing:1 smaller:2 across:1 making:1 modification:1 happens:1 restricted:1 invariant:1 turn:1 fail:1 mechanism:4 needed:1 fed:2 serf:2 photo:1 available:1 gaussians:1 multiplied:1 denker:1 appropriate:1 appearing:1 alternative:5 gate:1 assumes:1 top:1 publishing:1 maintaining:1 especially:1 implied:1 move:5 added:2 dependence:1 traditional:2 distance:1 separate:1 unable:1 street:1 code:1 rotational:1 demonstration:1 difficult:6 x20:3 memo:1 implementation:6 pomerleau:9 motivates:1 policy:1 adjustable:1 perform:10 vertical:1 neuron:1 howard:1 extended:2 hinton:1 frame:7 introduced:2 tive:1 required:4 connection:7 unmanned:1 hanson:1 california:1 distinction:3 learned:3 narrow:1 boser:1 beyond:2 able:3 below:1 pattern:3 perception:4 ghost:1 program:1 reliable:1 max:1 including:1 difficulty:1 representing:1 improve:1 technology:1 declined:1 created:2 lox:1 auto:3 coupled:1 extract:1 understanding:1 underway:1 highlight:3 filtering:4 unfiltered:1 foundation:2 degree:1 suspected:1 share:2 heavy:1 gl:1 surprisingly:1 supported:1 drastically:1 allow:2 institute:1 absolute:2 benefit:2 distributed:1 feedback:6 calculated:1 van:1 world:2 transition:4 forward:2 made:3 author:2 adaptive:1 party:1 reconstructed:1 overcomes:1 active:4 pittsburgh:2 alternatively:1 search:1 why:1 table:4 learn:3 channel:1 ca:1 necessarily:1 domain:2 official:1 main:1 motivation:1 noise:52 representative:1 egg:2 fashion:1 position:3 outdoor:1 third:2 learns:1 emphasizing:1 specific:4 emphasized:2 gating:3 explored:2 evidence:1 structed:1 sequential:1 effectively:2 importance:2 occurring:2 depicted:1 simply:1 appearance:2 army:1 visual:6 highlighting:3 prevents:1 contained:2 expressed:1 xlo:2 partially:1 tracking:1 chance:2 ma:1 goal:2 presentation:2 ann:1 shared:1 hard:3 change:1 baluja:7 determined:4 typical:3 except:1 principal:2 experimental:1 attempted:1 distraction:4 internal:4 support:2 tested:3 phenomenon:1
7,114
881
Analysis of Unstandardized Contributions in Cross Connected Networks Thomas R. Shultz Yuriko Oshima-Takane Yoshio Takane shultz@psych.mcgill.ca yuriko@psych.mcgill.ca takane@psych.mcgill.ca Department of Psychology McGill University Montreal, Quebec, Canada H3A IBI Abstract Understanding knowledge representations in neural nets has been a difficult problem. Principal components analysis (PCA) of contributions (products of sending activations and connection weights) has yielded valuable insights into knowledge representations, but much of this work has focused on the correlation matrix of contributions. The present work shows that analyzing the variance-covariance matrix of contributions yields more valid insights by taking account of weights. 1 INTRODUCTION The knowledge representations learned by neural networks are usually difficult to understand because of the non-linear properties of these nets and the fact that knowledge is often distributed across many units. Standard network analysis techniques, based on a network's connection weights or on its hidden unit activations, have been limited. Weight diagrams are typically complex and weights vary across mUltiple networks trained on the same problem. Analysis of activation patterns on hidden units is limited to nets with a single layer of hidden units without cross connections. Cross connections are direct connections that bypass intervening hidden unit layers. They increase learning speed in static networks by focusing on linear relations (Lang & Witbrock, 1988) and are a standard feature of generative algorithms such as cascadecorrelation (Fahlman & Lebiere, 1990). Because such cross connections do so much of the work, analyses that are restricted to hidden unit activations furnish only a partial picture of the network's knowledge. Contribution analysis has been shown to be a useful technique for multi-layer, cross connected nets. Sanger (1989) defined a contribution as the product of an output weight, the activation of a sending unit, and the sign of the output target for that input. Such contributions are potentially more informative than either weights alone or hidden unit activations alone since they take account of both weight and sending activation. Shultz and Elman (1994) used PCA to reduce the dimensionality of such contributions in several different types of cascade-correlation nets. Shultz and Oshima-Takane (1994) demonstrated that PCA of unscaled contributions produced even better insights into cascade-correlation solutions than did comparable analyses of contributions scaled by the sign of output targets. Sanger (1989) had recommended scaling contributions by the signs of output targets in order to determine whether the contributions helped or hindered the network's solution. But since the signs of output targets are only available to networks during error 602 Thomas R. Shultz, Yuriko Oshima-Takane, Yoshio Takane correction learning, it is more natural to use unscaled contributions in analyzing knowledge representations. There is an issue in PCA about whether to use the correlation matrix or the variancecovariance matrix. The correlation matrix contains Is in the diagonal and Pearson correlation coefficients between contributions off the diagonal. This has the effect of standardizing the variables (contributions) so that each has a mean of 0 and standard deviation of 1. Effectively, this ensures that the PCA of a correlation matrix exploits variation in input activation patterns but ignores variation in connection weights (because variation in connection weights is eliminated as the contributions are standardized). Here, we report on work that investigates whether more useful insights into network knowledge structures can be revealed by PCA of un standardized contributions. To do this, we apply PCA to the variance-covariance matrix of contributions. The variance-covariance matrix has contribution variances along the diagonal and covariances between contributions off the diagonal. Taking explicit account of the variation in connection weights in this way may produce a more valid picture of the network's knowledge. We use some of the same networks and problems employed in our earlier work (Shultz & Elman, 1994; Shultz & Oshima-Takane, 1994) to facilitate comparison of results. The problems include continuous XOR, arithmetic comparisons involving addition and mUltiplication, and distinguishing between two interlocking spirals. All of the nets were generated with the cascade-correlation algorithm (Fahlman & Lebiere, 1990). Cascade-correlation begins as a perceptron and recruits hidden units into the network as it needs them in order to reduce error. The recruited hidden unit is the one whose activations correlate best with the network's current error. Recruited units are installed in a cascade, each on a separate layer and receiving input from the input units and from any previously existing hidden units. We used the default values for all cascade-correlation parameters. The goal of understanding knowledge representations learned by networks ought to be useful in a variety of contexts. One such context is cognitive modeling, where the ability of nets to merely simulate psychological phenomena is not sufficient (McCloskey, 1991). In addition, it is important to determine whether the network representations bear any systematic relation to the representations employed by human subjects . 2 PCA OF CONTRIBUTIONS Sanger's (1989) original contribution analysis began with a three-dimensional array of contributions (output unit x hidden unit x input pattern). In contrast, we start with a twodimensional output weight x input pattern array of contributions. This is more efficient than the slicing technique used by Sanger to focus on particular output or hidden units and still allows identification of the roles of specific contributions (Shultz & Elman, 1994; Shultz & Oshima-Takane, 1994). We subject the variance-covariance matrix of contributions to PCA in order to identify the main dimensions of variation in the contributions (Jolliffe, 1986). A component is a line of best fit to a set of data points in multi-dimensional space. The goal of PCA is to summarize a multivariate data set with a relatively small number of components by capitalizing on covariance among the variables (in this case, contributions). We use the scree test (Cattell, 1966) to determine how many components are useful to include in the analysis. Varimax rotation is applied to improve the interpretability of the solution. Component scores are plotted to identify the function of each component 3 APPLICATION TO CONTINUOUS XOR The classical binary XOR problem does not have enough training patterns to make contribution analysis worthwhile. However, we constructed a continuous version of the XOR problem by dividing the input space into four quadrants. Starting from 0.1, input values were incremented in steps of 0.1, producing 100 x, y input pairs that can be partitioned into four quadrants of the input space. Quadrant a had values of x less than Analysis of Unstandardized Contributions in Cross Connected Networks 603 0.55 combined with values of y above 0.55. Quadrant b had values of x and y greater than 0.55. Quadrant c had values of x and y less than 0.55. Quadrant d had values of x greater than 0.55 combined with values of y below 0.55. Similar to binary XOR, problems from quadrants a and d had a positive output target (0.5) for the net, whereas problems from quadrants band c had a negative output target (-0.5). There was a single output unit with a sigmoid activation. Three cascade-correlation nets were trained on continuous XOR. Each of these nets generated a unique solution, recruiting five or six hidden units and taking from 541 to 765 epochs to learn to correctly classify all of the input patterns. Generalization to test patterns not in the training set was excellent. PCA of unscaled, unstandardized contributions yielded three components. A plot of rotated component scores for the 100 training patterns of net 1 is shown in Figure 1. The component scores are labeled according to their respective quadrant in the input space. Three components are required to account for 96.0% of the variance in the contributions. Figure 1 shows that component 1, with 44.3% of the variance in contributions, has the role of distinguishing those quadrants with a positive output target (a and d) from those with a negative output target (b and c). This is indicated by the fact that the black shapes are at the top of the component space cube in Figure 1 and the white shapes are at the bottom. Components 2 and 3 represent variation along the x and y input dimensions, respectively. Component 2 accounted for 26.1 % of the variance in contributions, and component 3 accounted for 25.6% of the variance in contributions. Input pairs from quadrants b and d (square shapes) are concentrated on the negative end of component 2, whereas input pairs from quadrants a and c (circle shapes) are concentrated on the positive end of component 2. Similarly, input pairs from quadrants a and b cluster on the negative end of component 3, and input pairs from quadrants c and d cluster on the positive end of component 3. Although the network was not explicitly trained to represent the x and y input dimensions, it did so as an incidental feature of its learning the distinction between quadrants a and d vs. quadrants band c. Similar results were obtained from the other two nets learning the continuous XOR problem. In contrast, PCA of the correlation matrix from these nets had yielded a somewhat less clear picture with the third component separating quadrants a and d from quadrants b and c, and the first two components representing variation along the x and y input dimensions (Shultz & Oshima-Takane, 1994). PCA of the correlation matrix of scaled contributions had performed even worse, with plots of component scores indicating interactive separation of the four quadrants, but with no clear roles for the individual components (Shultz & Elman, 1994). Standardized, rotated component loadings for net 1 are plotted in Figure 2. Such plots can be examined to determine the role played by each contribution in the network. For example, hidden units 2, 3, and 4 all playa major role in the job done by component 1, distinguishing positive from negative outputs. 4 APPLICATION TO COMPARATIVE ARITHMETIC Arithmetic comparison requires a net to conclude whether a sum or a product of two integers is greater than, less than, or equal to a comparison integer. Several psychological simulations have used neural nets to make additive and multiplicative comparisons and this has enhanced interest in this type of problem (McClelland, 1989; Shultz, Schmidt, Buckingham, & Mareschal, in press). The first input unit coded the type of arithmetic operation to be performed: 0 for addition and 1 for multiplication. Three additional linear input units encoded the integers. Two of these input units each coded a randomly selected integer in the range of 0 to 9, inclusive; another input unit coded a randomly selected comparison integer. For addition problems, comparison integers ranged from 0 to i9, inclusive; for multiplication, comparison integers ranged from 0 to 82, inclusive. Two sigmoid output units coded the results of the comparison operation. Target outputs of 0.5, -0.5 represented a greater than result, targets of -0.5, 0.5 represented less than, and targets of 0.5,0.5 represented equal to. 604 Thomas R. Shultz, Yuriko Oshima-Takane, Yoshio Takane 2 Component 1 o -1 3 2 Component 2 -2 Component 3 -2 Figure 1. Rotated component scores for a continuous XOR net. Component scores for the x, y input pairs in quadrant a are labeled with black circles, those from quadrant b with white squares, those from quadrant c with white circles, and those from quadrant d with black squares. The network's task is to distinguish pairs from quadrants a and d (the black shapes) from pairs from quadrants b and c (the white shapes). Some of the white shapes appear black because they are so densely packed, but all of the truly black shapes are relatively high in the cube. Hidden6 Hidden5 Hidden4 c: 0 '5 Hidden3 C Hidden2 :g 0 () Hidden1 Component III 3 II 2 Input2 ? Input1 -1.0 -0.5 0.0 0.5 1.0 Loading Figure 2. Standardized, rotated component loadings for a continuous XOR net. Rotated loadings were standardized by dividing them by the standard deviation of the respective contribution scores. Analysis of Unstandardized Contributions in Cross Connected Networks 605 The training patterns had 100 addition and 100 multiplication problems, randomly selected, with the restriction that 45 of each had correct answers of greater than, 45 of each had correct answers of less than, and 10 of each had correct answers of equal to. These constraints were designed to reduce the natural skew of comparative values in the high direction on multiplication problems. We ran three nets for 1000 epochs each, at which point they were very close to mastering the training patterns. Either seven or eight hidden units were recruited along the way. Generalization to previously unseen test problems was very accurate. Four components were sufficient to account for most the variance in un standardized contributions, 88.9% in the case of net 1. Figure 3 displays the rotated component scores for the first two components of net 1. Component I, accounting for 51.1 % of the variance, separated problems with greater than answers from problems with less than answers, and located problems with equal to answers in the middle, at least for addition problems. Component 2, with 20.2% of the variance, clearly separated multiplication from addition. Contributions from the first input unit were strongly associated with component 2. Similar results obtained for the other two nets. Components 3 and 4, with 10.6% and 7.0% of the variance, were sensitive to variation in the second and third inputs, respectively. This is supported by an examination of the mean input values of the 20 most extreme component scores on these two components. Recall that the second and third inputs coded the two integers to be added or multiplied. The negative end of component 3 had a mean second input value of 8.25; the positive end of this component had a mean second input value of 0.55. Component 4 had mean third input value of 2.00 on the negative end and 7.55 on the positive end. In contrast, PCA of the correlation matrix for these nets had yielded a far more clouded picture, with the largest components focusing on input variation and lesser components doing bits and pieces of the separation of answer types and operations in an interactive manner (ShUltz & Oshima-Takane, 1994). Problems with equal to answers were not isolated by any of the components. PCA of scaled contributions had produced three components that interactively separated the three answer types and operations, but failed to represent variation in input integers (ShUltz & Elman, 1994). Essentially similar advantages for using the variance-covariance matrix were found for nets learning either addition alone or multiplication alone. 5 APPLICATION TO THE TWO-SPIRALS PROBLEM The two-spirals problem requires a particularly difficult discrimination and a large number of hidden units. The input space is defined by two interlocking spirals that wrap around their origin three times. There are two sets of 97 real-valued x, y pairs, with each set representing one of the spirals, and a single sigmoid output unit coded for the identity of the spiral. Our three nets took between 1313 and 1723 epochs to master the distinction, and recruited from 12 to 16 hidden units. All three nets generalized well to previously unseen input pairs on the paths of the two spirals. PCA of the variance-covariance matrix for net 1 revealed that six components accounted for a total of 97.9% of the variance in contributions. The second and fourth of these components together distinguished one spiral from the other, with 20.7% and 9.8% of the variance respectively. Rotated component scores for these two components are plotted in Figure 4. A diagonal line drawn on Figure 4 from coordinates -2,2 to 2, -2 indicates that 11 points from each spiral were misclassified by components 2 and 4. This is only 11.3% of the data points in the training patterns. The fact that the net learned all of the training patterns implies that these exceptions were picked up by other components. Components 1 and 6, with 40.7% and 6.4% of the variance, were sensitive to variation in the x and y inputs, respectively. Again, this was confirmed by the mean input values of the 20 most extreme component scores on these two components. On component I, the negative end had a mean x value of 3.55 and the positive end had a mean y value of -3.55. 606 Thomas R. Shultz. Yuriko Oshima-Takane. Yoshio Takane 2 x> It" #~ ? . ?x Component 1 o -1 +< -2 L..-_ _......L_ _ _....L..._ _--iL..-_ _-J -2 -1 0 2 Component 2 Figure 3. Rotated component scores for an arithmetic comparison net. Greater than problems are symbolized by circles, less than problems by squares, addition by white shapes, and multiplication by black shapes. For equal to problems only, addition is represented by + and multiplication by X. Although some densely packed white shapes may appear black, they have no overlap with truly black shapes. All of the black squares are concentrated around coordinates -1, -1. 2 Component 2 o o o -1 o o Spiral 1 -2 1....-_ _- - '_ _ _......._ _ _........_ __ -2 -1 o 2 Component 4 Figure 4. Rotated component scores for a two-spirals net. Squares represent data points from spiral 1, and circles represent data points from spiral 2. Analysis of Unstandardized Contributions in Cross Connected Networks 607 On component 6, the negative end had a mean x value of 2.75 and the positive end had a mean y value of -2.75. The skew-symmetry of these means is indicative of the perfectly symmetrical representations that cascade-correlation nets achieve on this highly symmetrical problem. Every data point on every component has a mirror image negative with the opposite signed component score on that same component. This -x, -y mirror image point is always on the other spiral. Other components concentrated on particular regions of the spirals. The other two nets yielded essentially similar results. These results can be contrasted with our previous analyses of the two-spirals problem, none of which succeeded in showing a clear separation of the two spirals. PCAs based on scaled (Shultz & Elman, 1994) or unscaled (Shultz & Oshima-Takane, 1994) correlation matrices showed extensive symmetries but never a distinction between one spiral and another.1 Thus, although it was clear that the nets had encoded the problem's inherent symmetries, it was still unclear from previous work how the nets used this or other information to distinguish points on one spiral from points on the other spiral. 6 DISCUSSION On each of these problems, there was considerable variation among network solutions, as revealed, for example, by variation in numbers of hidden units recruited and signs and sizes of connection weights. In spite of such variation, the present technique of applying peA to the variance-covariance matrix of contributions yielded results that are sufficiently abstract to characterize different nets learning the same problem. The knowledge representations produced by this analysis clearly identify the essential information that the net is being trained to utilize as well as more incidental features of the training patterns such as the nature of the input space. This research strengthens earlier conclusions that PCA of network contributions is a useful technique for understanding network performance (Sanger, 1989), including relatively intractable multi-level cross connected nets (Shultz & Elman, 1994; Shultz & Oshima-Takane, 1994). However, the current study underscores the point that there are several ways to prepare a contribution matrix for PCA, not all of which yield equally valid or useful results. Rather than starting with a three dimensional matrix of output unit x hidden unit x input pattern and focusing on either one output unit at a time or one hidden unit at a time (Sanger, 1989), it is preferable to collapse contributions into a two dimensional matrix of output weight x input pattern. The latter is not only more efficient, but yields more valid results that characterize the network as a whole, rather than small parts of the network. Also, rather than scaling contributions by the sign of the output target (Sanger, 1989), it is better to use unsealed contributions. Unsealed contributions are not only more realistic, since the network has no knowledge of output targets during its feed-forward phase, but also produce clearer interpretations of the nefs knowledge representations (Shultz & Oshima-Takane, 1994). The latter claim is particularly true in terms of sensitivity to input dimensions and to operational distinctions between adding and multiplying. Plots of component scores based on unscaled contributions are typically not as dense as those based on sealed contributions but are more revealing of the network's knowledge. Finally, rather than applying peA to the correlation matrix of contributions, it makes more sense to apply it to the variance-covariance matrix. As noted in the introduction, using the correlation matrix effectively standardizes the contributions to have identical means and variances, thus obseuring the role of network connection weights. The present results indicate much clearer knowledge representations when the variance-covariance matrix is used since connection weight information is explicitly retained. Matrix differences were especially marked on the more difficult problems, such as two-spirals, where the only peAs to reveal how nets distinguished the spirals were those based on 1Results from un scaled contributions on the two-spirals problem were not actually presented in Shultz & Oshima-Takane (1994) since they were not very clear. 608 Thomas R. Shultz, Yuriko Oshima-Takane, Yoshio Takane variance-covariance matrices. But the relative advantages of using the variance-covariance matrix were evident on the easier problems too. There has been recent rapid progress in the study of the knowledge representations leamed by neural nets. Feed-forward nets can be viewed as function approximators for relating inputs to outputs. Analysis of their knowledge representations should reveal how inputs are encoded and transformed to produce the correct outputs. PCA of network contributions sheds light on how these function approximations are done. Components emerging from PCA are orthonormalized ingredients of the transformations of inputs that produce the correct outputs. Thus, PCA helps to identify the nature of the required transformations. Further progress might be expected from combining PCA with other matrix decomposition techniques. Constrained PCA uses external information to decompose multivariate data matrices before applying PCA (Takane & Shibayama, 1991). Analysis techniques emerging from this research will be useful in understanding and applying neural net research. Component loadings, for example, could be used to predict the results of lesioning experiments with neural nets. Once the role of a hidden unit has been identified by virtue of its association with a particular component, then one could predict that lesioning this unit would impair the function served by the component. Acknowledgments This research was supported by the Natural Sciences and Engineering Research Council of Canada. References Cattell, R. B. (1966). The scree test for the number of factors. Multivariate Behavioral Research, 1,245-276. Fahlman, S. E., & Lebiere, C. (1990.) The Cascade-Correlation learning architecture. In D. Touretzky (Ed.), Advances in neural information processing systems 2, (pp. 524532). Mountain View, CA: Morgan Kaufmann. Jolliffe, I. T. (1986). Principal component analysis. Berlin: Springer Verlag. Lang, K. J., & Wi tbrock , M. J. (1988). Learning to tell two spirals apart. In D. Touretzky, G. Hinton, & T. Sejnowski (Eds)., Proceedings of the Connectionist Models Summer School, (pp. 52-59). Mountain View, CA: Morgan Kaufmann. McClelland, J. L. (1989). Parallel distributed processing: Implications for cognition and development. In Morris, R. G. M. (Ed.), Parallel distributed processing: Implications for psychology and neurobiology, pp. 8-45. Oxford University Press. McCloskey, M. (1991). Networks and theories: The place of connectionism in cognitive science. Psychological Science, 2, 387-395. Sanger, D. (1989). Contribution analysis: A technique for assigning responsibilities to hidden units in connectionist networks. Connection Science, 1, 115-138. Shultz, T. R., & Elman, J. L. (1994). Analyzing cross connected networks. In J. D. Cowan, G. Tesauro, & J. Alspector (Eds.), Advances in Neural Information Processing Systems 6. San Francisco, CA: Morgan Kaufmann. ShUltz, T. R., & Oshima-Takane, Y. (1994). Analysis of un scaled contributions in cross connected networks. In Proceedings of the World Congress on Neural Networks (Vol. 3, pp. 690-695). Hillsdale, NJ: Lawrence Erlbaum. Shultz, T. R., Schmidt, W. C., Buckingham, D., & Mareschal, D. (In press). Modeling cognitive development with a generative connectionist algorithm. In G. Halford & T. Simon (Eds.), Developing cognitive competence: New approaches to process modeling. Hillsdale, NJ: Erlbaum. Takane, Y., & Shibayama, T. (1991). Principal component analysis with external information on both subjects and variables. Psychometrika, 56, 97-120.
881 |@word middle:1 version:1 loading:5 simulation:1 covariance:13 accounting:1 decomposition:1 contains:1 score:15 existing:1 current:2 activation:10 lang:2 buckingham:2 assigning:1 realistic:1 additive:1 informative:1 shape:12 plot:4 designed:1 v:1 alone:4 generative:2 selected:3 discrimination:1 indicative:1 leamed:1 five:1 along:4 constructed:1 direct:1 behavioral:1 manner:1 expected:1 rapid:1 alspector:1 elman:8 multi:3 psychometrika:1 begin:1 mountain:2 psych:3 emerging:2 recruit:1 transformation:2 nj:2 ought:1 every:2 interactive:2 shed:1 preferable:1 scaled:6 unit:37 appear:2 producing:1 positive:9 before:1 engineering:1 congress:1 installed:1 analyzing:3 oxford:1 path:1 black:10 ibi:1 signed:1 might:1 examined:1 limited:2 collapse:1 range:1 unique:1 acknowledgment:1 cascade:9 revealing:1 quadrant:25 spite:1 lesioning:2 close:1 twodimensional:1 context:2 applying:4 restriction:1 interlocking:2 demonstrated:1 starting:2 focused:1 slicing:1 insight:4 array:2 variation:14 coordinate:2 mcgill:4 target:13 enhanced:1 distinguishing:3 us:1 origin:1 strengthens:1 particularly:2 located:1 labeled:2 bottom:1 role:7 region:1 ensures:1 connected:8 incremented:1 valuable:1 ran:1 trained:4 represented:4 unsealed:2 separated:3 sejnowski:1 tell:1 pearson:1 whose:1 encoded:3 valued:1 ability:1 unseen:2 advantage:2 net:42 took:1 product:3 combining:1 achieve:1 intervening:1 cluster:2 produce:4 comparative:2 rotated:9 help:1 montreal:1 clearer:2 school:1 progress:2 job:1 dividing:2 implies:1 indicate:1 direction:1 hidden2:1 correct:5 pea:3 human:1 hillsdale:2 generalization:2 decompose:1 connectionism:1 correction:1 around:2 sufficiently:1 lawrence:1 cognition:1 predict:2 claim:1 recruiting:1 major:1 vary:1 prepare:1 sensitive:2 council:1 largest:1 orthonormalized:1 clearly:2 always:1 rather:4 focus:1 indicates:1 underscore:1 contrast:3 sense:1 typically:2 hidden:21 relation:2 misclassified:1 transformed:1 issue:1 among:2 development:2 constrained:1 cube:2 equal:6 once:1 never:1 eliminated:1 identical:1 report:1 connectionist:3 inherent:1 randomly:3 densely:2 individual:1 phase:1 oshima:15 interest:1 variancecovariance:1 highly:1 truly:2 extreme:2 light:1 implication:2 accurate:1 succeeded:1 partial:1 respective:2 circle:5 plotted:3 isolated:1 psychological:3 classify:1 earlier:2 modeling:3 witbrock:1 deviation:2 erlbaum:2 too:1 characterize:2 answer:9 combined:2 sensitivity:1 systematic:1 off:2 receiving:1 input2:1 together:1 again:1 interactively:1 worse:1 cognitive:4 external:2 account:5 standardizing:1 coefficient:1 explicitly:2 piece:1 performed:2 helped:1 multiplicative:1 picked:1 view:2 doing:1 responsibility:1 start:1 parallel:2 simon:1 contribution:59 square:6 il:1 xor:9 variance:24 kaufmann:3 yield:3 identify:4 identification:1 produced:3 none:1 multiplying:1 confirmed:1 pcas:1 served:1 touretzky:2 ed:5 pp:4 lebiere:3 associated:1 static:1 recall:1 knowledge:16 dimensionality:1 mareschal:2 actually:1 focusing:3 feed:2 done:2 strongly:1 correlation:19 reveal:2 indicated:1 facilitate:1 effect:1 ranged:2 true:1 furnish:1 white:7 shultz:26 during:2 noted:1 generalized:1 evident:1 image:2 began:1 sigmoid:3 rotation:1 association:1 interpretation:1 relating:1 sealed:1 similarly:1 had:23 playa:1 multivariate:3 showed:1 recent:1 apart:1 tesauro:1 verlag:1 binary:2 approximators:1 morgan:3 greater:7 somewhat:1 additional:1 employed:2 determine:4 recommended:1 arithmetic:5 ii:1 multiple:1 h3a:1 cross:11 equally:1 coded:6 involving:1 essentially:2 represent:5 addition:10 whereas:2 diagram:1 subject:3 recruited:5 cowan:1 quebec:1 yuriko:6 integer:9 input1:1 revealed:3 iii:1 spiral:24 enough:1 variety:1 fit:1 psychology:2 architecture:1 perfectly:1 opposite:1 hindered:1 reduce:3 identified:1 lesser:1 whether:5 six:2 pca:24 useful:7 clear:5 band:2 morris:1 concentrated:4 mcclelland:2 sign:6 correctly:1 halford:1 vol:1 four:4 drawn:1 utilize:1 merely:1 sum:1 master:1 fourth:1 place:1 separation:3 i9:1 scaling:2 investigates:1 comparable:1 bit:1 layer:4 summer:1 played:1 distinguish:2 display:1 yielded:6 symbolized:1 constraint:1 inclusive:3 speed:1 simulate:1 relatively:3 department:1 developing:1 according:1 across:2 cascadecorrelation:1 mastering:1 partitioned:1 wi:1 restricted:1 previously:3 skew:2 jolliffe:2 end:12 sending:3 capitalizing:1 available:1 operation:4 multiplied:1 apply:2 eight:1 worthwhile:1 distinguished:2 schmidt:2 thomas:5 original:1 standardized:6 scree:2 include:2 top:1 sanger:8 exploit:1 especially:1 classical:1 added:1 diagonal:5 unclear:1 wrap:1 separate:1 separating:1 berlin:1 seven:1 takane:23 retained:1 difficult:4 potentially:1 negative:10 incidental:2 packed:2 varimax:1 cattell:2 hinton:1 neurobiology:1 competence:1 canada:2 pair:10 required:2 extensive:1 connection:13 learned:3 distinction:4 impair:1 usually:1 pattern:15 below:1 summarize:1 interpretability:1 including:1 overlap:1 natural:3 examination:1 representing:2 improve:1 picture:4 yoshio:5 epoch:3 understanding:4 multiplication:9 relative:1 bear:1 ingredient:1 sufficient:2 bypass:1 unscaled:5 accounted:3 supported:2 fahlman:3 l_:1 understand:1 perceptron:1 taking:3 distributed:3 dimension:5 default:1 valid:4 world:1 ignores:1 forward:2 san:1 far:1 correlate:1 symmetrical:2 conclude:1 francisco:1 un:4 continuous:7 learn:1 nature:2 ca:6 operational:1 symmetry:3 excellent:1 complex:1 did:2 main:1 dense:1 whole:1 explicit:1 third:4 specific:1 showing:1 virtue:1 essential:1 intractable:1 adding:1 effectively:2 mirror:2 easier:1 failed:1 mccloskey:2 springer:1 goal:2 identity:1 marked:1 viewed:1 considerable:1 contrasted:1 principal:3 total:1 indicating:1 exception:1 latter:2 phenomenon:1
7,115
882
Effects of Noise on Convergence and Generalization in Recurrent Networks Kam Jim Bill G. Horne c. Lee Giles* NEC Research Institute, Inc., 4 Independence Way, Princeton, NJ 08540 {kamjim,horne,giles}~research.nj.nec.com *Also with UMIACS, University of Maryland, College Park, MD 20742 Abstract We introduce and study methods of inserting synaptic noise into dynamically-driven recurrent neural networks and show that applying a controlled amount of noise during training may improve convergence and generalization. In addition, we analyze the effects of each noise parameter (additive vs. multiplicative, cumulative vs. non-cumulative, per time step vs. per string) and predict that best overall performance can be achieved by injecting additive noise at each time step. Extensive simulations on learning the dual parity grammar from temporal strings substantiate these predictions. 1 INTRODUCTION There has been much research in applying noise to neural networks to improve network performance. It has been shown that using noisy hidden nodes during training can result in error-correcting codes which increase the tolerance of feedforward nets to unreliable nodes (Judd and Munro, 1992) . Also, randomly disabling hidden nodes during the training phase increases the tolerance of MLP's to node failures (Sequin and Clay, 1990). Bishop showed that training with noisy data is equivalent to Tikhonov Regularization and suggested directly minimizing the regularized error function as a practical alternative (Bishop , 1994) . Hanson developed a stochastic version of the delta rule which adapt weight means and standard deviations instead 650 Kam Jim, Bill G. Home, C. Lee Giles of clean weight values (Hanson, 1990). (Mpitsos and Burton, 1992) demonstrated faster learning rates by adding noise to the weight updates and adapting the magnitude of such noise to the output error. Most relevant to this paper, synaptic noise has been applied to MLP's during training to improve fault tolerance and training quality. (Murray and Edwards, 1993) In this paper, we extend these results by introducing several methods of inserting synaptic noise into recurrent networks, and demonstrate that these methods can improve both convergence and generalization. Previous work on improving these two performance measures have focused on ways of simplifying the network and methods of searching the coarse regions of state space before the fine regions. Our work shows that synaptic noise can improve convergence by searching for promising regions of state space, and enhance generalization by enforcing saturated states. 2 NOISE INJECTION IN RECURRENT NETWORKS In this paper, we inject noise into a High Order recurrent network (Giles et al., 1992) consisting of N recurrent state neurons Sj, L non-recurrent input neurons lie, and N 2 L weights Wijle. (For justification ofits use see Section 4.) The recurrent network g(Lj,1e WijleSJlD, where g(.) is a operation is defined by the state process S:+1 sigmoid discriminant function. During training, an error function is computed as Ep where Sb -dp, Sb is the output neuron, and dp is the target output value tor pattern p. = = !f;, fp = Synaptic noise has been simulated on Multi-Layered-Perceptrons by inserting noise to the weights of each layer during training (Murray et al., 1993). Applying this method to recurrent networks is not straightforward because effectively the same weights are propagated forward in time. This can be seen by recalling the BPTT representation of unrolling a recurrent network in time into T layers with identical weights, where T is the length of the input string. In Tables 2 and 3, we introduce the noise injection steps for eight recurrent network noise models representing all combinations of the following noise parameters: additive vs. multiplicative, cumulative vs. non-cumulative, per time step vs. per string. As their name imply, additive and multiplicative noise add or multiply the weights by a small noise term. In cumulative noise, the injected noise is accumulated, while in non-cumulative noise the noise from the current step is removed before more noise is injected on the next step. Per time step and per string noise refer to when the noise is inserted: either at each time step or only once for each training string respectively. Table 1 illustrates noise accumulation examples for all additive models (the multiplicative case is analogous). 3 ANALYSIS ON THE EFFECTS OF SYNAPTIC NOISE The effects of each noise model is analyzed by taking the Taylor expansion on the error function around the noise-free weight set. By truncating this expansion to second and lower order terms, we can interpret the effect of noise as a set of regularization terms applied to the error function. From these terms predictions can be made about the effects on generalization and convergence. A similar analysis was Effects of Noise on Convergence and Generalization in Recurrent Networks 651 performed on MLP's to demonstrate the effects of synaptic noise on fault tolerance and training quality (Murray et. al., 1993). Tables 2 and 3 list the noise injection step and the resulting first and second brder Taylor expansion terms for all noise models. These results are derived by assuming the noise to be zero-mean white with variance (F2 and uncorrelated in time. 3.1 Predictions on Generalization One common cause of bad generalization in recurrent networks is the presence of unsaturated state representations. Typically, a network cannot revisit the exact same point in state space, but tends to wander away from its learned state representation. One approach to alleviate this problem is to encourage state nodes to operate in the saturated regions of the sigmoid. The first order error expansion terms of most noise models considered are capable of encouraging the network to achieve saturated states. This can be shown by applying the chain rule to the partial derivative in the first order expansion terms: (1) where e~ is the net input to state node i at time step t. The partial derivatives g~ favor internal representations such that the effects of perturbations to the net are minimized. inputs e: Multiplicative noise implements a form of weight decay because the error expansion terms include the weight products Wt~ijk or Wt ,ijk Wu,ijk' Although weight decay has been shown to improve generalization on feedforward networks (Krogh and Hertz, 1992) we hypothesize this may not be the case for recurrent networks that are learning to solve FSA problems. Large weights are necessary to saturate the state nodes to the upper and lower limits of the sigmoid discriminant function. Therefore, we predict additive noise will allow better generalization because of its absence of weight decay. att Sb . Noise models whose first order error term contain the expression a~rb i , l,k ",'mn. will favor saturated states for those partials whose sign correspond to the sign of a majority of the partials. It will favor unsaturated states, operating in the linear region of the sigmoid, for partials whose sign is the minority. Such sign-dependent enforcement is not optimal. The error terms for cumulative per time step noises sum a product with the expresSb . Sb ,where = min(t + 1, U + 1). The effect of cumulative noise sion lm " increases more rapidly because of v and thus optimal generalization and detrimental noise effects will occur at lower amplitudes than non-cumulative noise. va:.r,.". a:.t v For cumulative per string noise models, the products (t+ l)(u+ 1) and Wt,ijk Wu,lmn in the expansion terms rapidly overwhelm the raw error term. Generalization improvement is not expected for these models. We also reason that all generalization enhancements will be valid only for a range of noise values, above which noise overwhelms the raw error information. 652 3.2 Kam Jim, Bill G. Horne, C. Lee Giles Predictions on Convergence Synaptic noise can improve convergence by favoring promising weights in the beginning stages of training. This can be demonstrated by examining the second order error expansion term for non-cumulative, multiplicative, per time step noise: When fp is negative, solutions with a negative second order state-weight partial derivative will be de-stabilized. In other words, when the output Sb is too small the network will favor updating in a direction such that the first order partial derivative is increasing. A corresponding relationship can be observed for the case when fp is positive. Thus the second order term of the error function will allow a higher raw error fp to be favored if such an update will place the weights in a more promising area, i.e. a region where weight changes are likely to move Sb in a direction to reduce the raw error. The anticipatory effect of this term is more important in the beginning stages of training where fp is large, and will become insignificant in the finishing stages of training as fp approaches zero. Similar to arguments in Section 3.1, the absence of weight decay will make the learning task easier and improve convergence. From this discussion it can be inferred that additive per time step noise models should yield the best generalization and convergence performance because of their sign-independent favoring of saturated states and the absence of weight decay. Furthermore, convergence and generalization performance is more sensitive to cumulative noise, i.e. optimal performance and detrimental effects will occur at lower amplitudes than in non-cumulative noise. 4 SIMULATION RESULTS In order to perform many experiments in a reasonable amount of computation time, we attempt to learn the simple "hidden-state" dual parity automata from sample strings encoded as temporal sequences. (Dual parity is a 4-state automata that recognizes binary strings containing an even number of ones and zeroes.) We choose a second-order recurrent network since such networks have demonstrated good performance on such problems (Giles et. al., 1992). Thus our experiments consist of 500 simulations for each data point and achieve useful (90%) confidence levels. Experiments are performed with both 3 and 4 state networks, both of which are adequate to learn the automata. The learning rate and momentum are set to 0.5, and the weights are initialized to random values between [-1.0, 1.0]. The data consists of 8191 strings of lengths 0 to 12. The networks are trained on a subset of the training set, called the working set, which gradually increases in size until the entire training set is classified correctly. Strings from the working set are presented in alphabetical order. The training set consists of the first 1023 strings of lengths 0 to 9, while the initial working set consists of 31 strings of lengths 0 to 4. During testing no noise is added to the weights of the network. Effects of Noise on Convergence and Generalization in Recurrent Networks 653 300rT-----,--~--rT----,_----. e ? ?? ???? ? ? b 2150 1 i 200 . ? S; ~150 ~ ~OO . 150 . ~ 2 TraininG Nole. Std Oev 3 o 0 0_15 ~ ~.15 Traln'nG Nol_. Std Oev 2 Figure 1: Best Convergence/Generalization for Additive and Multiplicative Noises. (a) multiplicative non-cumulative per time step; (b) additive cumulative per time step. 4.1 Convergence and Generalization Performance Simulated performance closely mirror our predictions. Improvements were observed for all noise models except for cumulative per string noises which failed to converge for all runs. Generalization improvement was more emphasized on networks with 4 states, while convergence enhancement was more noticeable on 3-state networks. The simulations show the following results: ? Additive noise is better tolerated than multiplicative noise, and achieves better convergence and generalization (Figure 1). ? Cumulative noise achieves optimal generalization and convergence at lower amplitudes than non-cumulative noise. Cumulative noise also has a narrower range of beneficial noise, which is defined as the range of noise amplitudes which yields better performance than that of a noiseless network (Figure 2a illustrates this for generalization). ? Per time step noise achieves better convergence/generalization and has a wider range of beneficial values than per string noise (Figure 2b). Overall, the best performance is obtained by applying cumulative and noncumulative additive noise at each time step. These results closely match the predictions of section 3.1. The only exceptions are that all multiplicative noise models seem to yield equivalent performance. This discrepancy between prediction and simulation may be due to the detrimental effects of weight decay in multiplicative noise , which can conflict with the advantages of cumulative and per time step noise. 4.2 The Payoff Picture: Generalization vs. Convergence Generalization vs. Convergence results are plotted in Figure 3. Increasing noise amplitudes proceed from the left end-point of each curve to the right end-point. 654 Kam Jim, Bill G. Horne, C. Lee Giles Table 1: Examples: Additive Noise Accumulation. ~i is the noise at time step ti TIME STEPS NOISE MODEL per time step non-cumulative per time step cumulative per sequence non-cumulative per sequence cumulative t1 t2 t3 W+~l W+~2 W +~1 +~2 W+~l W+2~1 W+~3 W+~l W+~l W+~l W ... ... +~1 +~2 + ~3 W+~l W+3~1 ... ... ... 300rT------r---r-~r---~----_. 2150 j200 .5 , 1150 ~ 100 1 00 1 2 Training Not_. S'td D_v Figure 2: (I) Best Generalization for Cumulative and Non-Cumulative Noises: a) cumulative additive per time step; b) non-cumulative additive per time step. (II) Best Generalization for Per Time Step and Per String Noises: a) non-cumulative per string additive; b) non-cumulative per time step additive. e e 15.15 5.5 5 .) a. .. .. .. 15 4 .5 ~ ~ .... i ~ ~ II 3.15 ~ 'l\J 1>5 Iii ~ 2.5 3 4 3 .5 3 CJ 2 .5 2 2 c 1.15 ~oo 1 .5 200 300 Convergence In epocha 400 ~OO 200 300 Convergence 'n epochs 400 Figure 3: Payoff: Mean Generalization vs. Convergence for 4-state (I) and 3state(lI) recurrent network. (I a) Worst i-state - non-cumulative multiplicative per string; (Ib, Ic) Best 4-state - cumulative and non-cumulative additive per time step, respectively; (lIa) Worst 3-state - non-cumulative multiplicative per string; (lib) Best 3-state - cumulative additive per time step. Effects of Noise on Convergence and Generalization in Recurrent Networks 655 Table 2: Noise injection step and error expansion terms for per time step nOIse models. v = min(t + 1, U + 1). W'" is the noise-free weight set. NON-CUMULATIVE Addlhve MU bpncahve Noise step l.tt order 2nd order ClIM~LATIVE Additive M ulhplicati ve t Noise step Wt,ijk = wtjl< + L t AT,iJI< = wtjl< Wt,ijl< T-O II (1 + AThl<) T-O l.tt order L T-1 ~~ ~2 2 P 2nd order ~ T ~ "Wt ,.JI< .. W ._ So Lu,.JI< 8W .. 8W .. .. t,.JI< u,.JI< t ,u=O .JI< NON-CUMULATIVE AddItIve Multlpllcahve Noise step 1 1st order -~ 2 2 1 2 P 2nd order -~ T-1 2: L T W .. t,u=O .JI< T 1 ~2 L 2: " t,.JI< T 850 u,.JI< 8W . - .. t,.JI< 8S0 8W . . _2 W t,u=O ijl< ;UM ILA IVE Additive W u,.JI< T crSO .. W .. t,.JI< u,.JI< 8W .. W . . t,.Jk u,.JI< Multiplicative Noise step Wt,ijl< = W i jl?l + Aijl<)' t = wijl< + 2: Ct,T(Aijl<)T+1 T-O 2: 2: TIT 1st order 1 2 -~ 2 .. t,u=O .JI< 0 w s85 -- T 85 0 8Wt ,ijl< 8Wu ,ijl< T-1 -2: 1 2 " .. " ~ L- t,.JI< t,u=O ijl<,lmn T-1 2~~ +2~p T T 85 85 u,lmn _ 8W0._ . 8W 0 t,.JI< u,lmn T L- L"t ' i j l8Wt <--.. ,ijl< 8S 0 t=O .JI< 2: L T-1 2nd order 1 -~ 2 P .. t,u=O .JI<,lmn _2 "t,.JI< .. "u ,lmn 8w .' t,.JI< T crS O 8W u,lmn 656 Kam Jim, Bill G. Home, C. Lee Giles These plots illustrate the cases where both convergence and generalization are improved. In figure 311 the curves clearly curl down and to the left for lower noise amplitudes before rising to the right at higher noise amplitudes. These lower regions are important because they represent noise values where generalization and convergence improve simultaneously and do not trade off. 5 CONCLUSIONS We have presented several methods of injecting synaptic noise to recurrent neural networks. We summarized the results of an analysis of these methods and empirically tested them on learning the dual parity automaton from strings encoded as temporal sequences. (For a complete discussion of results, see (Jim, Giles, and Horne, 1994) ). Results show that most of these methods can improve generalization and convergence simultaneously - most other methods previously discussed in literature cast convergence as a cost for improved generalization performance. References [1] Chris M. Bishop. Training with noise is equivalent to Tikhonov Regularization. Neural Computation, 1994. To appear. [2] Robert M. Burton, Jr. and George J. Mpitsos. Event-dependent control of noise enhances learning in neural networks. Neural Networks, 5:627-637, 1992. [3] C.L. Giles, C.B. Miller, D. Chen, H.H. Chen, G.Z . Sun, and Y.C. Lee. Learning and extracting finite state automata with second-order recurrent neural networks. Neural Computation, 4(3):393-405, 1992. [4] Stephen Jose Hanson. A stochastic version ofthe delta rule. Physica D., 42:265272, 1990. [5] Kam Jim, C.L. Giles, and B.G. Horne. Synaptic noise in dynamically-driven recurrent neural networks: Convergence and generalization. Technical Report UMIACS-TR-94-89 and CS-TR-3322, Institute for Advanced Computer Studies, University of Maryland, College Park, MD, 1994. [6] Stephen Judd and Paul W . Munro. Nets with unreliable hidden nodes learn error-correcting codes. In S.J Hanson, J.D. Cowan, and C.L. Giles, editors, Advances in Neural Information Processing Systems 5, pages 89-96, San Mateo, CA, 1993. Morgan Kaufmann Publishers. [7] Anders Krogh and John A. Hertz. A simple weight decay can improve generalization. In J .E. Moody, S.J. Hanson, and R.P. Lippmann, editors, Advances in Neural Information Processing Systems 4, pages 450-957, San Mateo, CA, 1992. Morgan Kaufmann Publishers. [8] Alan F. Murray and Peter J. Edwards. Synaptic weight noise during multilayer perceptron training: Fault tolerance and training improvements. IEEE Trans. on Neural Networks, 4(4):722-725, 1993. [9] Carlo H. Sequin and Reed D. Clay. Fault tolerance in artificial neural networks. In Proc. of IJCNN, volume I, pages 1-703-708, 1990.
882 |@word version:2 rising:1 nd:4 bptt:1 simulation:5 simplifying:1 aijl:2 tr:2 initial:1 att:1 current:1 com:1 john:1 additive:21 hypothesize:1 plot:1 update:2 v:9 beginning:2 coarse:1 node:8 become:1 consists:3 introduce:2 expected:1 multi:1 td:1 encouraging:1 unrolling:1 increasing:2 lib:1 horne:6 string:20 developed:1 nj:2 temporal:3 ti:1 um:1 control:1 appear:1 before:3 positive:1 t1:1 tends:1 limit:1 mateo:2 dynamically:2 range:4 practical:1 testing:1 alphabetical:1 implement:1 area:1 adapting:1 word:1 confidence:1 ila:1 cannot:1 layered:1 applying:5 accumulation:2 bill:5 equivalent:3 demonstrated:3 straightforward:1 truncating:1 automaton:5 focused:1 correcting:2 rule:3 searching:2 justification:1 analogous:1 target:1 exact:1 s85:1 sequin:2 jk:1 updating:1 std:2 ep:1 inserted:1 observed:2 burton:2 worst:2 region:7 sun:1 trade:1 removed:1 mu:1 trained:1 overwhelms:1 tit:1 f2:1 artificial:1 whose:3 encoded:2 solve:1 ive:1 grammar:1 favor:4 noisy:2 fsa:1 sequence:4 advantage:1 net:4 product:3 inserting:3 relevant:1 rapidly:2 achieve:2 convergence:29 enhancement:2 wider:1 oo:3 recurrent:21 illustrate:1 noticeable:1 disabling:1 krogh:2 edward:2 c:1 direction:2 closely:2 stochastic:2 generalization:35 alleviate:1 physica:1 around:1 considered:1 ic:1 predict:2 lm:1 tor:1 achieves:3 proc:1 injecting:2 sensitive:1 clearly:1 cr:1 sion:1 derived:1 finishing:1 improvement:4 dependent:2 anders:1 accumulated:1 sb:6 lj:1 typically:1 entire:1 hidden:4 favoring:2 overall:2 dual:4 favored:1 once:1 ng:1 identical:1 park:2 discrepancy:1 minimized:1 t2:1 report:1 randomly:1 simultaneously:2 ve:1 phase:1 consisting:1 recalling:1 attempt:1 mlp:3 multiply:1 saturated:5 analyzed:1 chain:1 encourage:1 partial:7 capable:1 necessary:1 kam:6 taylor:2 initialized:1 plotted:1 giles:12 cost:1 introducing:1 deviation:1 subset:1 examining:1 too:1 tolerated:1 st:2 lee:6 off:1 enhance:1 moody:1 containing:1 choose:1 inject:1 derivative:4 li:1 de:1 summarized:1 inc:1 multiplicative:14 performed:2 analyze:1 variance:1 kaufmann:2 miller:1 correspond:1 yield:3 t3:1 ofthe:1 raw:4 lu:1 carlo:1 classified:1 synaptic:11 failure:1 propagated:1 iji:1 cj:1 clay:2 amplitude:7 higher:2 unsaturated:2 improved:2 anticipatory:1 furthermore:1 stage:3 until:1 working:3 quality:2 name:1 effect:16 contain:1 regularization:3 white:1 during:8 substantiate:1 ijl:7 tt:2 demonstrate:2 complete:1 sigmoid:4 common:1 ji:20 empirically:1 volume:1 jl:1 extend:1 discussed:1 interpret:1 refer:1 curl:1 operating:1 add:1 showed:1 driven:2 tikhonov:2 binary:1 fault:4 seen:1 morgan:2 george:1 converge:1 ii:3 stephen:2 alan:1 technical:1 faster:1 adapt:1 match:1 controlled:1 va:1 prediction:7 multilayer:1 noiseless:1 represent:1 achieved:1 addition:1 fine:1 publisher:2 operate:1 umiacs:2 cowan:1 seem:1 extracting:1 presence:1 feedforward:2 iii:1 independence:1 reduce:1 expression:1 munro:2 peter:1 proceed:1 cause:1 adequate:1 useful:1 amount:2 stabilized:1 revisit:1 sign:5 delta:2 per:31 rb:1 correctly:1 clean:1 sum:1 run:1 jose:1 injected:2 place:1 reasonable:1 wu:3 home:2 layer:2 ct:1 occur:2 ijcnn:1 argument:1 min:2 injection:4 combination:1 hertz:2 beneficial:2 jr:1 gradually:1 previously:1 overwhelm:1 enforcement:1 ofits:1 end:2 operation:1 eight:1 away:1 alternative:1 include:1 recognizes:1 murray:4 move:1 added:1 rt:3 md:2 enhances:1 dp:2 detrimental:3 maryland:2 simulated:2 majority:1 w0:1 chris:1 discriminant:2 reason:1 enforcing:1 minority:1 assuming:1 code:2 length:4 relationship:1 reed:1 minimizing:1 robert:1 negative:2 perform:1 upper:1 neuron:3 finite:1 payoff:2 jim:7 perturbation:1 inferred:1 cast:1 extensive:1 hanson:5 conflict:1 learned:1 trans:1 suggested:1 pattern:1 fp:6 event:1 regularized:1 advanced:1 mn:1 representing:1 improve:11 imply:1 picture:1 epoch:1 literature:1 wander:1 s0:1 editor:2 uncorrelated:1 lmn:7 parity:4 free:2 allow:2 perceptron:1 institute:2 taking:1 tolerance:6 curve:2 judd:2 valid:1 cumulative:38 forward:1 made:1 san:2 lia:1 sj:1 wijl:1 lippmann:1 unreliable:2 table:5 promising:3 learn:3 ca:2 improving:1 expansion:9 noise:93 paul:1 momentum:1 lie:1 ib:1 saturate:1 down:1 bad:1 bishop:3 emphasized:1 list:1 decay:7 insignificant:1 consist:1 adding:1 effectively:1 mirror:1 nec:2 magnitude:1 illustrates:2 chen:2 easier:1 likely:1 failed:1 narrower:1 absence:3 change:1 except:1 wt:8 called:1 ijk:5 perceptrons:1 exception:1 college:2 internal:1 princeton:1 tested:1
7,116
883
Pairwise Neural Network Classifiers with Probabilistic Outputs David Price A2iA and ESPCI 3 Rue de l'Arrivee, BP 59 75749 Paris Cedex 15, France a2ia@dialup.francenet.fr Stefan Knerr ESPCI and CNRS (UPR AOOO5) 10, Rue Vauquelin, 75005 Paris, France knerr@neurones.espci.fr Leon Personnaz, Gerard Dreyfus ESPeI, Laboratoire d'Electronique 10, Rue Vauquelin, 75005 Paris, France dreyfus@neurones.espci.fr Abstract Multi-class classification problems can be efficiently solved by partitioning the original problem into sub-problems involving only two classes: for each pair of classes, a (potentially small) neural network is trained using only the data of these two classes. We show how to combine the outputs of the two-class neural networks in order to obtain posterior probabilities for the class decisions. The resulting probabilistic pairwise classifier is part of a handwriting recognition system which is currently applied to check reading. We present results on real world data bases and show that, from a practical point of view, these results compare favorably to other neural network approaches. 1 Introduction Generally, a pattern classifier consists of two main parts: a feature extractor and a classification algorithm. Both parts have the same ultimate goal, namely to transform a given input pattern into a representation that is easily interpretable as a class decision. In the case of feedforward neural networks, the interpretation is particularly easy if each class is represented by one output unit. For many pattern recognition problems, it suffices that the classifier compute the class of the input pattern, in which case it is common practice to associate the pattern to the class corresponding to the maximum output of the classifier. Other problems require graded (soft) decisions, such as probabilities, at the output of the 1110 David Price, Stefan Knerr, Leon Personnaz, Gerard Dreyfus classifier for further use in higher context levels: in speech or character recognition for instance, the probabilistic outputs of the phoneme (character) recognizer are often used by a Hidden-Markov-Model algorithm or by some other dynamic programming algorithm to compute the most probable word hypothesis. In the context of classification, it has been shown that the minimization of the Mean Square Error (MSE) yields estimates of a posteriori class probabilities [Bourlard & Wellekens, 1990; Duda & Hart, 1973]. The minimization can be performed by a feedforward multilayer perceptrons (MLP's) using the backpropagation algorithm, which is one of the reasons why MLP's are widely used for pattern recognition tasks. However, MLPs have well-known limitations when coping with real-world problems, namely long training times and unknown architecture. In the present paper, we show that the estimation of posterior probabilities for a K-class problem can be performed efficiently using estimates of posterior probabilities for K(K-1 )/2 two-class sub-problems. Since the number of sub-problems increases as K2, this procedure was originally intended for applications involving a relatively small number of classes, such as the 10 classes for the recognition of handwritten digits [Knerr et aI., 1992]. In this paper we show that this approach is also viable for applications with K? 10. The probabilistic pairwise classifier presented in this paper is part of a handwriting recognition system, discussed elsewhere [Simon, 1992], which is currently applied to check reading. The purpose of our character recognizer is to classify pre-segmented characters from cursive handwriting. The probabilistic outputs of the recognizer are used to estimate word probabilities. We present results on real world data involving 27 classes, compare these results to other neural network approaches, and show that our probabilistic pairwise classifier is a powerful tool for computing posterior class probabilities in pattern recognition problems. 2 Probabilistic Outputs from Two-class Classifiers Multi-class classification problems can be efficiently solved by "divide and conquer" strategies which partition the original problem into a set of K(K-l)/2 two-class problems. For each pair of classes (OJ and (OJ, a (potentially small) neural network with a single output unit is trained on the data of the two classes [Knerr et aI., 1990, and references therein]. In this section, we show how to obtain probabilistic outputs from each of the two-class classifiers in the pairwise neural network classifier (Figure 1). K(K-I)12 two-class networks inputs Figure 1: Pairwise neural network classifier. Pairwise Neural Network Classifiers with Probabilistic Outputs 1111 It has been shown that the \llinimization of the MSE cost function (or likewise a cost function based on an entropy measure, [Bridle, 1990]) leads to estimates of posterior probabilities. Of course, the quality of the estimates depends on the number and distribution of examples in the training set and on the minimization method used. In the theoretical case of two classes <01 and <02, each Gaussian distributed, with means m 1 and m2, a priori probabilities Pq and Pr2, and equal covariance matrices ~, the posterior probability of class <01 given the pattern x is: Pr(class=<o\ I X=x) = _ _ _ _ _ _ _ _ _ _--'1'----_ _ _ _ _ _ _ _ __ 1 + Pr2 exp( _ !-(2xT~-\(m\-m2) + m!~-\m2 - m T~-lm\)) PrJ 2 (1) Thus a single neuron with a sigmoidal transfer function can compute the posterior probabilities for the two classes. However, in the case of real world data bases, classes are not necessarily Gaussian distributed, and therefore the transformation of the K(K-l )/2 outputs of our pairwise neural network classifier to posterior probabilities proceeds in two steps. In the first step, a class-conditional probability density estimation is performed on the linear output of each two-class neural network: for both classes <OJ and <OJ of a given twoclass neural network, we fit the probability density over Vjj (the weighted sum of the inputs of the output neuron) to a function. We denote by <Ojj the union of classes <OJ and <OJ. The resulting class-conditional densities p(vij I <OJ) and p(Vjj I <OJ) can be transformed to probabilities Pr(<Oj I <OJ' /\ (Vij=Vjj? and Pr(<Oj I <Ojj /\ (Vjj=Vjj? via the Bayes rule (note that Pr(<Ojj /\ (Vij=Vjj) <OJ) = Pr?Vij=Vjj) I <OJ)): 1 p( VjJ" I <OJ) Pr( <OJ) Pr(<Oj I <Ojj/\(Vij=Vij? = - - - - - " - - - - - p(Vjj I <Ok) Pr(<Ok) ke{j,j} L (2) It is a central assumption of our approach that the linear classifier output Vij is as informative as the input vector x. Hence, we approximate Prij = Pr(<Oj I <Ojj /\ (X=x? by Pr(<Oi I <Ojj /\ (V=Vjj?. Note that Pji = I-Pjj. In the second step, the probabilities Prij are combined to obtain posterior probabilities Pr(<Oj I (X=x? for all classes <Oi given a pattern x. Thus, the network can be considered as generating an intermediate data representation in the recognition chain, subject to further processing [Denker & LeCun, 1991]. In other words, the neural network becomes part of the preprocessing and contributes to dimensionality reduction. 3 Combining the Probabilities Prij of the Two-class Classifiers to a posteriori Probabilities The set of two-class neural network classifiers discussed in the previous section results in probabilities Prjj for all pairs (i, j) with i j. Here, the task is to express the posterior probabilities Pr(<Oj I (X=x? as functions of the Prjj- * 1112 David Price, Stefan Knerr, Leon Personnaz, Gerard Dreyfus We assume that each pattern belongs to only one class: K (3) Pr( U Olj I (X=x? = 1 j=1 From the definition of Olij. it follows for any given i: K K Pr(U Olj I (X=x? = Pr( U Olij I (X=x? = 1 j=l,j*i J=I (4) U sing the closed form expression for the probability of the union of N events Ei: N N N L L Pr(U Ei) = Pr(Ej) +... + (_I)k.1 Pr(EhA ... AEh) +...+ (-I)N.l pr(EIA ... AEN) i= 1 i= 1 i}< ... <ik it follows from (4): K L Pr(Olij I (X=x? - (K-2) Pr(Oli I (X=x? = 1 (5) j=l,j*i With Prij = Pr(Oli I OlijA(X=X? = Pr(OliAOli'A(X=X? J Pr(OlijA(X=X? Pr(Oli I (X=x? = --'----~ Pr(Olij I (X=x? (6) one obtains the final expression for the K posterior probabilities given the K(K-l)12 twoclass probabilities Prji : Pr(Oli I (X=x? = _ _--"-1_ __ f (7) _1__ (K-2) j=I,#i Prij In [Refregier et aI., 1991], a method was derived which allows to compute the K posterior probabilities from only (K-l) two-class probabilities using the following relation between posterior probabilities and two-class probabilities: Prij = Pr(Oli I (X=x? Prji Pr(Olj I (X=x? (8) However, this approach has several practical drawbacks. For instance, in practice, the quality of the estimation of the posterior probabilities depends critically on the choice of the set of (K-l) two-class probabilities, and finding the optimal subset of (K-l) Prij is costly, since it has to be performed for each pattern at recognition time. Pairwise Neural Network Classifiers with Probabilistic Outputs 4 1113 Application to Cursive Handwriting Recognition We applied the concepts described in the previous sections to the classification of presegmented characters from cursive words originating from real-world French postal checks. For cursive word recognition it is important to obtain probabilities at the output of the character classifier since it is necessary to establish an ordered list of hypotheses along with a confidence value for further processing at the word recognition level: the probabilities can be passed to an Edit Distance algorithm [Wagner et at, 1974] or to a Hidden-Markov-Model algorithm [Kundu et aI., 1989] in order to compute recognition scores for words. For the recognition of the amounts on French postal checks we used an Edit Distance algorithm and made extensive use of the fact that we are dealing with a limited vocabulary (28 words). The 27 character classes are particularly chosen for this task and include pairs of letters such as "fr", "gttl, and "tr" because these combinations of letters are often difficult to presegment. Other characters, such as tlk" and "y" are not included because they do not appear in the given 28 word vocabulary. 0~)(" C.A~ t:; ~"U::l._ r;.., ~ \\:..~~~ (~tSl\h~~ \J'~ ~~~ ~~~~~~ &.nl' tL'i.upr rm'A Figure 2: Some examples of literal amounts from live French postal checks. A data base of about 3,300 literal amounts from postal checks (approximately 16,000 words) was annotated and, based on this annotation, segmented into words and letters using heuristic methods [Simon et aI., 1994]. Figure 2 shows some examples of literal amounts. The writing styles vary strongly throughout the data base and many checks are difficult to read even for humans. Note that the images of the pre-segmented letters may still contain some of the ligatures or other extraneous parts and do not in general resemble hand-printed letters. The total of about 55,000 characters was divided into three sets: training set (20,000), validation set (20,000), and test set (15,000). All three sets were used without any further data base cleaning. Therefore, many letters are not only of very bad quality, but they are truly ambiguous: it is not possible to recognize them uniquely without word context. Figure 3: Reference lines indicating upper and lower limit of lower case letters. Before segmentation, two reference lines were detected for each check (Figure 3). They indicate an estimated upper and lower limit of the lower case letters and are used for 1114 David Price, Stefan Knerr, Leon Personnaz, Gerard Dreyfus nonnalization of the pre-segmented characters (Figure 4) to 10 by 24 pixel matrices with 16 gray values (Figure 5). This is the representation used as input to the classifiers. Figure 4: Segmentation of words into isolated letters (ligatures are removed later). Figure 5: Size nonnalized letters: 10 by 24 pixel matrices with 16 gray values. The simplest two-class classifier is a single neuron; thus, 351 neurons of the resulting pairwise classifier were trained on the training data using the generalized delta rule (sigmoidal transfer function). In order to avoid overfitting, training was stopped at the minimum of MSE on the validation set. The probability densities P(Vij I IDi) were estimated on the validation set: for both classes IDi and IDj of a given neuron, we fitted the probability densities over the linear output Vij to a Gaussian. The two-class probabilities Prij and Prji were then obtained via Bayes rule. The 351 probabilities Prij were combined using equation (7) in order to obtain a posteriori probabilities Pr(IDi I (X=x?, i E {1, .. ,27}. However, the a priori probabilities for letters as given by the training set are different from the prior probabilities in a given word context [Bourlard & Morgan, 1994]. Therefore, we computed the posterior probabilities either by using, in Bayes rule, the prior probabilities of the letters in the training set, or by assuming that the prior probabilities are equal. In the first case, many infonnative letters, for instance those having ascenders or descenders, have little chance to be recognized at all due to small a priori probabilities. Table 1 gives the recognition perfonnances on the test set for classes assumed to have equal a priori probabilities as well as for the true a priori probabilities of the test set. For each pattern, an ordered list (in descending order) of posterior class probabilities was generated; the recognition perfonnance is given (i) in tenns of percentage of true classes found in first position, and (ii) in tenns of average position of the true class in the ordered list. As mentioned above, the results of the first column are the most relevant ones, since the classifier outputs are subsequently used for word recognition. Note that the recognition rate (first position) of isolated letters without context for a human reader can be estimated to be around 70% to 80%. We compared the results of the pairwise classifier to a number of other neural network classification algorithms. First, we trained MLPs with one and two hidden layers and various numbers of hidden units using stochastic backpropagation. Here again, training was stopped based on the minimum MSE on the validation set. Second, we trained MLPs with a single hidden layer using the Softmax training algorithm [Bridle, 1990]. As a third approach, we trained 27 MLPs with 10 hidden units each, each MLP separating one class from all others. Table 1 gives the recognition perfonnances on the test set. The Softmax 1115 Pairwise Neural Network Classifiers with Probabilistic Outputs training algorithm clearly gives the best results in terms of recognition performance. However, the pairwise classifier has three very attractive features for classifier design: (i) training is faster than for MLP's by more than one order of magnitude; therefore, many different designs (changing pattern representations for instance) can be tested at a small computational cost; (ii) in the same spirit, adding a new class or modifying the training set of an existing one can be done without retraining all two-class classifiers; (iii) at least as importantly, the procedure gives more insight into the classification problem than MLP's do. Classifier Pairwise Classifier MLP (100 hid. units) Softmax (100 hid. units) 27 MLPs AveragePosition First Position equal prior probs equal prior probs 2.9 48.9 % AveragePosition true prior probs 2.6 First Position true prior probs 52.2 % 3.6 48.9 % 2.7 60.0 % 2.6 54.9 % 2.2 61.9 % 3.2 41.6 % 2.4 55.8 % Table I: Recognition performances on the test set in terms of average position and recognition rate (first position) for the various neural networks used. Our pairwise classifier is part of a handwriting recognition system which is currently applied to check reading. The complete system also incorporates other character recognition algorithms as well as a word recognizer which operates without pre-segmentation. The result of the complete check recognition chain on a set of test checks is the following: (i) at the word level, 83.3% of true words are found in first position; (ii) 64.1 % of well recognized literal amounts are found in first position [Simon et al., 1994]. Recognizing also the numeral amount, we obtained 80% well recognized checks for 1% error. 5 Conclusion We have shown how to obtain posterior class probabilities from a set of pairwise classifiers by (i) performing class density estimations on the network outputs and using Bayes rule, and (ii) combining the resulting two-class probabilities. The application of our pairwise classifier to the recognition of real world French postal checks shows that the procedure is a valuable tool for designing a recognizer, experimenting with various data representations at a small computational cost and, generally, getting insight into the classification problem. Acknowledgments The authors wish to thank J.C. Simon, N. Gorsky, O. Baret, and J.C. Deledicq for many informative and stimulating discussions. 1116 David Price, Stefan Knerr, Leon Personnaz, Gerard Dreyfus References H.A. Bourlard, N. Morgan (1994). Connectionist Speech Recognition. Kluwer Academic Publishers. H.A. Bourlard, C. Wellekens (1990). Links between Markov Models and Multilayer Perceptrons. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 12, No. 12, 1167-1178. J.S. Bridle (1990). Probabilistic Interpretation of Feedforward Classification Network Outputs, with Relationships to Statistical Pattern Recognition. In Neurocomputing: Algorithms, Architectures and Applications, Fogelman-Soulie, and Herault (eds.). NATO ASI Series, Springer. J.S. Denker, Y.LeCun (1991). Transforming Neural-Net Output Levels to Probability Distributions. In Advances in Neural Information Processing Systems 3, Lippmann, Moody, Touretzky (eds.). Morgan Kaufman. R.O. Duda, P.E. Hart (1973). Pattern Classification and Scene Analysis. Wiley. S. Knerr, L. Personnaz, G. Dreyfus (1990). Single-Layer Learning Revisited: A Stepwise Procedure for Building and Training a Neural Network. In Neurocomputing: Algorithms, Architectures and Applications, Fogelman-Soulie and Herault (eds.). NATO AS! Series, Springer. S. Knerr, L. Personnaz, G. Dreyfus (1992). Handwritten Digit Recognition by Neural Networks with Single-Layer Training. IEEE Transactions on Neural Networks, Vol. 3, No.6, 962-968. A. Kundu, Y. He, P. Bahl (1989). Recognition of Handwritten Words: First and Second Order Hidden Markov Model Based Approach. Pattern Recognition, Vol. 22, No.3. lC. Simon (1992). Off-Line Cursive Word Recognition. Proceedings of the IEEE, Vol. 80, No.7, 1150-1161. Ph. Refregier, F. Vallet (1991). Probabilistic Approach for Multiclass Classification with Neural Networks. Int. Conference on Artificial Networks, Vol. 2, 1003-1007. J.C. Simon, O. Baret, N. Gorski (1994). Reconnaisance d'ecriture manuscrite. Compte Rendu Academie des Sciences, Paris, t. 318, Serie II, 745-752. R.A. Wagner, M.J. Fisher (1974). The String to String Correction Problem. J.A.C.M. Vol. 21, No.5, 168-173. .
883 |@word duda:2 retraining:1 covariance:1 tr:1 serie:1 reduction:1 series:2 score:1 existing:1 partition:1 informative:2 interpretable:1 intelligence:1 postal:5 idi:3 revisited:1 sigmoidal:2 along:1 viable:1 ik:1 consists:1 combine:1 upr:2 pairwise:17 multi:2 little:1 becomes:1 kaufman:1 string:2 finding:1 transformation:1 classifier:33 k2:1 eha:1 partitioning:1 unit:6 rm:1 appear:1 before:1 limit:2 approximately:1 therein:1 limited:1 practical:2 lecun:2 acknowledgment:1 practice:2 union:2 backpropagation:2 digit:2 procedure:4 coping:1 asi:1 printed:1 word:20 pre:4 confidence:1 nonnalization:1 context:5 live:1 writing:1 descending:1 ke:1 knerr:10 m2:3 rule:5 insight:2 importantly:1 cleaning:1 programming:1 designing:1 hypothesis:2 academie:1 associate:1 recognition:32 particularly:2 solved:2 removed:1 valuable:1 mentioned:1 transforming:1 dynamic:1 trained:6 easily:1 represented:1 various:3 probs:4 detected:1 artificial:1 heuristic:1 widely:1 transform:1 final:1 net:1 fr:4 hid:2 relevant:1 combining:2 getting:1 prj:1 gerard:5 generating:1 ojj:6 resemble:1 tlk:1 indicate:1 drawback:1 annotated:1 modifying:1 subsequently:1 stochastic:1 human:2 numeral:1 require:1 suffices:1 probable:1 correction:1 around:1 considered:1 exp:1 lm:1 vary:1 purpose:1 recognizer:5 estimation:4 currently:3 edit:2 tool:2 weighted:1 compte:1 stefan:5 minimization:3 clearly:1 gaussian:3 avoid:1 ej:1 derived:1 check:13 experimenting:1 posteriori:3 cnrs:1 hidden:7 relation:1 originating:1 transformed:1 france:3 a2ia:2 pixel:2 fogelman:2 classification:11 priori:5 extraneous:1 herault:2 softmax:3 equal:5 having:1 presegmented:1 others:1 connectionist:1 descender:1 recognize:1 neurocomputing:2 pjj:1 intended:1 mlp:6 truly:1 nl:1 chain:2 necessary:1 perfonnance:1 divide:1 isolated:2 theoretical:1 stopped:2 fitted:1 instance:4 classify:1 soft:1 column:1 infonnative:1 cost:4 subset:1 recognizing:1 refregier:2 combined:2 density:6 probabilistic:13 off:1 moody:1 again:1 central:1 pr2:2 literal:4 style:1 de:2 int:1 depends:2 performed:4 view:1 later:1 closed:1 bayes:4 annotation:1 simon:6 mlps:5 oi:2 square:1 phoneme:1 likewise:1 efficiently:3 yield:1 handwritten:3 critically:1 touretzky:1 ed:3 definition:1 handwriting:5 bridle:3 dimensionality:1 segmentation:3 ok:2 higher:1 originally:1 eia:1 done:1 strongly:1 hand:1 ei:2 french:4 bahl:1 quality:3 gray:2 building:1 concept:1 contain:1 true:6 hence:1 read:1 attractive:1 uniquely:1 ambiguous:1 generalized:1 complete:2 image:1 dreyfus:8 common:1 oli:5 discussed:2 interpretation:2 he:1 kluwer:1 ai:5 pq:1 base:5 posterior:17 belongs:1 tenns:2 morgan:3 minimum:2 espci:4 recognized:3 ii:5 gorski:1 segmented:4 faster:1 academic:1 long:1 divided:1 hart:2 involving:3 multilayer:2 laboratoire:1 tsl:1 publisher:1 cedex:1 subject:1 incorporates:1 spirit:1 vallet:1 feedforward:3 intermediate:1 easy:1 iii:1 fit:1 architecture:3 twoclass:2 multiclass:1 expression:2 ultimate:1 passed:1 speech:2 neurones:2 generally:2 cursive:5 amount:6 ph:1 simplest:1 vauquelin:2 percentage:1 estimated:3 delta:1 vol:6 express:1 changing:1 sum:1 letter:14 powerful:1 throughout:1 reader:1 decision:3 layer:4 bp:1 scene:1 leon:5 performing:1 relatively:1 combination:1 character:11 pr:30 equation:1 wellekens:2 denker:2 pji:1 original:2 include:1 graded:1 establish:1 conquer:1 personnaz:7 strategy:1 costly:1 distance:2 link:1 thank:1 separating:1 reason:1 assuming:1 relationship:1 difficult:2 potentially:2 favorably:1 design:2 unknown:1 upper:2 neuron:5 markov:4 sing:1 nonnalized:1 david:5 pair:4 paris:4 namely:2 extensive:1 proceeds:1 pattern:17 reading:3 oj:19 event:1 bourlard:4 kundu:2 ascender:1 prior:7 limitation:1 validation:4 vij:9 elsewhere:1 course:1 idj:1 wagner:2 electronique:1 distributed:2 soulie:2 vocabulary:2 world:6 author:1 made:1 preprocessing:1 transaction:2 perfonnances:2 approximate:1 obtains:1 lippmann:1 nato:2 dealing:1 overfitting:1 olj:3 assumed:1 why:1 table:3 transfer:2 contributes:1 mse:4 necessarily:1 rue:3 main:1 ligature:2 tl:1 wiley:1 lc:1 sub:3 position:9 wish:1 third:1 extractor:1 bad:1 xt:1 list:3 vjj:10 stepwise:1 adding:1 magnitude:1 entropy:1 ordered:3 springer:2 chance:1 stimulating:1 conditional:2 goal:1 price:5 fisher:1 included:1 operates:1 total:1 perceptrons:2 indicating:1 tested:1
7,117
884
Spatial Representations in the Parietal Cortex May Use Basis Functions Alexandre Pouget alex@salk.edu Terrence J. Sejnowski terry@salk.edu Howard Hughes Medical Institute The Salk Institute La Jolla, CA 92037 and Department of Biology University of California, San Diego Abstract The parietal cortex is thought to represent the egocentric positions of objects in particular coordinate systems. We propose an alternative approach to spatial perception of objects in the parietal cortex from the perspective of sensorimotor transformations. The responses of single parietal neurons can be modeled as a gaussian function of retinal position multiplied by a sigmoid function of eye position, which form a set of basis functions. We show here how these basis functions can be used to generate receptive fields in either retinotopic or head-centered coordinates by simple linear transformations. This raises the possibility that the parietal cortex does not attempt to compute the positions of objects in a particular frame of reference but instead computes a general purpose representation of the retinal location and eye position from which any transformation can be synthesized by direct projection. This representation predicts that hemineglect, a neurological syndrome produced by parietal lesions, should not be confined to egocentric coordinates, but should be observed in multiple frames of reference in single patients, a prediction supported by several experiments. Alexandre Pouget, Terrence J. Sejnowski 158 1 Introduction The temporo-parietal junction in the human cortex and its equivalent in monkeys, the inferior parietal lobule, are thought to playa critical role in spatial perception. Lesions in these regions typically result in a neurological syndrome, called hemineglect, characterized by a lack of motor exploration toward the hemispace contralateral to the site of the lesion. As demonstrated by Zipser and Andersen [11), the responses of single cells in the monkey parietal cortex are also consistent with this presumed role in spatial perception. In the general case, recovering the egocentric position of an object from its multiple sensory inputs is difficult because of the multiple reference frames that must be integrated . In this paper, we consider a simpler situation in which there is only visual input and all body parts are fixed but the eyes, a condition which has been extensively used for neurophysiological studies in monkeys. In this situation, the head-centered position of an object, X, can be readily recovered from the retinal location,R, and current eye position, E, by vector addition: (1) If the parietal cortex contains a representation of the egocentric position of objects, then one would expect to find a representation of the vectors, X, associated with these objects. There is an extensive literature on how to encode a vector with a population of neurons, and we first present two schemes that have been or are used as working hypothesis to study the parietal cortex. The first scheme involves what is typically called a computational map, whereas the second uses a vectorial representation [9]. This paper shows that none of these encoding schemes accurately accounts for all the response properties of single cells in the parietal cortex. Instead, we propose an alternative hypothesis which does not aim at representing X per se; instead, the inputs Rand E are represented in a particular basis function representation. We show that this scheme is consistent with the way parietal neurons respond to the retinal position of objects and eye position, and we give computational arguments for why this might be an efficient strategy for the cortex. 2 Maps and Vectorial Representations One way to encode a two-dimensional vector is to use a lookup table for this vector which, in the case of a two-dimensional vector, would take the form of a twodimensional neuronal map. The parietal cortex may represent the egocentric location of object, X, in a similar fashion. This predicts that the visual receptive field of parietal neurons have a fixed position with respect to the head (figure IB). The work of Andersen et al. (1985) have clearly shown that this is not the case. As illustrated in figure 2A, parietal neurons have retinotopic receptive fields . In a vectorial representation, a vector is encoded by N units, each of them coding for the projection of the vector along its preferred direction. This entails that the activity, h, of a neuron is given by: Spatial Representations in the Parietal Cortex May Use Basis Functions B Map Representation A ~ V 159 C Vectorial Representation ?io j[ o 0 0 o vx ?1110 -'JO 0 90 180 a(I)'gr") Figure 1: Two neural representations of a vector . A) A vector if in cartesian and polar coordinates. B) In a map representation, units have a narrow gaussian tuning to the horizontal and vertical components of if . Moreover, the position of the peak response is directly related to the position of the units on the map. C) In a vectorial representation , each unit encodes the projection of if along its preferred direction ( central arrows) . This results in a cosine tuning to the vector angle, () . (2) Wa is usually called the preferred direction of the cells because the activity is maximum whenever () = 0; that is, when A points in the same direction as Wa. Such neurons have a cosine tuning to the direction of the egocentric location of objects, as shown also in figure lC. Cosine tuning curves have been reported in the motor cortex by Georgopoulos et al. (1982) , suggesting that the motor cortex uses a vectorial code for the direction of hand movement in extrapersonal space. The same scheme has been also used by Goodman and Andersen (1990), and Touretzski et al. (1993) to model the encoding of egocentric position of objects in the parietal cortex. Touretzski et al. (1993) called their representation a sinusoidal array instead of a vectorial representation. Using Eq. 1, we can rewrite Eq. 2: (3) This second equation is linear in Ii and if and uses the same vectors , dot products. This leads to three important predictions: Wa , in both 1) The visual receptive fields of parietal neurons should be planar. 2) The eye position receptive fields of parietal neurons should also be planar; that is, for a given retinal positions, the response of parietal neuron should be a linear function of eye position. 160 Alexandre Pouget, Terrence J. Sejnowski B A @ ? o -101'--'---'--~-'--'--'---'---' -40 -20 0 20 40 Retinal Position (Deg) Figure 2: Typical response of a neuron in the parietal cortex of a monkey. A) Visual receptive field has a fixed position on the retina, but the gain of the response is modulated by eye position (ex). (Adpated from Andersen et al., 1985) B) Example of an eye position receptive field, also called gain field, for a parietal cell. The nine circles indicate the amplitude of the response to an identical retinal stimulation for nine different eye positions. Outer circles show the total activity, whereas black circles correspond to the total response minus spontaneous activity prior to visual stimulation. (Adpated from Zipser et al., 1988) 3) The preferred direction for retinal location and eye position should be identical. For example, if the receptive field is on the right side of the visual field , the gain field should also increase with eye positon to the right side. The visual receptive fields and the eye position gain fields of single parietal neurons have been extensively studied by Andersen et al. [2]. In most cases, the visual receptive fields were bell-shaped with one or several peaks and an average radius of 22 degrees of visual angle [1], a result that is clearly not consistent with the first prediction above . We show in figure 2A an idealized visual receptive field of a parietal neuron. The effect of eye position on the visual receptive field is also illustrated. The eye position clearly modulates the gain of the visual response. The prediction regarding the receptive field for eye position has been borne out by statistical analysis. The gain fields of 80% of the cells had a planar component [1 , 11] . One such gain field is shown in figure 2B. There is not enough data available to determine whether or not the third prediction is valid. However, indirect evidence suggests that if such a correlation exists between preferred direction for retinal location and for eye position, it is probably not strong. Cells with opposite preferred directions [2, 3] have been observed. Furthermore, although each hemisphere represents all possible preferred eye position directions, there is a clear tendency to overrepresent the contralateral retinal hemifield [1]. In conclusion, the experimental data are not fully consistent with the predictions of the vectorial code. The visual receptive fields, in particular, are strongly nonlinear. If these nonlinearities are computationally neutral, that is, they are averaged out in subsequent stages of processing in the cortex, then the vectorial code could capture Spatial Representations in the Parietal Cortex May Use Basis Functions 161 the essence of what the parietal cortex computes and, as such, would provide a valid approximation of the neurophysiological data. We argue in the next section that the nonlinearities cannot be disregarded and we present a representational scheme in which they have a central computational function. 3 3.1 Basis Function Representation Sensorimotor Coordination and Nonlinear Function Approximation The function which specified the pattern of muscle activities required to move a limb, or the body, to a specific spatial location is a highly nonlinear function of the sensory inputs. The cortex is not believed to specify patterns of muscle activation, but the intermediate transformations which are handled by the cortex are often themselves nonlinear . Even if the transformations are actually linear, the nature of cortical representations often makes the problem a nonlinear mapping. For example, there exists in the putamen and premotor cortex cells with gaussian head-centered visual receptive fields [7J which means that these cells compute gaussians of A or, equivalently, gaussians of R + E, which is nonlinear in Rand E. There are many other examples of sensory remappings involving similar computations. If the parietal cortex is to have a role in these remappings, the cells should respond to the sensory inputs in a way that can be used to approximate the nonlinear responses observed elsewhere. One possibility would be for parietal neurons to represent input signals such as eye position and retinal location with basis functions. A basis function decomposition is a well-known method for approximating nonlinear functions which is, in addition, biologically plausible [8J. In such a representation, neurons do not encode the head-centered locations of objects, A; instead, they compute functions of the input variables, such as Rand E, which can be used subsequently to approximate any functions of these variables. 3.2 Predictions of the Basis Function Representation Not all functions are basis functions. Linear functions do not qualify, nor do sums of functions which, individually, would be basis functions, such as gaussian functions of retinal location plus a sigmoidal functions of eye position. If the parietal cortex uses a basis function representation, two conditions have to be met: 1) The visual and the eye position receptive fields should be smooth nonlinear function of Rand E. 2) The selectivities to Rand E should interact nonlinearly The visual receptive fields of parietal neurons are typically smooth and nonlinear. Gaussian or sum of gaussians appear to provide good models of their response profiles [2]. The eye position receptive field on the other hand, which is represented by the gain field, appears to be approximately linear. We believe, however, that the published data only demonstrate that the eye position receptive field is monotonic, 162 Alexandre Pouget, Terrence J. Sejnowski Head-Centered Retinotopic o o Figure 3: Approximation of a gaussian head-centered (top-left) and a retinotopic (top-right) receptive field, by a linear combination of basis function neurons. The bottom 3-D plots show the response to all possible horizontal retinal position, r x , and horizontal eye positions, ex, of four typical basis function units. These units are meant to model actual parietal neurons but not necessarily linear. In published experiments, eye position receptive fields (gain fields) were sampled at only nine points, which makes it difficult to distinguish between a plane and other functions such as a sigmoidal function or a piecewise linear function. The hallmark of a nonlinearity would be evidence for saturation of activity within working range of eye position. Several published gain fields show such saturations [3, 11], but a rigorous statistical analysis would be desirable. Andersen et al. (1985) have have shown that the responses of parietal neurons are best modeled by a multiplication between the retinal and eye position selectivities which is consistent with the requirements for basis functions. Therefore, the experimental data are consistent with our hypothesis that the parietal cortex uses a basis function representation. The response of most gain-modulated neurons in the parietal cortex could be modeled by multiplying a gaussian tuning to retinal position by a sigmoid of eye position, a function which qualifies as a basis function. 3.3 Simulations We simulated the response of 121 parietal gain-modulated neurons modeled by multiplying a gaussian of retinal position, r x , with a sigmoid of eye position, ex : Spatial Representations in the Parietal Cortex May Use Basis Functions e- (rz-rra):il ~ ~ .. h,o=----1 +e- t e,,;-e,l'j 163 (4) where the centers of the gaussians for retinalloction rxi and the positions of the sigmoids for eye postions exi were uniformly distributredo The widths of the gaussian (T and the sigmoid t were fixed. Four of these functions are shown at the bottom of figure 3. We used these basis functions as a hidden layer to approximate two kinds of output functions: a gaussian head-centered receptive field and a gaussian retinotopic receptive field . Neurons with these response properties are found downstream of the parietal cortex in the premotor cortex [7] and superior colliculus, two structures believed to be involved in the control of, respectively, arm and eye movements. The weights for a particular output were obtained by using the delta rule. Weights were adjusted until the mean error was below 5% of the maximum output value. Figure 3 shows our best approximations for both the head-centered and retinotopic receptive fields. This demonstrates that the same pool of neurons can be used to approximate several diffferent nonlinear functions. 4 Discussion Neurophysiological data support our hypothesis that the parietal cortex represents its inputs, such as the retinal location of objects and eye position, in a format suitable to non-linear function approximation, an operation central to sensorimotor coordination. Neurons have gaussian visual receptive fields modulated by monotonic function of eye position leading to response function that can be modeled by product of gaussian and sigmoids. Since the product of gaussian and sigmoids forms basis functions, this representation is good for approximating nonlinear functions of the input variables. Previous attempts to characterize spatial representations have emphasized linear encoding schemes in which the location of objects is represented in egocentric coordinates. These codes cannot be used for nonlinear function approximation and, as such, may not be adequate for sensorimotor coordination [6, 10]. On the other hand, such representations are computationally interesting for certain operations, like addition or rotation. Some part of the brain more specialized in navigation like the hippocampus might be using such a scheme [10]. In figure 3, a head-centered or a retinotopic receptive field can be computed from the same pool of neurons. It would be arbitrary to say that these neurons encode the positions of objects in egocentric coordinates. Instead, these units encode a position in several frames of reference simultaneously. If the parietal cortex uses this basis function representation, we predict that hemineglect, the neurological syndrome which results from lesions in the parietal cortex, should not be confined to any particular frame of reference. This is precisely the conclusion that has emerged from recent studies of parietal patients [4]. Whether the behavior of parietal patients can be fully explained by lesions of a basis function representation remains to be investigated. 164 Alexandre Pouget, Terrence J. Sejnowski Acknowledgments We thank Richard Andersen for helpful conversations and with access to unpublished data. References [1] R.A. Andersen, C. Asanuma, G. Essick, and R.M. Siegel. Corticocortical connections of anatomically and physiologically defined subdivisions within the inferior parietal lobule. Journal of Comparative Neurology, 296(1):65-113, 1990. [2] R.A . Andersen, G.K. Essick, and R.M. Siegel. Encoding of spatial location by posterior parietal neurons. Science, 230:456-458 , 1985. [3] R.A. Andersen and D. Zipser. The role of the posterior parietal cortex in coordinate transformations for visual-motor integration. Canadian Journal of Physiology and Pharmacology, 66:488-501, 1988. [4] M. Behrmann and M. Moscovitch. Object-centered neglect in patient with unilateral neglect: effect of left-right coordinates of objects. Journal of Cognitive Neuroscience , 6:1-16, 1994. [5] A.P. Georgopoulos, J.F. Kalaska, R. Caminiti, and J.T. Massey. On the relations between the direction of two-dimensional arm movements and cell discharge in primate motor cortex. Journal of Neuroscience, 2(11):1527-1537, 1982. [6] S.J. Goodman and R.A. Andersen. Algorithm programmed by a neural model for coordinate transformation. In International Joint Conference on Neural Networks, San Diego, 1990. [7] M.S. Graziano, G.s. Yap, and C.G. Gross. Coding of visual space by premotor neurons. Science, 266:1054-1057, 1994. [8] T. Poggio. A theory of how the brain might work. Cold Spring Harbor Symposium on Quantitative Biology, 55:899-910, 1990. [9] J .F. Soechting and M. Flanders. Moving in three-dimensional space: frames of reference, vectors and coordinate systems. Annual Review in Neuroscience, 15:167-91, 1992. [10] D.S. Touretzky, A.D . Redish, and H.S . Wan. Neural representation of space using sinusoidal arrays. Neural Computation, 5:869-884, 1993. [11] D. Zipser and R.A. Andersen. A back-propagation programmed network that stimulates reponse properties of a subset of posterior parietal neurons. Nature, 331:679- 684, 1988.
884 |@word hippocampus:1 simulation:1 decomposition:1 minus:1 contains:1 extrapersonal:1 recovered:1 current:1 activation:1 must:1 readily:1 subsequent:1 motor:5 plot:1 plane:1 location:13 sigmoidal:2 simpler:1 along:2 asanuma:1 direct:1 symposium:1 presumed:1 behavior:1 themselves:1 nor:1 brain:2 actual:1 retinotopic:7 moreover:1 what:2 kind:1 monkey:4 transformation:7 quantitative:1 demonstrates:1 control:1 unit:7 medical:1 appear:1 io:1 encoding:4 approximately:1 might:3 black:1 plus:1 studied:1 suggests:1 programmed:2 range:1 averaged:1 lobule:2 acknowledgment:1 hughes:1 cold:1 bell:1 thought:2 physiology:1 projection:3 cannot:2 caminiti:1 twodimensional:1 equivalent:1 map:6 demonstrated:1 center:1 pouget:5 rule:1 array:2 population:1 coordinate:10 discharge:1 spontaneous:1 diego:2 us:6 hypothesis:4 yap:1 corticocortical:1 predicts:2 observed:3 role:4 bottom:2 capture:1 region:1 remappings:2 movement:3 gross:1 raise:1 rewrite:1 basis:23 exi:1 indirect:1 joint:1 represented:3 sejnowski:5 encoded:1 premotor:3 plausible:1 emerged:1 say:1 propose:2 product:3 representational:1 requirement:1 comparative:1 object:17 eq:2 strong:1 recovering:1 involves:1 indicate:1 met:1 direction:11 radius:1 subsequently:1 centered:10 human:1 exploration:1 vx:1 adjusted:1 mapping:1 predict:1 purpose:1 polar:1 soechting:1 coordination:3 individually:1 clearly:3 gaussian:14 aim:1 putamen:1 encode:5 rigorous:1 helpful:1 typically:3 integrated:1 hidden:1 relation:1 spatial:10 integration:1 field:35 shaped:1 essick:2 biology:2 identical:2 represents:2 piecewise:1 richard:1 retina:1 simultaneously:1 attempt:2 possibility:2 highly:1 navigation:1 poggio:1 circle:3 rxi:1 neutral:1 subset:1 contralateral:2 gr:1 characterize:1 reported:1 stimulates:1 peak:2 international:1 terrence:5 pool:2 graziano:1 hemifield:1 jo:1 andersen:12 central:3 wan:1 borne:1 qualifies:1 cognitive:1 leading:1 suggesting:1 account:1 sinusoidal:2 nonlinearities:2 lookup:1 retinal:17 coding:2 redish:1 idealized:1 il:1 correspond:1 accurately:1 produced:1 none:1 multiplying:2 published:3 touretzky:1 whenever:1 sensorimotor:4 involved:1 associated:1 gain:12 sampled:1 conversation:1 amplitude:1 actually:1 back:1 appears:1 alexandre:5 planar:3 response:18 specify:1 rand:5 reponse:1 strongly:1 furthermore:1 stage:1 correlation:1 until:1 working:2 hand:3 horizontal:3 nonlinear:13 lack:1 propagation:1 believe:1 effect:2 illustrated:2 width:1 inferior:2 essence:1 cosine:3 demonstrate:1 hallmark:1 sigmoid:4 superior:1 rotation:1 specialized:1 stimulation:2 synthesized:1 tuning:5 nonlinearity:1 had:1 dot:1 moving:1 access:1 entail:1 cortex:34 playa:1 rra:1 posterior:3 recent:1 perspective:1 jolla:1 hemisphere:1 selectivity:2 certain:1 qualify:1 muscle:2 syndrome:3 determine:1 signal:1 ii:1 multiple:3 desirable:1 smooth:2 characterized:1 believed:2 kalaska:1 prediction:7 involving:1 patient:4 represent:3 confined:2 cell:10 addition:3 whereas:2 goodman:2 probably:1 zipser:4 intermediate:1 canadian:1 enough:1 harbor:1 opposite:1 regarding:1 whether:2 handled:1 unilateral:1 nine:3 adequate:1 se:1 clear:1 extensively:2 generate:1 delta:1 neuroscience:3 per:1 four:2 massey:1 egocentric:9 downstream:1 sum:2 colliculus:1 angle:2 respond:2 layer:1 distinguish:1 annual:1 activity:6 hemispace:1 vectorial:9 precisely:1 alex:1 georgopoulos:2 encodes:1 argument:1 spring:1 format:1 department:1 combination:1 biologically:1 primate:1 explained:1 anatomically:1 computationally:2 equation:1 remains:1 junction:1 available:1 gaussians:4 operation:2 multiplied:1 limb:1 alternative:2 rz:1 top:2 neglect:2 approximating:2 move:1 receptive:26 strategy:1 thank:1 simulated:1 outer:1 argue:1 toward:1 code:4 modeled:5 equivalently:1 difficult:2 hemineglect:3 vertical:1 neuron:29 howard:1 parietal:47 situation:2 head:10 frame:6 arbitrary:1 nonlinearly:1 required:1 specified:1 extensive:1 unpublished:1 connection:1 california:1 narrow:1 usually:1 perception:3 pattern:2 below:1 saturation:2 terry:1 critical:1 suitable:1 arm:2 representing:1 scheme:8 eye:33 prior:1 literature:1 review:1 multiplication:1 fully:2 expect:1 interesting:1 degree:1 consistent:6 elsewhere:1 supported:1 side:2 institute:2 curve:1 cortical:1 valid:2 computes:2 sensory:4 san:2 approximate:4 preferred:7 deg:1 neurology:1 physiologically:1 why:1 table:1 nature:2 ca:1 interact:1 investigated:1 necessarily:1 arrow:1 profile:1 pharmacology:1 lesion:5 body:2 neuronal:1 site:1 siegel:2 fashion:1 salk:3 lc:1 position:49 ib:1 flanders:1 third:1 behrmann:1 specific:1 emphasized:1 evidence:2 exists:2 modulates:1 sigmoids:3 cartesian:1 disregarded:1 neurophysiological:3 visual:19 neurological:3 monotonic:2 typical:2 uniformly:1 total:2 called:5 tendency:1 la:1 experimental:2 subdivision:1 support:1 moscovitch:1 modulated:4 meant:1 ex:3
7,118
885
A Comparison of Discrete-Time Operator Models for Nonlinear System Identification Andrew D. Back, Ah Chung Tsoi Department of Electrical and Computer Engineering, University of Queensland St. Lucia, Qld 4072. Australia. e-mail: {back.act}@elec.uq.oz.au Abstract We present a unifying view of discrete-time operator models used in the context of finite word length linear signal processing. Comparisons are made between the recently presented gamma operator model, and the delta and rho operator models for performing nonlinear system identification and prediction using neural networks. A new model based on an adaptive bilinear transformation which generalizes all of the above models is presented. 1 INTRODUCTION The shift operator, defined as qx(t) ~ x(t + 1), is frequently used to provide time-domain signals to neural network models. Using the shift operator, a discrete-time model for system identification or time series prediction problems may be constructed. A common method of developing nonlinear system identification models is to use a neural network architecture as an estimator F(Y(t), X(t); 0) of F(Y(t), X(t?, where 0 represents the parameter vector of the network. Shift operators at the input of the network provide the regression vectors Y(t-l) [yet-I), ... , y(t-N)]', andX(t) [x(t), ... , x(t-M)]' in a manner analogous to linear filters, where [.], represents the vector transpose. = = It is known that linear models based on the shift operator q suffer problems when used to modellightly-damped-Iow-frequency (LDLF) systems, with poles near (1,0) on the unit circle in the complex plane [5]. As the sampling rate increases, coefficient sensitivity and round-off noise become a problem as the difference between successive sampled inputs becomes smaller and smaller. 884 Andrew D. Back, Ah Chung Tsoi A method of overcoming this problem is to use an alternative discrete-time operator. Agarwal and Burrus first proposed the use of the delta operator in digital filters to replace the shift operator in an attempt to overcome the problems described above [1]. The delta operator is defined as q-l {) = (1) ~ where ~ is the discrete-time sampling interval. Williamson showed that the delta operator allows better performance in terms of coefficient sensitivity for digital filters derived from the direct form structure [19], and a number of authors have considered using it in linear filtering, estimation and control [5, 7, 8] More recently, de Vries, Principe at. al. proposed the gamma operator [2, 3] as a means of studying neural network models for processing time-varying patterns. This operator is defined by = 'Y q - (1 - c) (2) C It may be observed that it is a generalization of the delta operator with adjustable parameters c. An extension to the basic gamma operator introducing complex poles using a second order operator, was given in [18]. This raises the question, is the gamma operator capable of providing better neural network modelling capabilities for LDLF systems ? Further, are there any other operators which may be better than these for nonlinear modelling and prediction using neural networks? In the context of robust adaptive control, Palaniswami has introduced the rho operator which has shown useful improvements over the performance ofthe delta operator [9, 10]. The rho operator is defined as p = (3) where Cl, C2 are adjustable parameters. The rho operator generalizes the delta and gamma operators. For the case where Cl~ C2~ 1, the rho operator reduces to the usual shift 0, and C2 1, the rho operator reduces to the delta operator [10]. operator. When c) For Cl ~ = C2~ = c, the rho operator is equivalent to the gamma operator. = = = = One advantage of the rho operator over the delta operator is that it is stably invertible, allowing the derivation of simpler algorithms [9]. The p operator can be considered as a stable low pass filter, and parameter estimation using the p operator is low frequency biased. For adaptive control systems, this gives robustness advantages for systems with unmodelled high frequency characteristics [9] . By defining the bilinear transformation (BLT) as an operator, it is possible to introduce an operator which generalizes all of the above operators. We can therefore define the pi operator as 11" = 2 (Clq - ~ (C3Q C2) + C4) (4) with the restriction that Cl C4 f C2C3 (to ensure 11" is not a constant function [14]). The bilinear mapping produced has a pole at q = -C4/C3. By appropriate setting of the Cl, C2, C3, C4 parameters each operator, the pi operator can be reduced to each of the previous operators. In the work reported here, we consider these alternative discrete-time operators in feedforward neural network models for system identification tasks. We compare the popular A Comparison of Discrete-Time Operator Models for Nonlinear System Identification 885 gamma model [4] with other models based on the shift, delta, rho and pi operators. A framework of models and Gauss-Newton training algorithms is provided, and the models are compared by simulation experiments. 2 OPERATOR MODELS FOR NONLINEAR SIGNAL PROCESSING A model which generalizes the usual discrete-time linear moving average model, ie, a single layer network is given by yet) G(v,O)x(t) (5) M C(v,O) L: bw- I i i=O q-~ 8-' ,-i p- i 7r- i shift operator delta operator gamma operator rho operator pi operator (6) This general class of moving average model can be termed MA(v). We define uo(t) ~ x(t), and Ui(t) ~ V-IUi_l (t) I and hence obtain x(t - i) - 1) + Ui(t - 1) CUi-l(t - 1) + (1 - C)Ui(t - 1) C2~Ui-l(t - 1) + (1- Cl~)Ui(t - 1) 2~1 (C3 Ui-l(t) + C4 Ui-l(t - 1?) - ~Ui(t - 1) ~Ui-l(t shift operator delta operator gamma operator rho operator pi operator (7) A nonlinear model may be defined using a multilayer perceptron (MLP) with the v-operator elements at the input stage. The input vector ZP( t) to the network is Z?(t) = [Xi(t), V-1Xi(t), ... , V-MXi(t)]' (8) where Xi(t) is the ith input to the system. This model is termed the v-operator multilayer perceptron or MLP(v) model. An MLP(v) model having L layers with No, N I , ... , NL nodes per layer, is defined in the same manner as a usual MLP, with (9) Nz L WiiZJ-I(t) (10) i=l where each neuron i in layer 1 has an output of z!(t); a layer consists of N/ neurons (1 = 0 denotes the input layer, and 1 = L denotes the output layer, zJvz 1.0 may be used for a bias); 10 is a sigmoid function typically evaluated as tanh(?), and a synaptic connection between unit i in the previous layer and unit k in the current layer is represented by The notation t may be used to represent a discrete time or pattern instance. While the case = wt. 886 Andrew D. Back, Ah Chung Tsoi we consider employs the v-operator at the input layer only, it would be feasible to use the operators throughout the network as required. On-line algorithms to update the operator parameters in the MA(v) model can be found readily. In the case of the MLP(v) model, we approach the problem by backpropagating the error information to the input layer and using this to update the operator coefficients. de Vries and Principe et. al., proposed stochastic gradient descent type algorithms for adjusting the c operator coefficient using a least-squares error criterion [2, 12]. For brevity we omit the updating procedures for the MLP network weights; a variety of methods may be applied (see for example [13, 15]). We define an instantaneous output error criterion J(t) = !e 2(t), where e(t) = y(t) - f)(t). Defining 0 as the estimated operator parameter vector at time t of the parameter vector 0, we have c gamma operator { [CI , C2]' rho operator (11) [CI, C2, C3, 134 ]' pi operator o= A first order algorithm to update the coefficients is Oi(t + 1) ~Oi (t) Oi(t) + ~Oi(t) -71\1 eJ ((}; t) = (12) (13) where the adjustment in weights is found as ~Oi(t) = -71 oJ(t) o(}j M 71 I: 1/J1'(t)Oj(t) (14) i=1 where OJ (t) is the backpropagated error at the jth node of input layer, and 1/J1' (t) is the first order sensitivity vector of the model operator parameters, defined by &Ui(t) gamma operator ./~ (t) !Pi = I [::~(t) &Cjl &Cj2 [ &Ui(t) &Ui(t) &Cjl rho operator &Ui(t)]' ' ' &Cj2 &Ui(t) ' &Cj3 &Ui(t)] ' &Cj4 I (15) Pi operator Substituting Ui(t) in from (7), the recursive equations for 1/J1 (t) (noting that 1/J1 (t)= tPHt) 'Vj) are tPi(t) = Ui_l(t - 1) - Ui(t - 1) + CitPi-l(t - 1) + (1 - C)tPi(t - 1) gamma operator tPi(t) = - 1) + (1- CI~)tf;i I(t - 1) - ~Ui(t - 1) ] [ C2~tPi-I,I(t ~Ui-I(t - 1) + C2~tPi-I,2(t - 1) + (1 - CI~)tPi,2(t _ 1) rho operator 2t (C3tPi-I,I(t) + C4tPi-l,1(t - 1?) + ~tPi,l(t - 1) + C4Ui_l(t - 1?) - ~Ui(t - 1), ~C3tPi-I,2(t) + C4tPi-I,2(t - 1?) + ~tPi,2(t - 1) + tUi(t - 1), (Ui_l(t) + C3tPi-l,3(t) + C4tPi-I,3(t - 1?) + ~tPi~3(t - 1), (C3tPi-I,4(t) + Ui-I(t - 1) + C4tPi-l,4(t - 1? + T.-tPi,4(t - 1) -2~2 (C3Ui-l(t) c} 2~1 2t 2t C1 pi operator A Comparison of Discrete-Time Operator Models for Nonlinear System Identification 887 for the gamma, rho, and pi operators respectively, and where ""i ,j (t) refers to the jth element of the ith "" vector, with ""i ,o(t) = O. A more powerful updating procedure can be obtained by using the Gauss-Newton method [6]. In this case, we replace (14) with (omitting i subscripts for clarity), OCt + 1) = OCt) + 'Y(t)R- 1(t)",,(t)A -16(t) (16) where 'Y(t) is the gain sequence (see [6] for details), A-I is a weighting matrix which may be replaced by the identity matrix [16], or estimated as [6] A(t) = A(t - 1) + 'Y(t) (6 2(t) - A(t - 1?) (17) R( t) is an approximate Hessian matrix, defined by R(t + 1) = A(t)R(t) + ?(t)""(t),,,,'(t) (18) where A(t) = 1 - ?(t). Efficient computation of R- 1 may be performed using the matrix inversion lemma [17], factorization methods such as Cholesky decomposition or other fast algorithms. Using the well known matrix inversion lemma [6], we substitute pet) = R-l(t), where Pt () - _1_p t _ ?(t) ( P(t)",,(t)""'(t)P(t) ) A(t) () A(t) A(t) + ?(t)""'(t)P(t)",,(t) (19) The initial values of the coefficients are important in determining convergence. Principe et. al. [12] note that setting the coefficients for the gamma operator to unity provided the best approach for certain problems. 3 SIMULATION EXAMPLES We are primarily interested in the differences between the operators themselves for modelling and prediction, and not the associated difficulties of training multilayer perceptrons (recall that our models will only differ at the input layer). For the purposes of a more direct comparison, in this paper we test the models using a single layer network. Hence these linear system examples are used to provide an indication of the operators' performance. 3.1 EXPERIMENT 1 The first problem considered is a system identification task arising in the context of high bit rate echo cancellation [5]. In this case, the system is described by H(z) = 0.0254- 0.0296z- 1+ 0.00425z- 2 1 _ 1.957z-1 + 0.957z- 2 (20) This system has poles on the real axis at 0.9994, and 0.9577, thus it is an LDLF system. The input signal to the system in each case consisted of uniform white noise with unit variance. A Gauss-Newton algorithm was used to determine all unknown weights. We conducted Monte-Carlo tests using 20 runs of differently seeded training samples each of 2000 points to obtain the results reported. We assessed the performance of the models by using the Signal-to-NoiseRatio (SNR) defined as 1010g( E[d(t)2Jj E[e(t)2D, where E[?l is the expectation operator, and d(t) is the desired signal. For each run, we used the last 500 samples to compute a SNR figure. 888 Andrew D. Back, Ah Chung Tsoi 0.04 0.03 ? 0.02 0.01 0.04 0.03 0.02 ? o .~.. -0.01 -0.02 -0.03 -O''?900 ~. .~ ? o.ot o.ot 0 -0.01 -0.02 -0.03 0 -0.01 -0.02 -0.03 -0?11900 1920 (a) 0.04 0.03 0.02 0.01 0 -{l.01 -0.02 -0.03 -0?11900 0.04 0.03 0.02 ..? -0?11900 1920 1920 (b) (c) 0.04 0.03 0.02 0.01 0 -0.01 -0.02 -0.03 -O''?900 1920 (d) 1920 (e) Figure 1: Comparison of typical model output results for Experiment 1 with models based on the following operators: (a) shift, (b) delta (c) gamma, (d) rho, and (e) pi. Table 1: System Identification Experiment 1 Results For the purposes of this experiment, we conducted several trials and selected 0(0) values which provided stable convergence. The values chosen for this experiment were: 0(0) {0.75, [0.5,0.75]' [0.75,0.7,0.35,-0.25]} for the gamma, rho and pi operator models respectively. In each case we used model order M = 8. = Results for this experiment are shown in Table 1 and Figure 1. We observe that the pi operator gives the best performance overall. Some difficulties with instability occurring were encountered, thereby requiring a stability correction mechanism to be used on the operator updates. The next best performance was observed in the rho and then gamma models, with fewer instability problems occurring. 3.2 EXPERIMENT 2 The second experiment used a model described by H(z) = 1 - 0.8731z- 1 - 0.8731z-2 + z-3 1 - 2.8653z- 1 + 2.7505z- 2 - 0.8843z- 3 (21) This system is a 3rd order lowpass filter tested in [11]. The same experimental procedures as used in Experiment 1 were followed in this case. For the second experiment (see Table 2), it was found that the pi operator gave the best results A Comparison of Discrete-Time Operator Models for Nonlinear System Identification 889 Table 2: System Identification Experiment 2 Results recorded over all the tests. On average however, the improvement for this identification problem is less. It is observed that that the pi model is only slightly better than the gamma and rho models. Interestingly, the gamma and rho models had no problems with stability, while the pi model still suffered from convergence problems due to instability. As before, the delta model gave a wide variation in results and performed poorly. From these and other experiments performed it appears that performance advantages can be obtained through the use of the more complex operators. As observed from the best recorded runs, the extra degrees of freedom in the rho and pi operators appear to provide the means to give better performance than the gamma model. The improvements of the more complex operators come at the expense of potential convergence problems due to instabilities occurring in the operators and a potentially multimodal mean square output error surface in the operator parameter space. Clearly, there is a need for further investigation into the performance of these models on a wider range of tasks. We present these preliminary examples as an indication of how these alternative operators perform on some system identification problems. 4 CONCLUSIONS Models based on the delta operator, rho operator, and pi operator have been presented and new algorithms derived. Comparisons have been made to the previously presented gamma model introduced by de Vries, Principe et. al. [4] for nonlinear signal processing applications. While the simulation examples considered show are only linear, it is important to realize that the derivations are applicable for multilayer perceptrons, and that the input stage of these networks is identical to what we have considered here. We treat only the linear case in the examples in order not to complicate our understanding of the results, knowing that what happens in the input layer is important to higher layers in network structures. The results obtained indicate that the more complex operators provide a potentially more powerful modelling structure, though there is a need for further work into mechanisms of maintaining stability while retaining good convergence properties. The rho model was able to perform better than the gamma model on the problems tested, and gave similar results in terms of susceptibility to convergence and instability problems. The pi model appears capable of giving the best performance overall, but requires more attention to ensure the stability of the coefficients. For future work it would be of value to analyse the convergence of the algorithms, in order to design methods which ensure stability can be maintained, while not disrupting the convergence of the model. 890 Andrew D. Back, Ah Chung Tsoi Acknowledgements The first author acknowledges financial support from the Australian Research Council. The second author acknowledges partial support from the Australian Research Council. References [1] R.C. Agarwal and C.S. Burrus, ''New recursive digital filter structures having very low sensitivity and roundoff noise", IEEE Trans. Circuits and Systems, vol. cas-22, pp. 921-927, Dec. 1975. [2] de Vries, B. Principe, J.C. "A theory for neural networks with time delays", Advances in Neural Information Processing Systems, 3, R.P. Lippmann (Ed.), pp 162 - 168, 1991. [3] de Vries, B., Principe, J. and P.G. de Oliveira "Adaline with adaptive recursive memory", Neural Networks for Signal Processing I. Juang, B.H., Kung, S.Y., Kamm, C.A. (Eds) IEEE Press, pp. 101-110, 1991. [4] de Vries, B. Principe, J. "The Gamma Model - a new neural model for temporal processing". Neural Networks. Vol 5, No 4, pp 565 - 576, 1992. [5] H. Fan and Q. Li, "A () operator recursive gradient algorithm for adaptive signal processing", Proc. IEEE Int. Conf. Acoust. Speech and Signal Proc., vol. nI, pp. 492-495, 1993. [6] L. Ljung, and T. SOderstrtlm, Theory and Practice of Recursive Identification, Cambridge, Massachusetts: The MIT Press, 1983. [7] R.H. Middleton, and G.C. Goodwin, Digital Control and Estimation, Englewood Cliffs: Prentice Hall, 1990. [8] V. Peterka, "Control of Uncertain Processes: Applied Theory and Algorithms", Kybernetika, vol. 22, pp. 1-102, 1986. [9] M. Palaniswami, "A new discrete-time operator for digital estimation and control". The Uni versity of Melbourne, Department of Electrical Engineering, Technical Report No.1, 1989. [10] M. Palaniswami, "Digital Estimation and Control with a New Discrete Time Operator", Proc. 30th IEEE Conf. Decision and Control, pp. 1631-1632, 1991. [11] J.C. Principe, B. de Vries, J-M. Kuo and P. Guedes de Oliveira, "Modeling Applications with the Focused Gamma Net", Advances in Neural Information Processing Systems, vol. 4, pp. 143-150,1991. [12] J.C. Principe, B. de Vries, and P. Guedes de Oliveira, "The Gamma Filter - a new class of adaptive IIR filters with restricted feedback", IEEE Trans. Signal Processing, vol. 41, pp. 649-656, 1993. [13] G. V. Puskorius, and L.A. Feldkamp, "Decoupled Extended Kalman Filter Training of Feedforward Layered Networks", Proc. Int Joint Conf. Neural Networks, Seattle, vol I, pp. 771-777,1991. [14] E.B. Saff and A.D. Snider, Fundamentals of Complex Analysis for Mathematics, Science and Engineering. Englewood Cliffs, NJ: Prentice-Hall, 1976. [15] S. Shah and F. Palmieri, "MEKA - A Fast Local Algorithm for Training Feedfoward Neural Networks", Proc Int Joint Conf. on Neural Networks, vol m, pp. 41-46, 1990. [16] J.J. Shynk, "Adaptive IIR filtering using parallel-form realizations", IEEE Trans. Acoust. Speech Signal Proc., vol. 37, pp. 519-533,1989. [17] Soderstrom and Stoica, "System Identification", London: Prentice Hall, 1989. [18] T.O. de Silva, P.G. de Oliveira, J.e. Principe, and B. de Vries, " Generalized feedforward filters with complex poles", Neural Networks for Signal Processing n, S.Y. Kung et. al. (Eds) Piscataway,NJ: IEEE Press, 1992. [19] D. Williamson, "Delay replacement in direct form structures", IEEE Trans. Acoust., Speech, Signal Processing, vol. ASSP-34, pp. 453-460, Aprl. 1988. PARTvm VISUAL PROCESSING
885 |@word trial:1 inversion:2 simulation:3 queensland:1 decomposition:1 thereby:1 initial:1 series:1 interestingly:1 current:1 yet:2 readily:1 realize:1 j1:4 update:4 selected:1 fewer:1 plane:1 ith:2 feedfoward:1 node:2 successive:1 simpler:1 constructed:1 direct:3 become:1 c2:11 consists:1 introduce:1 manner:2 themselves:1 frequently:1 feldkamp:1 kamm:1 versity:1 becomes:1 provided:3 notation:1 circuit:1 what:2 kybernetika:1 acoust:3 transformation:2 nj:2 temporal:1 act:1 control:8 unit:4 uo:1 omit:1 appear:1 before:1 engineering:3 local:1 treat:1 bilinear:3 cliff:2 subscript:1 au:1 nz:1 factorization:1 roundoff:1 range:1 tsoi:5 recursive:5 practice:1 procedure:3 word:1 refers:1 layered:1 operator:99 prentice:3 context:3 instability:5 restriction:1 equivalent:1 middleton:1 attention:1 focused:1 estimator:1 financial:1 stability:5 variation:1 analogous:1 pt:1 element:2 updating:2 observed:4 electrical:2 ui:21 raise:1 multimodal:1 lowpass:1 joint:2 differently:1 represented:1 derivation:2 elec:1 fast:2 london:1 monte:1 rho:23 analyse:1 echo:1 advantage:3 sequence:1 indication:2 net:1 tpi:10 realization:1 adaline:1 poorly:1 oz:1 seattle:1 convergence:8 juang:1 zp:1 wider:1 andrew:5 come:1 indicate:1 australian:2 differ:1 filter:10 stochastic:1 australia:1 generalization:1 investigation:1 preliminary:1 extension:1 correction:1 considered:5 hall:3 mapping:1 substituting:1 susceptibility:1 purpose:2 estimation:5 proc:6 applicable:1 tanh:1 council:2 tf:1 mit:1 clearly:1 ej:1 varying:1 derived:2 improvement:3 modelling:4 typically:1 interested:1 overall:2 retaining:1 having:2 sampling:2 identical:1 represents:2 future:1 report:1 employ:1 primarily:1 gamma:25 replaced:1 bw:1 replacement:1 attempt:1 freedom:1 mlp:6 englewood:2 nl:1 damped:1 puskorius:1 capable:2 partial:1 decoupled:1 soderstrom:1 circle:1 desired:1 uncertain:1 melbourne:1 instance:1 modeling:1 pole:5 introducing:1 snr:2 uniform:1 delay:2 conducted:2 reported:2 iir:2 st:1 fundamental:1 sensitivity:4 ie:1 off:1 invertible:1 recorded:2 cjl:2 conf:4 chung:5 li:1 potential:1 de:14 coefficient:8 int:3 performed:3 view:1 stoica:1 capability:1 parallel:1 palaniswami:3 square:2 oi:5 ni:1 variance:1 characteristic:1 ofthe:1 identification:15 produced:1 carlo:1 ah:5 synaptic:1 complicate:1 ed:3 frequency:3 pp:13 associated:1 sampled:1 gain:1 adjusting:1 popular:1 massachusetts:1 palmieri:1 recall:1 back:6 appears:2 higher:1 evaluated:1 though:1 stage:2 nonlinear:10 stably:1 omitting:1 consisted:1 requiring:1 hence:2 seeded:1 white:1 cj4:1 round:1 backpropagating:1 maintained:1 criterion:2 generalized:1 disrupting:1 silva:1 instantaneous:1 recently:2 common:1 sigmoid:1 blt:1 cambridge:1 meka:1 rd:1 mathematics:1 cancellation:1 had:1 moving:2 stable:2 surface:1 showed:1 termed:2 certain:1 peterka:1 determine:1 signal:14 reduces:2 technical:1 prediction:4 regression:1 basic:1 multilayer:4 expectation:1 guedes:2 represent:1 agarwal:2 dec:1 c1:1 interval:1 suffered:1 biased:1 ot:2 extra:1 near:1 noting:1 feedforward:3 variety:1 gave:3 architecture:1 knowing:1 shift:10 suffer:1 speech:3 hessian:1 jj:1 useful:1 oliveira:4 backpropagated:1 reduced:1 cj2:2 delta:15 estimated:2 per:1 arising:1 discrete:13 vol:10 clarity:1 snider:1 run:3 powerful:2 throughout:1 decision:1 bit:1 layer:16 followed:1 ldlf:3 fan:1 encountered:1 lucia:1 performing:1 department:2 developing:1 piscataway:1 cui:1 smaller:2 slightly:1 unity:1 happens:1 restricted:1 equation:1 previously:1 mechanism:2 studying:1 generalizes:4 observe:1 appropriate:1 uq:1 alternative:3 robustness:1 shah:1 substitute:1 denotes:2 ensure:3 maintaining:1 newton:3 unifying:1 giving:1 question:1 usual:3 gradient:2 mail:1 pet:1 length:1 kalman:1 providing:1 potentially:2 expense:1 design:1 adjustable:2 unknown:1 allowing:1 perform:2 neuron:2 finite:1 descent:1 defining:2 extended:1 assp:1 overcoming:1 introduced:2 tui:1 goodwin:1 required:1 c3:4 connection:1 c4:5 trans:4 able:1 pattern:2 oj:3 memory:1 difficulty:2 mxi:1 axis:1 shynk:1 acknowledges:2 saff:1 understanding:1 acknowledgement:1 determining:1 ljung:1 filtering:2 digital:6 degree:1 pi:19 burrus:2 last:1 transpose:1 jth:2 bias:1 perceptron:2 wide:1 overcome:1 feedback:1 author:3 made:2 adaptive:7 qx:1 approximate:1 lippmann:1 uni:1 qld:1 xi:3 table:4 robust:1 ca:1 williamson:2 complex:7 cl:6 domain:1 vj:1 noise:3 weighting:1 unmodelled:1 ci:4 vries:9 occurring:3 visual:1 adjustment:1 ma:2 oct:2 identity:1 replace:2 feasible:1 typical:1 wt:1 lemma:2 pas:1 kuo:1 gauss:3 experimental:1 perceptrons:2 principe:10 cholesky:1 support:2 assessed:1 brevity:1 kung:2 tested:2
7,119
886
Implementation of Neural Hardware with the Neural VLSI of URAN in Applications with Reduced Representations ll-Song Han Korea Telecom Research Laboratories 17, Woomyun-dong, Suhcho-ku Seoul 137-140, KOREA Ki-Chul Kim Dept. of Info and Comm KAIST Seoul, 130-012, Korea Hwang-Soo Lee Dept. of Info and Comm KAIST Seoul, 130-012, Korea Abstract This paper describes a way of neural hardware implementation with the analog-digital mixed mode neural chip. The full custom neural VLSI of Universally Reconstructible Artificial Neural network (URAN) is used to implement Korean speech recognition system. A multi-layer perceptron with linear neurons is trained successfully under the limited accuracy in computations. The network with a large frame input layer is tested to recognize spoken korean words at a forward retrieval. Multichip hardware module is suggested with eight chips or more for the extended performance and capacity. 812 ll-Song Han, Hwang-Soo Lee, Ki-Chul Kim 1 INTRODUCTION In general, the neural network hardware or VLSI has been preferred in respects of its relatively fast speed, huge network size and effective cost comparing to software simulation. Universally Reconstructible Artificial Neural-network(URAN), the new analog-digital mixed VLSI neural network, can be used for the implementation of the real world neural network applications with digital interface. The basic electronic synapse circuit is based on the electrically controlled MOSFET resistance and is operated with discrete pulses. The URAN's adaptability is tested for the multi-layer perceptron with the reduced precision of connections and states. The linear neuron function is also designed for the real world applications. The multi-layer network with back propagation learning is designed for the speaker independent digit/word recognition. The other case of application is for the servo control, where the neural input and output are extended to 360 levels for the suitable angle control. With the servo control simulation, the flexibility of URAN is proved to extend the accuracy of input and output from external. 2. Analog-Digital Mixed Chip - URAN In the past, there have been improvements in analog or analog-digital mixed VLSI chips. Analog neural chips or analog-digital mixed neural chips are still suffered from the lack of accuracy, speed or flexibility. With the proposed analog-digital mixed neural network circuit of URAN, the accuracy is improved by using the voltage-controlled linear MOSFET resistance for the synapse weight emulation. The speed in neural computation is also improved by using the simple switch controlled by the neural input as described in previous works. The general flexibility is attained by the independent characteristic of each synapse cell and the modular structure of URAN chip. As in Table 1 of URAN chip feature, the chip is operated under the flexible control, that is, the various mode of synaptic connection per neuron or the extendable weight accuracy can be implemented. It is not limited for the asynchronous/direct interchip expansion in size or speed. In fact, 16 fully connected module of URAN is selected from external and independently - it is possible to select either one by one or all at once. Table 1. URAN Chip Features Total Synapses Computation Speed Weight Accuracy Module No. Module Size 135,424 connections 200 Giga Connections Per S 8 Bit 16 92 X 92 Implementation of Neural Hardware with the Neural VLSI of URAN 813 As all circuits over the chip except digital decoder unit are operated in analog transistor level, the computation speed is relatively high and even can be improved substantially. The cell size including interconnection area in conventional short-channel technology is reduced less than 900 1.1 m2 . From its expected and measured linear characteristic, URAN has the accuracy more than 256 linear levels. The accuracy extendability and flexible modularity are inherent in electrical wired- OR characteristics as each synapse is an independent bipolar current source with switch. No additional clocking or any limited synchronous operation is required in this case, while it is indispensible in most of conventional digital neural hardware or analog-digital neural chip. Therefore, any size of neural network can be integrated in VLSI or module hardware merely by placing the cell in 2 dimensional array without any timing limitation or loading effect. 3. Neural Hardware with URAN - Module Expansion URAN is the full custom VLSI of analog-digital mixed operation. The prototype of URAN chip is fabricated in 1.0tI digital CMOS technology. The chip contains 135,424 synapses with 8 bit weight accuracy on a 13 X 13 mm2 die size using single poly double metal technology. As summarized in Table 1 of chip features, the chip allows the variety of configuration. In the prototype chip, 16 fully connected module of 92 X 92 can be selected from external and independently - selecting independent module either one by one, several or all at a time IS possible. With URAN's synapse circuit of linear Voltage-controlled bipolar current source, the synaptic multiplication with weight value is done with the switching transistor, in a similar way of analog-sampled data type. The accuracy enhancement and flexible modularity of URAN are inherent in its electrical wired-OR interface from each independent bipolar current source. And the neural network hardware module can be realized in any size with the multi URAN chips. 4. Considerations on the Reduced Precision URAN chip is applied for the case of Korean speaker independent speech recognition. By changing numbers of hidden units and input accuracy, the result of simulations has not shown any problems in It means that the overall performance is not recognition accuracy. severely affected from the accuracy of weight, input, and output with URAN. Also, it was possible to train with 2 or 1 decimal accuracy for input and output, which is equivalent to 8 bit or 4 bit precisions. With 20 hidden units for the Korean spoken 10 digit recognition, 2 decimal input accuracy yields 99.2% and l ' decimal input accuracy yields 98.6%, 814 II-Song Han, Hwang-Soo Lee, Ki-Chul Kim while binary I-bit input results 96.6%. The following is the condition for the experimentation. The general result is summarized in Table 2. Conditions for Training and Test ? 2,000 samples from 10 women and 10 men ( 10 times X 10 digits X 20 persons) [] Training with 500 spoken samples of 10 digits in Korean from 10 persons (5 times X 10 digits X [ 5 women and 5 men ] ) from 2,000 samples [] Recognition Test with 1,000 spoken samples from the other 10 persons of women and men. Preprocessing of samples [] sampled at 10KHz with 12bit accuracy [] preemphasis with 0.95 [] Hamming window of 20ms [] 17 channel critical-band filter bank [] noise added for the SNR of 3OdB, 2OdB, 10dB, OdB Table 2. Low Accuracy Connection with Linear Neuron Input / Output Accuracy SNR Ratio clean 30 dB 20 dB 10 dB o dB 2 decimal 1 decimal 1 bit 97.5% 96.2% 90.1% 59.8% 30.8% 97.2% 96.6% 91.3% 59.9% 29.5% 90.7% 90.5% 86.6% 68.0% 38.5% In case of servo control, the digital VCR for industrial purpose is Six inputs are used to minimize the modelled for the application. number of hidden units and 20 hidden units are configured for one output. For the adaptation to URAN, the linear neuron function is used during the simulation. The weight accuracy during the learning phase using conventional computer is 4 byte and that in the recall phase using URAN chip is 1 byte. With this limitation, the overall performance is not severely degraded, that is, the reduction of error is attained up to 70% improvement comparing to the conventional method. The nonideal factor of 30% results from the limitation in learning data as well as the limited hardware. Current results are suitable for the digital VCR or compact cam coder in noisy environment Implementation of Neural Hardware with the Neural VLSI of URAN 815 5. Conclusion In this paper, it is proved to be suitable for the application to the multi-layer perceptron with the use of URAN chip, which is fabricated in conventional digital CMOS technology - 1.0# single poly double metal. The reduced weight accuracy of 1 byte is proved to be enough to obtain high perfonnance using the linear neuron and URAN. With 8 test chips of 135,424 connections, it is now under development of the practical module of neural hardware with million connections and tera connections per second - comparable to the power of biological neuro-system of some insects. The size of the hardware is smaller than A4 size and is designed for more general recognition system. The flexible modularity of URAN makes it possible to realize a 1,000,000 connections neural chip in 0.5# CMOS technology and a general purpose neural hardware of hundreds of tera connections or more. References II Song Han and Ki-Hwan Ahn, "Neural Network VLSI Chip Implementation of Analog-Digital Mixed Operation for more than 100,000 Connections" MicroNeuro'93, pp. 159-162, 1993 M. Brownlow, L. Tara s senko , A. F. Murray, A. Hamilton, I SHan, H. M. Reekie, "Pulse Firing Neural Chips Implementing Hundreds of Neurons," NIPS2, pp. 785-792, 1990
886 |@word reconstructible:2 implemented:1 effect:1 murray:1 loading:1 added:1 emulation:1 laboratory:1 realized:1 filter:1 pulse:2 simulation:4 ll:2 during:2 implementing:1 speaker:2 die:1 capacity:1 reduction:1 m:1 configuration:1 contains:1 decoder:1 selecting:1 biological:1 indispensible:1 past:1 interface:2 current:4 comparing:2 consideration:1 decimal:5 ratio:1 korean:5 realize:1 info:2 khz:1 designed:3 purpose:2 million:1 analog:13 extend:1 implementation:6 selected:2 neuron:7 short:1 successfully:1 extended:2 frame:1 odb:3 han:4 ahn:1 direct:1 voltage:2 required:1 improvement:2 connection:11 industrial:1 expected:1 kim:3 binary:1 multi:5 suggested:1 additional:1 integrated:1 hidden:4 vlsi:10 window:1 ii:2 soo:3 full:2 circuit:4 overall:2 flexible:4 coder:1 insect:1 uran:27 including:1 development:1 substantially:1 power:1 suitable:3 critical:1 retrieval:1 spoken:4 once:1 fabricated:2 hwan:1 technology:5 mm2:1 placing:1 controlled:4 neuro:1 ti:1 basic:1 bipolar:3 control:5 unit:5 inherent:2 byte:3 cell:3 hamilton:1 multiplication:1 recognize:1 timing:1 fully:2 severely:2 switching:1 phase:2 source:3 suffered:1 mixed:8 men:3 limitation:3 microneuro:1 firing:1 huge:1 digital:16 db:5 clocking:1 custom:2 metal:2 reekie:1 bank:1 limited:4 operated:3 enough:1 switch:2 practical:1 variety:1 asynchronous:1 vcr:2 implement:1 korea:4 prototype:2 digit:5 perfonnance:1 perceptron:3 area:1 synchronous:1 six:1 tera:2 world:2 word:2 song:4 forward:1 resistance:2 universally:2 speech:2 preprocessing:1 cost:1 compact:1 conventional:5 equivalent:1 snr:2 hundred:2 preferred:1 band:1 independently:2 hardware:14 reduced:5 m2:1 extendable:1 person:3 array:1 modularity:3 nips2:1 per:3 table:5 lee:3 dong:1 multichip:1 discrete:1 ku:1 channel:2 affected:1 expansion:2 poly:2 changing:1 woman:3 clean:1 recognition:7 external:3 noise:1 merely:1 module:10 angle:1 electrical:2 summarized:2 telecom:1 interchip:1 configured:1 connected:2 electronic:1 precision:3 servo:3 comparable:1 environment:1 comm:2 bit:7 ki:4 layer:5 shan:1 cam:1 trained:1 minimize:1 accuracy:21 degraded:1 characteristic:3 yield:2 software:1 modelled:1 chip:25 speed:6 various:1 train:1 brownlow:1 mosfet:2 fast:1 effective:1 relatively:2 nonideal:1 artificial:2 synapsis:2 electrically:1 synaptic:2 describes:1 smaller:1 modular:1 kaist:2 pp:2 interconnection:1 hamming:1 noisy:1 sampled:2 proved:3 recall:1 transistor:2 adaptability:1 adaptation:1 back:1 attained:2 operation:3 except:1 experimentation:1 flexibility:3 improved:3 synapse:5 eight:1 done:1 total:1 chul:3 double:2 enhancement:1 select:1 tara:1 wired:2 giga:1 cmos:3 propagation:1 lack:1 a4:1 seoul:3 mode:2 measured:1 dept:2 hwang:3 tested:2
7,120
887
Finding Structure in Reinforcement Learning Sebastian Thrun University of Bonn Department of Computer Science nr R6merstr. 164, D-53117 Bonn, Germany E-mail: thrun@carbon.informatik.uni-bonn.de Anton Schwartz Dept. of Computer Science Stanford University Stanford, CA 94305 Email: schwartz@cs.stanford.edu Abstract Reinforcement learning addresses the problem of learning to select actions in order to maximize one's performance in unknown environments. To scale reinforcement learning to complex real-world tasks, such as typically studied in AI, one must ultimately be able to discover the structure in the world, in order to abstract away the myriad of details and to operate in more tractable problem spaces. This paper presents the SKILLS algorithm. SKILLS discovers skills, which are partially defined action policies that arise in the context of multiple, related tasks. Skills collapse whole action sequences into single operators. They are learned by minimizing the compactness of action policies, using a description length argument on their representation. Empirical results in simple grid navigation tasks illustrate the successful discovery of structure in reinforcement learning. 1 Introduction Reinforcement learning comprises a family of incremental planning algorithms that construct reactive controllers through real-world experimentation. A key scaling problem of reinforcement learning, as is generally the case with unstructured planning algorithms, is that in large real-world domains there might be an enormous number of decisions to be made, and pay-off may be sparse and delayed. Hence, instead of learning all single fine-grain actions all at the same time, one could conceivably learn much faster if one abstracted away the myriad of micro-decisions, and focused instead on a small set of important decisions. But this immediately raises the problem of how to recognize the important, and how to distinguish it from the unimportant. This paper presents the SKILLS algorithm. SKILLS finds partially defined action policies, called skills, that occur in more than one task. Skills, once found, constitute parts of solutions to multiple reinforcement learning problems. In order to find maximally useful skills, a description length argument is employed. Skills reduce the number of bytes required to describe action policies. This is because instead of having to describe a complete action policy for each task separately, as is the case in plain reinforcement learning, skills constrain multiple task to pick the same actions, and thus reduce the total number of actions required 386 Sebastian Thrun, Anton Schwartz for representing action policies. However, using skills comes at a price. In general, one cannot constrain actions to be the same in multiple tasks without ultimately suffering a loss in performance. Hence, in order to find maximally useful skills that infer a minimum loss in performance, the SKILLS algorithm minimizes a function of the form E = PERFORMANCE LOSS + 1J' DESCRIPTION LENGTH. (1) This equation summarizes the rationale of the SKILLS approach. The reminder of this paper gives more precise definitions and learning rules for the terms" PERFORMANCE LOSS" and "DESCRIPTION LENGTH," using the vocabulary of reinforcement learning. In addition, experimental results empirically illustrate the successful discovery of skills in simple grid navigation domains. 2 Reinforcement Learning Reinforcement learning addresses the problem of learning, through experimentation, to act so as to maximize one's pay-off in an unknown environment. Throughout this paper we will assume that the environment of the learner is a partially controllable Markov chain [1]. At any instant in time the learner can observe the state of the environment, denoted by s E S, and apply an action, a E A. Actions change the state of the environment, and also produce a scalar pay-off value, denoted by rs,a E ~. Reinforcement learning seeks to identify an action policy, 7f : S --+ A, i.e., a mapping from states s E S to actions a E A that, if actions are selected accordingly, maximizes the expected discounted sum offuture pay-off R = E [f ,t-til (2) r t ]. t=tu Here I (with 0 S,S 1) is a discount factor that favors pay-offs reaped sooner in time, and rt refers to the expected pay-off at time t. In general, pay-off might be delayed. Therefore, in order to learn an optimal 7f, one has to solve a temporal credit assignment problem [11]. To date, the single most widely used algorithm for learning from delayed pay-off is QLearning [14]. Q-Learning solves the problem of learning 7f by learning a value function, denoted by Q : S x A --+~. Q maps states s E S and actions a E A to scalar values. After learning, Q(s, a) ranks actions according to their goodness: The larger the expected cumulative pay-offfor picking action a at state s, the larger the value Q(s, a). Hence Q, once learned, allows to maximize R by picking actions greedily with respect to Q: 7f(s) = argmax Q(s, a) aEA The value function Q is learned on-line through experimentation. Initially, all values Q( s, a) are set to zero. Suppose during learning the learner executes action a at state s, which leads to a new state s' and the immediate pay-off rs ,a' Q-Learning uses this state transition to update Q(s, a): Q(s, a) with V(s') ;- (1 - a) . Q(s, a) m~x Q(s', a + a? (rs,a + I' V(s')) (3) a) The scalar a (O<aSl) is the learning rate, which is typically set to a small value that is decayed over time. Notice that if Q(s, a) is represented by a lookup-table, as will be the case throughout this paper, the Q-Learning rule (3) has been shown] to converge to a value function Qopt( s, a) which measures the future discounted pay-off one can expect to receive upon applying action a in state s, and acting optimally thereafter [5, 14]. The greedy policy 7f(s) = argmax a Qopt(s, a) maximizes R. I under certain conditions concerning the exploration scheme, the environment and the learning rate Finding Structure in Reinforcement Learning 3 387 Skills Suppose the learner faces a whole collection of related tasks, denoted by B, with identical states 5 and actions A. Suppose each task b E B is characterized by its individual payoff function, denoted by rb(s, a). Different tasks may also face different state transition probabilities. Consequently, each task requires a task-specific value function, denoted by Qb(S, a), which induces a task-specific action policy, denoted by '!rb. Obviously, plain QLearning, as described in the previous section, can be employed to learn these individual action policies. Such an approach, however, cannot discover the structure which might inherently exist in the tasks. In order to identify commonalities between different tasks, the SKILLS algorithm allows a learner to acquire skills. A skill, denoted by k, represents an action policy, very much like '!rb. There are two crucial differences, however. Firstly, skills are only locally defined, on a subset Sk of all states S. Sk is called the domain of skill k. Secondly, skills are not specific to individual tasks. Instead, they apply to entire sets of tasks, in which they replace the task-specific, local action policies. Let f{ denote the set of all skills. In general, some skills may be appropriate for some tasks, but not for others. Hence, we define a vector of usage values Uk,b (with 0 ~ Uk,b ~ 1 for all kEf{ and all b E B). Policies in the SKILLS algorithm are stochastic, and usages Uk,b determine how frequently skill k is used in task b. At first glance, Uk,b might be interpreted as a probability for using skill k when performing task b, and one might always want to use skill k in task b if Uk ,b = 1, and never use skill k if Uk ,b = 0. 2 However, skills might overlap, i.e., there might be states S which occurs in several skill domains, and the usages might add to a value greater than 1. Therefore, usages are normalized, and actions are drawn probabilistical1y according to the normalized distribution: U~,k . mk(s) (4) (with 8= 0) k'EK Here Pb( kls) denotes the probability for using skill k at state s, if the learner faces task b. The indicator function mk (s) is the membership function for skill domains, which is 1 if s E Sk and 0 otherwise. The probabilistic action selection rule (4) makes it necessary to redefine the value Vb (s) of a state s. If no skill dictates the action to be taken, actions will be drawn according to the Qb-optimal policy '!r;(s) argmax Qb(S, a) , .lEA as is the case in plain Q-Learning. The probability for this to happen is Pb*(s) L Pb(kls) . kEK Hence, the value of a state is the weighted sum 1- Vb(S) P;(s) . vt(s) +L Pb(kls) . Qb(S, '!rk(S)) (5) kEK with vt(s) Qb(S, '!rb(s)) = TEa; Qb(S, a) Why should a learner use skills, and what are the consequences? Skills reduce the freedom to select actions, since mUltiple policies have to commit to identical actions. Obviously, such 2This is exactly the action selection mechanism in the SKILLS algorithm if only one skill is applicable at any given state s. 388 Sebastian Thrun, Anton Schwartz a constraint will generally result in a loss in peiformance. This loss is obtained by comparing the actual value of each state s, Vb(S), and the value ifno skill is used, VtCs): LOSS = ~ ~ Vb*(s) - Vb(S) sES bEB '''---v,----' = LO S S( s ) (6) If actions prescribed by the skills are close to optimal, i.e., if Vb*(s) ~ Vb(S)('v'S E S), the loss will be small. If skill actions are poor, however, the loss can be large. Counter-balancing this loss is the fact that skills give a more compact representation of the learner's policies. More specifically, assume (without loss of generality) actions can be represented by a single byte, and consider the total number of bytes it takes to represent the policies of all tasks b E B. In the absence of skills, representing all individual policies requires IBI . lSI bytes, one byte for each state in S and each task in B. If skills are used across multiple tasks, the description length is reduced by the amount of overlap between different tasks. More specifically, the total description length required for the specification of all policies is expressed by the following term : ~ ~ P;(s) + ~ ISkl DL sE S bEB (7) kEK ~~-----v~---------j = DL( s ) If all probabilities are binary, i.e. , Pb(kls) and P;(s) E {O, I}, DL measures precisely the number of bytes needed to represent all skill actions, plus the number of bytes needed to represent task-specific policy actions where no skill is used. Eq. (7) generalizes this measure smoothly to stochastic policies. Notice that the number of skills If{ I is assumed to be constant and thus plays no part in the description length DL. Obviously, minimizing LOSS maximizes the pay-off, and minimizing DL maximizes the compactness of the representation of the learner's policies. In the SKILLS approach, one seeks to minimize both (cf Eq. (1? E = LOSS + TJDL = ~ LOSS(s) + TJDL(s) . (8) sES 11 > 0 is a gain parameter that trades off both target functions. E-optimal policies make heavily use of large skills, yet result in a minimum loss in performance. Notice that the state space may be partitioned completely by skills, and solutions to the individual tasks can be uniquely described by the skills and its usages. If such a complete partitioning does not exist, however, tasks may instead rely to some extent on task-specific, local pOlicies. 4 Derivation of the Learning Algorithm Each skill k is characterized by three types of adjustable variables: skill actions 7rk (s), the skill domain Sk, and skill usages Ub,k. one for each task b E B. In this section we will give update rules that perform hill-cli:nbing in E for each of these variables. As in Q-Learning these rules apply only at the currently visited state (henceforth denoted by s) . Both learning action policies (cf Eq. (3? and learning skills is fully interleaved. Actions. Determining skill actions is straightforward, since what action is prescribed by a skill exclusively affects the performance loss, but does not play any part in the description length. Hence, the action policy 7rk(S) minimizes LOSS(s) (cf Eqs. (5) and (6?: 7rk(S) argmax ~ Pb(kls) . Qb(S. a) .lEA bEB (9) 389 Finding Structure in Reinforcement Learning Domains. Initially, each skill domain Sk contains only a single state that is chosen at random. Sk is changed incrementally by minimizing E(s) for states s which are visited during learning. More specifically, for each skill k, it is evaluated whether or not to include sin Sk by considering E(s) = LOSS(s) + TJDL(s). s E Sk, ifandonlyif E(s)lsESk < E(s)lsf'Sk (otherwises ~ Sk) (10) If the domain of a skill k vanishes completely, i.e., if Sk = 0, it is re-initialized by a randomly selected state. In addition all usage values {Ub ,klb E B} are initialized randomly. This mechanism ensures that skills, once overturned by other skills, will not get lost forever. Usages. Unlike skill domains, which are discrete quantities, usages are real-valued numbers. Initially, they are chosen at random in [0, 1]. Usages are optimized by stochastic gradient descent in E. According to Eq. (8), the derivative of E(s) is the sum of aL~SS(s ) and Ub,k a~uL(s ) . The first term is governed by b,k 8LOSS(s) 8 Ub ,k _ 8Vb(S) 8 'Ub ,k _ - _ 8P;(s) . Q ( *() ) _ " 8Pb(j!S) . Q ( .(? 8 b 7rb s , s ~ 8 b S,7rJ S Ub,k jEK Ub ,k with 8Pb (j!s) (11 ) 8Ub ,k and 8Pb*(s) (12) Here Dkj denotes the Kronecker delta function, which is 1 if k second term is given by 8DL(s) 8Ubk , 8Pb*(s) 8Ukb, ' =j and 0 otherwise. The (13) which can be further transformed using Eqs. (12) and (11). In order to minimize E, usages are incrementally refined in the opposite direction of the gradients: Uk ,b ;....-.- Uk,b - {J. (8V(s) 8Ukb, + TJ 8DL(S?) 8Ukb, (14) Here {J > 0 is a small learning rate. This completes the derivation of the SKILLS algorithm. After each action execution, Q-Learning is employed to update the Q-function. SKILLS also re-calculates, for any applicable skill, the skill policy according to Eq . (9), and adjusts skill domains and usage values based upon Eqs. (10) and (14). 5 Experimental Results The SKILLS algorithm was applied to discover skills in a simple, discrete grid-navigation domain, depicted in Fig. 1. At each state, the agent can move to one of at most eight adjacent grid cells. With a 10% chance the agent is carried to a random neighboring state, regardless of the commanded action. Each corner defines a starting state for one out of four task, with the corresponding goal state being in the opposite corner. The pay-off (costs) for executing actions is -1, except for the goal state, which is an absorbing state with zero pay-off. In a first experiment, we supplied the agent with two skills f{ = {kJ, k2 }. All four tasks were trained in a time-shared manner, with time slices being 2,000 steps long. We used the following parameter settings: TJ = 1.2, 'Y = 1, a = 0.1, and {J = 0.001. After 30000 training steps for each task, the SKILLS algorithm has successfully discovered the two ski11s shown in Figure 1. One of these skills leads the agent to the right door, and 390 Sebastian Thrun. Anton Schwartz 1 I 1/ I 1'/ I //~ I // - /// ",. - - / / r, / - - .- - - - Figure 1: Simple 3-room environment. Start and goal states are marked by circles. The diagrams also shows two skills (black states), which lead to the doors connecting the rooms. the second to the left. Each skill is employed by two tasks. By forcing two tasks to adopt a single policy in the region of the skill, they both have to sacrifice performance, but the loss in performance is considerably small. Beyond the door, however, optimal actions point into opposite directions. There, forcing both tasks to select actions according to the same policy would result in a significant performance loss, which would clearly outweigh the savings in description length. The solution shown in Fig. 1 is (approximately) the global minimum of E, given that only two skills are available. It is easy to be seen that these skills establish helpful building blocks for many navigation tasks. When using more than two skills, E can be minimized further. We repeated the experiment using six skills, which can partition the state space in a more efficient way. Two of the resulting skills were similar to the skills shown in Fig. 1, but they were defined only between the doors. The other four skills were policies for moving out of a corner, one for each corner. Each of the latter four skills can be used in three tasks (unlike two tasks for passing through the middle room), resulting in an improVed description length when compared to the two-skill solution shown in Fig. 1. We also applied skill learning to a more complex grid world, using 25 skills for a total of 20 tasks. The environment, along with one of the skills, is depicted in Fig. 2. Different tasks were defined by different starting positions, goal positions and door configurations which could be open or closed. The training time was typically an order of magnitude slower than in the previous task, and skills were less stable over time. However, Fig. 2 illustrates that modular skills could be discovered even in such complex a domain. 6 Discussion This paper presents the SKILLS algorithm. SKILLS learns skills, which are partial policies that are defined on a subset of all states. Skills are used in as many tasks as possible, while affecting the performance in these tasks as little as possible. They are discovered by minimizing a combined measure, which takes a task performance and a description length argument into account. While our empirical findings in simple grid world domains are encouraging, there are several open questions that warrant future research. Learning speed. In our experiments we found that the time required for finding useful skills is up to an order of magnitude larger than the time it takes to find close-to-optimal policies. Finding Structure in Reinforcement Learning 391 Figure 2: SkiIl found in a more complex grid navigation task. Similar findings are reported in [9]. This is because discovering skills is much harder than learning control. Initially, nothing is know about the structure of the state space, and unless reasonably accurate Q-tables are available, SKILLS cannot discover meaningful skills. Faster methods for learning skills, which might precede the development of optimal value functions, are clearly desirable. Transfer. We conjecture that skills can be helpful when one wants to learn new, related tasks. This is because if tasks are related, as is the case in many natural learning environments, skills allow to transfer knowledge from previously learned tasks to new tasks. In particular, if the learner faces tasks with increasing complexity, as proposed by Singh [10], learning skills could conceivable reduce the learning time in complex tasks, and hence scale reinforcement learning techniques to more complex tasks. Using function approximators. In this paper, performance loss and description length has been defined based on table look-up representations of Q. Recently, various researchers have applied reinforcement learning in combination with generalizing function approximators, such as nearest neighbor methods or artificial neural networks (e.g., [2, 4, 12, 13]). In order to apply the SKILLS algorithm together with generalizing function approximators, the notions of skill domains and description length have to be modified. For example, the membership function mk, which defines the domain of a skill, could be represented by a function approximator which allows to derive gradients in the description length. Generalization in state space. In the current form, SKILLS exclusively discovers skills that are used across mUltiple tasks. However, skills might be useful under multiple circumstances even in single tasks. For example, the (generalized) skill of climbing a staircase may be useful several times in one and the same task. SKILLS, in its current form, cannot represent such skills. The key to learning such generalized skills is generalization. Currently, skills generalize exclusively over tasks, since they can be applied to entire sets of tasks. However, they cannot generalize over states. One could imagine an extension to the SKILLS algorithm, in which skills are free to pick what to generalize over. For example, they could chose to ignore certain state information (like the color of the staircase). It remains to be seen if effective learning mechanisms can be designed for learning such generalized skills. Abstractions and action hierarchies. In recent years, several researchers have recognized the importance of structuring reinforcement learning in order to build abstractions and action 392 Sebastian Thrun, Anton Schwartz hierarchies. Different approaches differ in the origin of the abstraction, and the way it is incorporated into learning. For example, abstractions have been built upon previously learned, simpler tasks [9, 10], previously learned low-level behaviors [7], subgoals, which are either known in advance [15] or determined at random [6], or based on a pyramid of different levels of perceptual resolution, which produces a whole spectrum of problem solving capabilities [3]. For all these approaches, drastically improved problem solving capabilities have been reported, which are far beyond that of plain, unstructured reinforcement learning. This paper exclusively focuses on how to discover the structure inherent in a family of related tasks. Using skills to form abstractions and learning in the resulting abstract problem spaces is beyond the scope of this paper. The experimental findings indicate, however, that skills are powerful candidates for operators on a more abstract level, because they collapse whole action sequences into single entities. References [I] A. G. Barto, S. J. Bradtke, and S. P. Singh. Learning to act using real-time dynamic programming. Artijiciallntelligence, to appear. [2] J. A. Boyan. Generalization in reinforcement learning: Safely approximating the value function. Same volume. [3] P. Dayan and G. E. Hinton. Feudal reinforcement learning. In J. E. Moody, S. J. Hanson, and R. P. Lippmann, editors, Advances in Neural Information Processing Systems 5,1993. Morgan Kaufmann. [4) V. Gullapalli, 1. A. Franklin, and Hamid B. Acquiring robot skills via reinforcement learning. IEEE Control Systems, 272( 1708), 1994. [5) T. Jaakkola, M. I. Jordan, and S. P. Singh. On the convergence of stochastic iterative dynamic programming algorithms. Technical Report 9307, Department of Brain and Cognitive Sciences, MIT, July 1993. [6] L. P. Kaelbling. Hierarchical learning in stochastic domains: Preliminary results. In Paul E. Utgoff, editor, Proceedings of the Tenth International Conference on Machine Learning, 1993. Morgan Kaufmann. [7] L.-J. Lin. Self-supervised Learning by Reinforcementand Artijicial Neural Networks. PhD thesis, Carnegie Mellon University, School of Computer Science, 1992. [8) M. Ring. Two methods for hierarchy learning in reinforcement environments. In From Animals to Animates 2: Proceedings of the Second International Conference on Simulation of Adaptive Behavior. MIT Press, 1993. [9] S. P. Singh. Reinforcement learning with a hierarchy of abstract models. In Proceeding of the Tenth National Conference on Artijiciallntelligence AAAI-92, 1992. AAAI Pressffhe MIT Press. [10] S. P. Singh. Transfer of learning by composing solutions for elemental sequential tasks. Machine Learning, 8,1992. [II] R. S. Sutton. Temporal Credit Assignment in Reinforcement Learning. PhD thesis, Department of Computer and Information Science, University of Massachusetts, 1984. [12) G. J. Tesauro. Practical issues in temporal difference learning. Machine Learning, 8, 1992. [13] S. Thrun and A. Schwartz. Issues in using function approximation for reinforcement learning. In M. Mozer, Pa. Smolensky, D. Touretzky, J. Elman, and A. Weigend, editors, Proceedings of the J993 Connectionist Models Summer School, 1993. Erlbaum Associates. [14] C. J. C. H. Watkins. Learningfrom Delayed Rewards. PhD thesis, King's College, Cambridge, England, 1989. [15] S. Whitehead, J. Karlsson, and J. Tenenberg. Learning multiple goal behavior via task decomposition and dynamic policy merging. In J. H. Connell and S. Mahadevan, editors, Robot Learning. Kluwer Academic Publisher, 1993.
887 |@word middle:1 open:2 seek:2 simulation:1 r:3 decomposition:1 pick:2 harder:1 configuration:1 contains:1 exclusively:4 franklin:1 current:2 comparing:1 artijiciallntelligence:2 yet:1 must:1 grain:1 partition:1 happen:1 designed:1 update:3 greedy:1 selected:2 discovering:1 accordingly:1 firstly:1 simpler:1 along:1 redefine:1 manner:1 sacrifice:1 expected:3 behavior:3 elman:1 planning:2 frequently:1 brain:1 discounted:2 actual:1 little:1 encouraging:1 considering:1 increasing:1 discover:5 maximizes:4 what:3 interpreted:1 minimizes:2 finding:8 temporal:3 safely:1 act:2 exactly:1 k2:1 schwartz:7 uk:8 partitioning:1 control:2 appear:1 local:2 consequence:1 sutton:1 approximately:1 might:10 ibi:1 plus:1 black:1 studied:1 chose:1 collapse:2 commanded:1 practical:1 lost:1 block:1 empirical:2 dictate:1 refers:1 get:1 cannot:5 close:2 selection:2 operator:2 context:1 applying:1 outweigh:1 map:1 peiformance:1 straightforward:1 regardless:1 starting:2 focused:1 resolution:1 unstructured:2 immediately:1 rule:5 adjusts:1 notion:1 target:1 suppose:3 play:2 heavily:1 imagine:1 hierarchy:4 programming:2 us:1 origin:1 pa:1 associate:1 region:1 ensures:1 counter:1 trade:1 mozer:1 environment:10 vanishes:1 complexity:1 utgoff:1 reward:1 dynamic:3 ultimately:2 ukb:3 trained:1 raise:1 singh:5 solving:2 myriad:2 upon:3 learner:10 completely:2 represented:3 various:1 derivation:2 describe:2 effective:1 artificial:1 refined:1 modular:1 stanford:3 solve:1 widely:1 larger:3 s:3 otherwise:2 valued:1 favor:1 commit:1 obviously:3 sequence:2 tu:1 neighboring:1 date:1 description:14 cli:1 elemental:1 convergence:1 produce:2 incremental:1 executing:1 ring:1 illustrate:2 derive:1 kef:1 nearest:1 school:2 eq:8 solves:1 c:1 come:1 indicate:1 differ:1 direction:2 stochastic:5 exploration:1 generalization:3 hamid:1 preliminary:1 secondly:1 extension:1 credit:2 mapping:1 scope:1 commonality:1 adopt:1 klb:1 applicable:2 precede:1 currently:2 visited:2 successfully:1 weighted:1 offs:1 clearly:2 mit:3 always:1 modified:1 beb:3 barto:1 jaakkola:1 structuring:1 focus:1 rank:1 greedily:1 helpful:2 abstraction:5 dayan:1 membership:2 typically:3 entire:2 compactness:2 initially:4 transformed:1 germany:1 issue:2 denoted:9 development:1 animal:1 construct:1 once:3 having:1 never:1 saving:1 identical:2 represents:1 look:1 warrant:1 future:2 minimized:1 others:1 report:1 connectionist:1 micro:1 inherent:1 randomly:2 recognize:1 national:1 individual:5 delayed:4 argmax:4 freedom:1 karlsson:1 navigation:5 tj:2 chain:1 accurate:1 partial:1 necessary:1 unless:1 sooner:1 initialized:2 re:2 circle:1 mk:3 goodness:1 assignment:2 cost:1 kaelbling:1 subset:2 successful:2 erlbaum:1 optimally:1 reported:2 considerably:1 combined:1 decayed:1 international:2 probabilistic:1 off:13 picking:2 connecting:1 together:1 moody:1 thesis:3 aaai:2 henceforth:1 corner:4 cognitive:1 ek:1 derivative:1 til:1 jek:1 account:1 de:1 lookup:1 closed:1 start:1 capability:2 minimize:2 kek:3 kaufmann:2 identify:2 climbing:1 generalize:3 anton:5 informatik:1 researcher:2 executes:1 touretzky:1 sebastian:5 email:1 definition:1 gain:1 massachusetts:1 reminder:1 knowledge:1 color:1 supervised:1 maximally:2 improved:2 evaluated:1 generality:1 glance:1 incrementally:2 defines:2 usage:12 building:1 normalized:2 asl:1 dkj:1 staircase:2 hence:7 adjacent:1 sin:1 during:2 self:1 uniquely:1 generalized:3 hill:1 complete:2 bradtke:1 discovers:2 recently:1 absorbing:1 empirically:1 volume:1 subgoals:1 kluwer:1 significant:1 mellon:1 cambridge:1 ai:1 grid:7 moving:1 specification:1 stable:1 robot:2 add:1 recent:1 forcing:2 tesauro:1 certain:2 binary:1 vt:2 approximators:3 seen:2 minimum:3 greater:1 morgan:2 employed:4 recognized:1 converge:1 maximize:3 determine:1 july:1 ii:1 multiple:9 desirable:1 rj:1 infer:1 technical:1 faster:2 characterized:2 england:1 academic:1 long:1 lin:1 concerning:1 calculates:1 overturned:1 controller:1 circumstance:1 represent:4 pyramid:1 cell:1 lea:2 receive:1 addition:2 want:2 fine:1 separately:1 affecting:1 completes:1 diagram:1 lsf:1 crucial:1 publisher:1 qopt:2 operate:1 unlike:2 jordan:1 door:5 mahadevan:1 easy:1 affect:1 opposite:3 reduce:4 ubk:1 gullapalli:1 whether:1 six:1 ul:1 aea:1 passing:1 constitute:1 action:54 generally:2 useful:5 se:1 unimportant:1 amount:1 discount:1 locally:1 induces:1 reduced:1 supplied:1 exist:2 lsi:1 notice:3 delta:1 rb:5 discrete:2 carnegie:1 tea:1 key:2 thereafter:1 four:4 enormous:1 pb:10 drawn:2 tenth:2 sum:3 year:1 weigend:1 powerful:1 family:2 throughout:2 decision:3 summarizes:1 scaling:1 vb:8 interleaved:1 pay:14 summer:1 distinguish:1 occur:1 constraint:1 precisely:1 constrain:2 kronecker:1 feudal:1 bonn:3 speed:1 argument:3 prescribed:2 qb:7 performing:1 connell:1 conjecture:1 department:3 according:6 combination:1 poor:1 across:2 partitioned:1 conceivably:1 taken:1 equation:1 previously:3 remains:1 mechanism:3 needed:2 know:1 tractable:1 whitehead:1 generalizes:1 available:2 experimentation:3 apply:4 observe:1 eight:1 away:2 appropriate:1 hierarchical:1 slower:1 artijicial:1 denotes:2 cf:3 include:1 instant:1 learningfrom:1 build:1 establish:1 approximating:1 move:1 question:1 quantity:1 occurs:1 rt:1 nr:1 gradient:3 conceivable:1 thrun:7 entity:1 mail:1 extent:1 length:14 minimizing:5 acquire:1 carbon:1 policy:33 unknown:2 adjustable:1 perform:1 markov:1 descent:1 immediate:1 payoff:1 hinton:1 incorporated:1 precise:1 discovered:3 required:4 optimized:1 hanson:1 learned:6 address:2 able:1 beyond:3 smolensky:1 built:1 overlap:2 natural:1 rely:1 boyan:1 indicator:1 representing:2 scheme:1 carried:1 kj:1 byte:7 discovery:2 determining:1 loss:22 expect:1 fully:1 rationale:1 approximator:1 agent:4 editor:4 balancing:1 lo:1 changed:1 free:1 drastically:1 allow:1 neighbor:1 face:4 sparse:1 slice:1 plain:4 vocabulary:1 world:6 cumulative:1 transition:2 made:1 reinforcement:26 collection:1 adaptive:1 far:1 skill:121 uni:1 qlearning:2 compact:1 forever:1 ignore:1 lippmann:1 abstracted:1 global:1 assumed:1 spectrum:1 iterative:1 sk:11 why:1 table:3 learn:4 reasonably:1 transfer:3 ca:1 controllable:1 inherently:1 composing:1 complex:6 domain:17 whole:4 arise:1 paul:1 nothing:1 suffering:1 repeated:1 fig:6 position:2 comprises:1 candidate:1 governed:1 perceptual:1 watkins:1 learns:1 rk:4 specific:6 dl:7 sequential:1 merging:1 importance:1 phd:3 magnitude:2 execution:1 illustrates:1 smoothly:1 depicted:2 generalizing:2 expressed:1 partially:3 scalar:3 acquiring:1 chance:1 kls:5 goal:5 marked:1 king:1 consequently:1 room:3 price:1 replace:1 absence:1 change:1 shared:1 specifically:3 except:1 determined:1 acting:1 called:2 total:4 experimental:3 meaningful:1 select:3 college:1 latter:1 reactive:1 ub:8 dept:1
7,121
888
A Novel Reinforcement Model of Birdsong Vocalization Learning Kenji Doya ATR Human Infonnation Processing Research Laboratories 2-2 Hikaridai, Seika, Kyoto 619-02, Japan Terrence J. Sejnowski Howard Hughes Medical Institute UCSD and Salk Institute, San Diego, CA 92186-5800, USA Abstract Songbirds learn to imitate a tutor song through auditory and motor learning. We have developed a theoretical framework for song learning that accounts for response properties of neurons that have been observed in many of the nuclei that are involved in song learning. Specifically, we suggest that the anteriorforebrain pathway, which is not needed for song production in the adult but is essential for song acquisition, provides synaptic perturbations and adaptive evaluations for syllable vocalization learning. A computer model based on reinforcement learning was constructed that could replicate a real zebra finch song with 90% accuracy based on a spectrographic measure. The second generation of the birdsong model replicated the tutor song with 96% accuracy. 1 INTRODUCTION Studies of motor pattern generation have generally focussed on innate motor behaviors that are genetically preprogrammed and fine-tuned by adaptive mechanisms (Harris-Warrick et al., 1992). Birdsong learning provides a favorable opportunity for investigating the neuronal mechanisms for the acquisition of complex motor patterns. Much is known about the neuroethology of birdsong and its neuroanatomical substrate (see Nottebohm, 1991 and Doupe, 1993 for reviews), but relatively little is known about the overall system from a computational viewpoint. We propose a set of hypotheses for the functions of the brain nuclei in the song system and explore their computational strength in a model based on biological constraints. The model could reproduce real and artificial birdsongs in a few hundred learning trials. 102 Kenji Doya, Terrence 1. Sejnowski ----+ Direct Motor Pathway ----+ Anterior Forebrain Pathway Figure 1: Major songbird brain nuclei involved in song control. The dark arrows show the direct motor control pathway and the gray arrows show the anterior forebrain pathway. Abbreviations: Uva, nucleus uvaeforrnis of the thalamus; NIf, nucleus interface of the neostriatum; L, field L (primary auditory are of the forebrain); HVc, higher vocal center; RA, robust nucleus of the archistriatum; DM, dorso-medial part of the nucleus intercollicularis; nXllts, tracheosyringeal part of the hypoglossal nucleus; AVT, ventral area of Tsai of the midbrain; X, area X of lobus parolfactorius; DLM, medial part of the dorsolateral nucleus of the thalamus; LMAN, lateral magnocellular nucleus of the anterior neostriatum. 2 NEUROETHOLOGY OF BIRDSONG Although songs from individual birds of the same species may sound quite similar, a young male songbird learns to sing by imitating the song of a tutor, which is usually the father or another adult male in the colony. If a young bird does not hear a tutor song during a critical period, it will sing short, poorly structured songs, and if a bird is deafened in the period when it practices vocalization, it develops highly abnormal songs. These observations indicate that there are two phases in song learning: the sensory learning phase when a young bird memorizes song templates and the motor learning phase in which the bird establishes the motor programs using auditory feedback (Konishi, 1965). These two phases can be separated by several months in some species, implying that birds have remarkable capability for memorizing complex temporal sequences. Once a song is crystallized, its pattern is very stable. Even deafening the bird has little immediate effect. The brain nuclei involved in song learning are shown in Figure 1. The primary motor control pathway is composed ofUva, NIf, HVc, RA, DM, and nXllts. If any of these nuclei is lesioned, a bird cannot sing normally. Experimental studies suggest that HVc is involved in generating syllable sequences and that RA produces motor commands for each syllable (Vu et aI., 1994). Interestingly, neurons in HVc, RA and nXllts show vigorous auditory responses, suggesting that the motor control system is closely coupled with the auditory system (Nottebohm, 1991). There is also a "bypass" from HVc to RA which consists of area X, DLM, and LMAN called the anterior forebrain pathway (Doupe, 1993). This pathway is not directly involved A Novel Reinforcement of Birdsong Vocalization Learning auditory syllable encoding preprocessing sequence generation reinforcement attention 103 motor pattern generation memory of tutor song normalized evaluation synaptic perturbation gradient estimate Figure 2: Schematic of primary song control nuclei and their proposed functions in the present model of bird song learning. in vocalization because lesions in these nuclei in adult birds do not impair their crystallized songs. However, lesions in area X and LMAN during the motor learning phase result in contrasting deficits. The songs of LMAN-Iesioned birds crystallize prematurely, whereas the songs of area X-Iesioned birds remain variable (Scharff and Nottebohm, 1991). It has been suggested that this pathway is responsible for the storage of song templates (Doupe and Konishi, 1991) or guidance of the synaptic connection from HVc to RA (Mooney, 1992). 3 FUNCTIONAL NEUROANATOMY OF BIRDSONG The song learning process can be decomposed into three stages. In the first stage, suitable internal acoustic representations of syllables and syllable combinations are constructed. This "auditory template" can be assembled by unsupervised learning schemes like clustering and principal components analysis. The second stage involves the encoding of phonetic sequences using the internal representation. If the representation is sparse or nearly orthogonal, sequential transition can be easily encoded by Hebbian learning. The third stage is an inverse mapping from the internal auditory representation into spatio-temporal patterns of motor commands. This can be accomplished by exploration in the space of motor commands using reinforcement learning. The responses of the units that encode the acoustic primitives can be used to the evaluate the resulting auditory signal and direct the exploration. How are these three computational stages organized within the brain areas and pathways of the songbird? Figure 2 gives an overview of our current working hypothesis. Auditory inputs are pre-processed in field L. Some higher-order representations, such as syllables and syllable combinations, are established in HVc depending on the bird's auditory experience. Moreover, transitions between syllables are encoded in the HVc network. The sequential activation of syllable coding units in HVc are transformed into the time course of motor commands in RA. DM and nXIIts control breathing and the muscles in syrinx, bird's vocal organ. 104 Kenji Doya, Terrence 1. Sejnowski a bronchus Figure 3: (a) The syrinx of songbirds. (b) The model syrinx. The consequences of selective lesions of areas in the anterior forebrain pathway (Scharff and Nottebohm, 1991) are consistent with the failures expected for a reinforcement learning system. In particular, we suggest that this pathway serves the function of an adaptive critic with stochastic search elements (Barto et al., 1983). We propose that LMAN perturbs the synaptic connections from HVc to RA and area X regulates LMAN by the song evaluation. Modulation of HVc to RA connection by LMAN is biologically plausible since LMAN input to RA is mediated mainly by NMDA type synapses, which can modulate the amplitude of mainly non-NMDA type synaptic input from HVc (Mooney, 1992). The assumption that area X provides evaluation is supported by the fact that it receives catecholaminergic projection (dopamine of norepinephrine) from a midbrain nucleus AVT (Lewis et al., 1981). These neurotransmitters are used in many species for reinforcement or attention signals. It is known that auditory learning is enhanced when associated with visual or social interaction with the tutor. Area X is a candidate region where auditory inputs from HVc are associated with reinforcing input from AVT during auditory learning. 4 CONSTRUCTION OF SONG LEARNING MODEL In order to test the above hypothesis, we constructed a computer model of the bird song learning system. The specific aim was to simulate the process of explorative motor learning, in which the time course of motor command for each syllable is determined by auditory template matching. We assumed that orthogonal representations for syllables and their sequential activation were already established in HVc and that an auditory template matching mechanism exists in area X. 4.1 The syrinx The bird's syrinx is located near the junction of the trachea and the bronchi (Vicario, 1991). Its sound source is the tympani form membrane which faces to the bronchus on one side and the air sac on the other (Figure 3a). When some of the syringeal muscles contract, the lumen of the bronchus is throttled and produces vibration in the membrane. When stretched along one dimension, the membrane produces harmonic sounds, but when stretched along two dimensions, the sound contains non-harmonic components (Casey and Gaunt, 1985). Accordingly, we provided two sound sources for the model syrinx (Figure 3b). The fundamental frequency of the harmonic component was controlled by the membrane tension in one direction (Tl). The amplitu6e of the noisy component was proportional to the A Novel Reinforcement of Birdsong Vocalization Learning 105 Ex ! ! ! ~ ~Th ? ~T' ? T2 HVc RA OM nXlIts Figure 4: RA units with different spatio-temporal output profiles are driven by locallycoded RVc units. LMAN perturbs the weights W between RVc and RA. The output units in DM and nXIIts drive the model syrinx. membrane tension in an orthogonal direction (T2). Mixture of these sounds went through a bandpass filter whose resonance frequency was controlled by the throttling of the bronchus (Th). The overall sound amplitude was determined by the strength of expiration (Ex). By controlling the time course of these four variables (Ex,Th,Tl,T2), the model could produce wide variety of "bird-like" chirps and warbles. 4.2 Motor pattern generation in RA RA is topographically organized, each part projecting to different motoneuron pools in nXIIts (Vicario, 1991). Also, RA neurons have complex temporal responses to the inputs from RVc (Mooney, 1992). Therefore, we assumed that RA consists of groups of neurons with specific spatial and temporal output properties, as shown in Figure 4. For each of the four motor command variables, we provided several units with different temporal response kernels. The sequential activation of syllable coding units in RVc drove the RA units through the weights W. Their responses were linearly combined and squashed between 0 and 1 to make the final motor commands. 4.3 Weight space search by LMAN and area X With the above model of the motor output, the task is to find a connection matrix W that maximizes a template matching measure. One way for doing this is to perturb the output of the units and correlate it with the input and the evaluation (Barto et aI., 1983). An alternative way, adopted here, is to perturb the weights and correlate them with the evaluation. We used the following stochastic gradient ascent algorithm. The weight matrix W is modulated by ~ W, given by the sum of the evaluation gradient estimate G and a random component. The modulated weight persists if the resulting vocalization is better than the recent average evaluation E[r]. The evaluation gradient estimate G is updated by the sum 106 Kenji Doya, Terrence J. Sejnowski of the perturbations ~ W weighted with the normalized evaluation. ~ W := r := G + random perturbation evaluation of the song generated with W W:= G := W+~W if r + ~W > E[r] E[r] JVR" ~W + (1- o:)G, V[r] r 0: where 0 < 0: < 1 provides a form of "momentum" in weight space similar to that used in supervised learning. The average and the variance of evaluation are also estimated on-line as follows. 4.4 + (1 - 0: )E[r] E[r] := o:r V[r] := o:(r - E[r]? + (1 - o:)V[r] . Spectrographic template matching We assumed that evaluation for vocalization is given separately for each of the syllables in a song. The sound signal was analyzed by an 80 channel spectrogram. Each output channel was sent to an analog delay line similar to the gamma filter (de Vries and Principe, 1992). The snapshot image of this (80 channels) x (12 steps) delay line at the end of each syllable was stored as the template. The same delay line image of the syllable generated by the model was compared with the template. This allowed some compensation for variable syllable duration. The direction cosine between the two delay line images was used for the reinforcement signal, r. 5 SIMULATION RESULTS One phrase of a recorded zebra finch song (Figure 5a) was the target. Templates were stored for the five syllables in the phrase. Five HVc units coded the five different syllables and 16 RA units represented the four output variables and four different temporal kernels. Learning was started with small random weights. After 200 to 300 trials, the syllable evaluation by direction cosine reached 0.9 (Figure 5d, solid line). The syllables of the learned song resembled the overall frequency profiles of the original syllables. The complex spectrographic structure of the original syllables were, however, not accurate (Figure 5b). One reason for this imperfect replication could be the difference between the vocal organs of the real zebra finch syrinx and our model syrinx. In order to check the significance of this difference, we took syllable templates from the model song (Figure 5b) and trained another model with random initial weights. In this case, the direction cosine went up to 0.96 (Figure 5d, dotted line) and they sounded quite similar to human ears (Figure 5c). We also checked the importance of the gradient estimate G in our algorithm. The dashed line in Figure 5d shows the performance of the model with G 0: a simple random walk learning. The learning was hopelessly slow and resembles the deficit seen after lesion of area X (Figure 2). = A Novel Reinforcement of Birdsong Vocalization Learning 1.0 107 ~-~--~--~-~---, d - zebra finch JOng model song - - - - random walk trilla Figure 5: Spectrograms of (a) the original zebra finch song, (b) the learned song based on the tutor in (a), and (c) the second generation learned song based on the tutor in (b). (d) Learning curves for two tutors: a zebra finch song (solid line) and a model song (dotted line) compared with an undirected search in weight space (dashed line). Weight perturbation was given by a Gaussian distribution with (J 0.1. The averaging parameter was a 0.2. Simulating 500 trials took 30 minutes on Sparc Station 10. = 6 = Discussion We have assumed that each vocalized syllable was separately evaluated. If the evaluation is given only at the end of one song or a phrase, learning can be much more difficult because of the temporal credit assignment problem. If we assume that birds take the easiest strategy available, there should be syllable specific evaluation and separate perturbation mechanisms. In some songbirds, individual syllables are practiced out of order at an early stage, and only later is the sequence matched to the auditory template. Selectivity of auditory responses in both HVc and area X develop during motor learning (Volman, 1993; Doupe, 1993). We can expect such change in response tuning in area X if the evaluations of syllables or syllable sequences are normalized with respect to recent average performance, as we assumed in our model. Many simplifying assumptions were made in the present model: syllables were unary coded in HVc; simple spectrographic template matching was used; the number of motor output variables and temporal kernels were fairly small; and the sound synthesizer was much simpler than a real syrinx. However, it is not difficult to replace these idealizations with more biologically accurate models. Since the number of learning trials needed in the present model was much less than in the real birdsong learning (tens of thousands of trials), there is margin for further elaboration. 108 Kenji Doya, Terrence 1. Sejnowski Acknowledgments We thank M. Lewicki for the zebra finch song data and M. Konishi, A. Doupe, M. Lewicki, E. Vu, D. Perkel and G. Striedter for their helpful discussions. References Barto, A. G., Sutton, R. S., and Anderson, C. W. (1983). Neuronlike adaptive elements that can solve difficult learning control problems. IEEE Transactions on System, Man, and Cybernetics, SMC-13:834-846. Casey, R. M. and Gaunt, A. S. (1985). Theoretical models of the avian syrinx. Journal of Theoretical Biology, 116:45-64. de Vries, B. and Principe, J. C. (1992). The gamma model-A new neural model for temporal processing. Neural Networks, 5:565-576. Doupe, A. J. (1993). A neural circuit specialized for vocal learning. Current Opinion in Neurobiology, 3:104-111. Doupe, A. J. and Konishi, M. (1991). Song-selective auditory circuits in the vocal control system of the zebra finch. Proceedings of the National Academy of Sciences, USA, 88: 11339-11343. Harris-Warrick, R. M., Marder, E., Selverston, A. I., and Moulins, M. (1992). Dynamic Biological Networks-The Stomatogastric Nervous System. MIT Press, Cambridge, MA. Konishi, M. (1965). fhe role of auditory feedback in the control of vocalization in the white-crowned sparrow. Zeitschrift fur Tierpsychologie, 22:770-783. Lewis, J. W., Ryan, S. M., Arnold, A. P., and Butcher, L. L. (1981). Evidence for a catecholarninergic projection to area x in the zebra finch. Journal of Comparative Neurology, 196:347-354. Mooney, R. (1992). Synaptic basis of developmental plasticity in a birdsong nucleus. Journal of Neuroscience, 12:2464-2477. Nottebohm, F. (1991). Reassessing the mechanisms and origins of vocal learning in birds. Trends in Neurosciences, 14:206-211. Scharff, C. and Nottebohm, F. (1991). A comparative study of the behavioral deficits following lesions of various parts of the zebra finch song systems: Implications for vocal learning. Journal of Neuroscience, 11 :2896-2913. Vicario, D. S. (1991). Neural mechanisms of vocal production in songbirds. Current Opinion in Neurobiology, 1:595-600. Volman, S. F. (1993). Development of neural selectivity for birdsong during vocal learning. Journal of Neuroscience, 13:4737--4747. Vu, E. T., Mazurek, M. E., and Kuo, Y.-C. (1994). Identification of a forebrain motor programming network for the learned song of zebra finches. Journal of Neuroscience, 14:6924-6934.
888 |@word trial:5 replicate:1 simulation:1 simplifying:1 solid:2 initial:1 contains:1 practiced:1 tuned:1 interestingly:1 current:3 anterior:5 activation:3 synthesizer:1 explorative:1 plasticity:1 motor:25 medial:2 implying:1 nervous:1 imitate:1 accordingly:1 short:1 provides:4 simpler:1 five:3 along:2 constructed:3 direct:3 replication:1 consists:2 pathway:12 behavioral:1 expected:1 ra:19 behavior:1 seika:1 brain:4 perkel:1 decomposed:1 little:2 provided:2 moreover:1 matched:1 maximizes:1 circuit:2 easiest:1 developed:1 contrasting:1 selverston:1 temporal:10 control:9 normally:1 medical:1 unit:11 persists:1 consequence:1 zeitschrift:1 sutton:1 encoding:2 modulation:1 bird:19 reassessing:1 chirp:1 resembles:1 smc:1 acknowledgment:1 responsible:1 vu:3 hughes:1 practice:1 area:16 projection:2 matching:5 pre:1 vocal:9 suggest:3 cannot:1 nxiits:3 storage:1 center:1 primitive:1 attention:2 duration:1 sac:1 stomatogastric:1 konishi:5 expiration:1 updated:1 diego:1 enhanced:1 construction:1 controlling:1 drove:1 substrate:1 target:1 programming:1 hypothesis:3 origin:1 element:2 trend:1 located:1 observed:1 role:1 thousand:1 region:1 went:2 developmental:1 lesioned:1 dynamic:1 preprogrammed:1 trained:1 topographically:1 iesioned:2 basis:1 easily:1 represented:1 neurotransmitter:1 various:1 separated:1 sejnowski:5 artificial:1 crystallized:2 quite:2 encoded:2 whose:1 plausible:1 solve:1 noisy:1 vocalization:10 final:1 sequence:6 took:2 propose:2 interaction:1 poorly:1 academy:1 mazurek:1 produce:4 generating:1 comparative:2 depending:1 develop:1 avian:1 colony:1 kenji:5 involves:1 indicate:1 syrinx:11 throttling:1 direction:5 closely:1 filter:2 stochastic:2 exploration:2 human:2 opinion:2 biological:2 ryan:1 credit:1 mapping:1 major:1 ventral:1 early:1 favorable:1 infonnation:1 vibration:1 organ:2 establishes:1 weighted:1 mit:1 catecholaminergic:1 gaussian:1 aim:1 neuroethology:2 moulins:1 command:7 barto:3 encode:1 lman:10 casey:2 fur:1 check:1 mainly:2 helpful:1 unary:1 butcher:1 reproduce:1 transformed:1 selective:2 overall:3 warrick:2 development:1 resonance:1 spatial:1 fairly:1 field:2 once:1 biology:1 unsupervised:1 nearly:1 t2:3 develops:1 few:1 composed:1 gamma:2 national:1 individual:2 phase:5 neuronlike:1 highly:1 evaluation:17 male:2 mixture:1 analyzed:1 implication:1 accurate:2 experience:1 orthogonal:3 walk:2 guidance:1 theoretical:3 hvc:18 assignment:1 phrase:3 hundred:1 father:1 delay:4 stored:2 finch:11 combined:1 fundamental:1 contract:1 terrence:5 pool:1 recorded:1 ear:1 japan:1 account:1 suggesting:1 de:2 coding:2 later:1 memorizes:1 doing:1 reached:1 capability:1 om:1 air:1 accuracy:2 variance:1 identification:1 drive:1 cybernetics:1 mooney:4 archistriatum:1 synapsis:1 synaptic:6 checked:1 failure:1 acquisition:2 frequency:3 involved:5 dm:4 associated:2 auditory:20 organized:2 nmda:2 amplitude:2 vicario:3 higher:2 supervised:1 tension:2 response:8 evaluated:1 anderson:1 stage:6 nif:2 working:1 receives:1 gray:1 innate:1 usa:2 effect:1 spectrographic:4 normalized:3 laboratory:1 white:1 during:5 songbird:7 cosine:3 interface:1 image:3 harmonic:3 novel:4 specialized:1 functional:1 overview:1 regulates:1 analog:1 crystallize:1 cambridge:1 zebra:11 ai:2 stretched:2 tuning:1 crowned:1 stable:1 recent:2 driven:1 sparc:1 phonetic:1 selectivity:2 accomplished:1 muscle:2 seen:1 motoneuron:1 spectrogram:2 neuroanatomy:1 period:2 dashed:2 signal:4 fhe:1 sound:9 kyoto:1 thalamus:2 hebbian:1 sparrow:1 elaboration:1 coded:2 controlled:2 schematic:1 dopamine:1 kernel:3 syringeal:1 whereas:1 fine:1 separately:2 source:2 ascent:1 sent:1 undirected:1 near:1 variety:1 imperfect:1 gaunt:2 birdsong:13 reinforcing:1 song:45 generally:1 dark:1 ten:1 processed:1 dotted:2 estimated:1 neuroscience:5 nottebohm:6 group:1 four:4 sum:2 idealization:1 inverse:1 doya:5 dorsolateral:1 abnormal:1 syllable:30 strength:2 marder:1 constraint:1 simulate:1 relatively:1 structured:1 volman:2 combination:2 membrane:5 remain:1 deafened:1 biologically:2 midbrain:2 memorizing:1 projecting:1 imitating:1 dlm:2 forebrain:6 mechanism:6 needed:2 avt:3 serf:1 end:2 adopted:1 junction:1 available:1 simulating:1 alternative:1 original:3 neuroanatomical:1 clustering:1 opportunity:1 scharff:3 perturb:2 hikaridai:1 magnocellular:1 tutor:9 already:1 squashed:1 primary:3 strategy:1 gradient:5 perturbs:2 deficit:3 separate:1 atr:1 lateral:1 thank:1 reason:1 difficult:3 neuron:4 observation:1 snapshot:1 howard:1 neostriatum:2 sing:3 compensation:1 immediate:1 neurobiology:2 prematurely:1 ucsd:1 perturbation:6 station:1 connection:4 acoustic:2 learned:4 established:2 assembled:1 adult:3 impair:1 suggested:1 usually:1 pattern:6 breathing:1 hear:1 genetically:1 program:1 memory:1 critical:1 suitable:1 scheme:1 started:1 mediated:1 coupled:1 review:1 expect:1 generation:6 proportional:1 remarkable:1 nucleus:16 consistent:1 tympani:1 viewpoint:1 bypass:1 critic:1 production:2 course:3 supported:1 side:1 institute:2 wide:1 template:13 focussed:1 face:1 arnold:1 sparse:1 feedback:2 dimension:2 curve:1 transition:2 vocalized:1 sensory:1 made:1 reinforcement:10 san:1 adaptive:4 replicated:1 preprocessing:1 social:1 correlate:2 transaction:1 investigating:1 assumed:5 spatio:2 neurology:1 search:3 norepinephrine:1 learn:1 channel:3 robust:1 ca:1 complex:4 uva:1 significance:1 linearly:1 arrow:2 profile:2 lesion:5 allowed:1 neuronal:1 tl:2 salk:1 slow:1 momentum:1 bandpass:1 candidate:1 third:1 learns:1 young:3 minute:1 specific:3 resembled:1 bronchus:5 evidence:1 essential:1 exists:1 sequential:4 importance:1 vries:2 margin:1 vigorous:1 explore:1 visual:1 hopelessly:1 lewicki:2 lewis:2 harris:2 ma:1 abbreviation:1 modulate:1 month:1 replace:1 man:1 change:1 specifically:1 determined:2 averaging:1 principal:1 called:1 specie:3 kuo:1 experimental:1 jong:1 principe:2 internal:3 doupe:7 modulated:2 tsai:1 evaluate:1 ex:3
7,122
889
Reinforcement Learning Methods for Continuous-Time Markov Decision Problems Steven J. Bradtke Computer Science Department University of Massachusetts Amherst, MA 01003 bradtkeGcs.umass.edu Michael O. Duff Computer Science Department University of Massachusetts Amherst, MA 01003 duffGcs.umass.edu Abstract Semi-Markov Decision Problems are continuous time generalizations of discrete time Markov Decision Problems. A number of reinforcement learning algorithms have been developed recently for the solution of Markov Decision Problems, based on the ideas of asynchronous dynamic programming and stochastic approximation. Among these are TD(,x), Q-Iearning, and Real-time Dynamic Programming. After reviewing semi-Markov Decision Problems and Bellman's optimality equation in that context, we propose algorithms similar to those named above, adapted to the solution of semi-Markov Decision Problems. We demonstrate these algorithms by applying them to the problem of determining the optimal control for a simple queueing system. We conclude with a discussion of circumstances under which these algorithms may be usefully applied. 1 Introduction A number of reinforcement learning algorithms based on the ideas of asynchronous dynamic programming and stochastic approximation have been developed recently for the solution of Markov Decision Problems. Among these are Sutton's TD(,x) [10], Watkins' Q-Iearning [12], and Real-time Dynamic Programming (RTDP) [1, 394 Steven Bradtke, Michael O. Duff 3]. These learning alogorithms are widely used, but their domain of application has been limited to processes modeled by discrete-time Markov Decision Problems (MDP's). This paper derives analogous algorithms for semi-Markov Decision Problems (SMDP's) - extending the domain of applicability to continuous time. This effort was originally motivated by the desire to apply reinforcement learning methods to problems of adaptive control of queueing systems, and to the problem of adaptive routing in computer networks in particular. We apply the new algorithms to the well-known problem of routing to two heterogeneous servers [7]. We conclude with a discussion of circumstances under which these algorithms may be usefully applied. 2 Semi-Markov Decision Problems A semi-Markov process is a continuous time dynamic system consisting of a countable state set, X, and a finite action set, A. Suppose that the system is originally observed to be in state z EX, and that action a E A is applied. A semi-Markov process [9] then evolves as follows: ? The next state, y, is chosen according to the transition probabilities Pz,(a) ? A reward rate p(z, a) is defined until the next transition occurs ? Conditional on the event that the next state is y, the time until the transition from z to y occurs has probability distribution Fz,(?Ja) One form of the SMDP is to find a policy the minimizes the expected infinite horizon discounted cost, the "value" for each state: e {IoOO e-.B t p(z(t), a(t?dt}, where z(t) and aCt) denote, respectively, the state and action at time t. For a fixed policy 71', the value of a given state z must satisfy v,..(z) L P ,(7I'(z? 10(00 10re-.B? p(z, 71'(z?dsdFz,(tJ7I'(z? + z ,E X L Pz,(7I'(Z? fooo e-.B t V,..(y)dFz,(tJ7I'(z?. (1) ,EX Defining R(z, y, a) = foOO fot e-.B? p(z, 71'(z?dsdFz, (tJ7I'(z?, the expected reward that will be received on transition from state z to state y on action a, and Reinforcement Learning Methods for Continuous-Time Markov Decision Problems 395 the expected discount factor to be applied to the value of state y on transition from state z on action a, it is clear that equation (1) is nearly identical to the value-function equation for discrete time Markov reward processes, Vw(z) = R(z, 1I"(z? + "Y I: Pzr (1I"(z?Vw (Y), (2) rEX where R(z, a) = :ErEx Pzr(a)R(z, y, a). If transition times are identically one for an SMDP, then a standard discrete-time MDP results. Similarly, while the value function associated with an optimal policy for an MDP satisfies the Bellman optimality equation I: Ve(z) = max {R(Z' a) + "Y pzr(a)v*(y)} , ilEA X rE (3) the optimal value function for an SMDP satisfies the following version of the Bellman optimality equation: V*(z) = max { ilEA I:X Pzr(a) 1 re-fJa p(z, a)dsdFzr(tJa) 0 10 00 + rE I: Pzr(a) loo e-fJtv*(y)dFzr(tJa)} . (4) rEX 3 Temporal Difference learning for SMDP's Sutton's TD(O) [10] is a stochastic approximation method for finding solutions to the system of equations (2). Having observed a transition from state z to state y with sample reward r(z, y, 1I"(z?, TD(O) updates the value function estimate V(A:)(z) in the direction of the sample value r(z, y, 1I"(z?+"YV(A:)(y). The TD(O) update rule for MDP's is V(A:+l)(Z) = V(A:)(z) + QA:[r(z, y, 1I"(z? + "YV(A:)(y) - V(A:)(z)], (5) where QA: is the learning rate. The sequence of value-function estimates generated by the TD(O) proceedure will converge to the true solution, Vw , with probability one [5,8, 11] under the appropriate conditions on the QA: and on the definition of the MDP. The TD(O) learning rule for SMDP's, intended to solve the system of equations (1) given a sequence of sampled state transitions, is: 1 -fJT ] V(A:+1)(z) = V(A:)(z) + QA: [ - ; r(z, y, 1I"(z? + e-fJTV(A:)(y) - V(A:)(z) , (6) where the sampled transition time from state z to state y was T time units, r(z, y, 1I"(z? is the sample reward received in T time units, and e- fJT is the sample discount on the value of the next state given a transition time of T time units. The TD(>.) learning rule for SMDP's is straightforward to define from here. I_p-tl. Steven Bradtke. Michael 0. Duff 396 4 Q-Iearning for SMDP's Denardo [6] and Watkins [12] define Q.f) the Q-function corresponding to the policy as (7) Q'II"(z, a) = R(z, a) + 'Y PzJ(a)V'II"(Y) 71", 2: YEX Notice that a can be any action. It is not necesarily the action 7I"(z) that would be chosen by policy 71". The function Q. corresponds to the optimal policy. Q'II"(z, a) represents the total discounted return that can be expected if any action is taken from state z, and policy 71" is followed thereafter. Equation (7) can be rewritten as Q'II"(z, a) = R(z, a) + 'Y 2: PZJ(a)Q'II"(Y' 7I"(Y?, (8) yEX and Q. satisfies the Bellman-style optimality equation Q?(z, a) = R(z, a) + 'Y 2: Pzy(a) max Q.(y, a'), JEX (9) A'EA Q-Iearning, first described by Watkins [12], uses stochastic approximation to iteratively refine an estimate for the function Q ?. The Q-Iearning rule is very similar to TD(O). Upon a sampled transition from state z to state y upon selection of a, with sampled reward r(z, y, a), the Q-function estimate is updated according to Q(A:+l)(Z, a) = Q(J:)(z, a) + etJ: [r(z, y, a) + 'Y ~~ Q(J:)(y, a') - Q(J:)(z, a)]. (10) Q-functions may also be defined for SMDP's. The optimal Q-function for an SMDP satisfies the equation Q?(z, a) 2: PZJ(a) roo t 10 10 'V JE"- e- tJ ? p(z, a)dsdFzJ(tla) + 2: Pz1I (a) roo e- tJt max Q.(y, a')dFzJ(tla). 10 'V JE"- (11) A'EA This leads to the following Q-Iearning rule for SMDP's: Q(A:+l)(Z, a) = Q(J:)(z, a)+etJ: [1 - ;-tJ'r' r(z, y, a) + e-tJ'r' ~~ Q(J:)(y, a') _ Q(J:)(z, a)] (12) 5 RTDP and Adaptive RTDP for SMDP's The TD(O) and Q-Iearning algorithms are model-free, and rely upon stochastic approximation for asymptotic convergence to the desired function (V'll" and Q., respectively). Convergence is typically rather slow. Real-Time Dynamic Programming (RTDP) and Adaptive RTDP [1,3] use a system model to speed convergence. Reinforcement Learning Methods for Continltolts-Time Markov Decision Problems 397 RTDP assumes that a system model is known a priori; Adaptive RTDP builds a model as it interacts with the system. As discussed by Barto et al. [1], these asynchronous DP algorithms can have computational advantages over traditional DP algorithms even when a system model is given. Inspecting equation (4), we see that the model needed by RTDP in the SMDP domain consists of three parts: 1. the state transition probabilities Pzy(a), 2. the expected reward on transition from state z to state y using action a, R(z, y, a), and 3. the expected discount factor to be applied to the value of the next state on transition from state z to state y using action a, 'Y(z, y, a). If the process dynamics are governed by a continuous time Markov chain, then the model needed by RTDP can be analytically derived through uniJormization [2]. In general, however, the model can be very difficult to analytically derive. In these cases Adaptive RTD P can be used to incrementally build a system model through direct interaction with the system. One version of the Adaptive RTDP algorithm for SMDP's is described in Figure 1. 1 2 3 4 Set k = 0, and set Zo to some start state. Initialize P, R, and ~. repeat forever { For all actions a, compute Q(Ie)(ZIe,a) = L P.."v(a) [ R(zIe,y,a) +~(zIe,y,a)V(Ie)(y) ] veX Perform the update V(le+l)(ZIe) = minoeA Q(Ie)(zIe,a) Select an action, ale. Perform ale and observe the transition to ZIe+l after T time units. Update P. Use the sample reward 1__;;11'" r(ZIe,Zle+l,ale) and the sample discount 5 6 7 factor e- f3T to update k=k+l 8 9 R and ~. } Figure 1: Adaptive RTDP for SMDP's. by Adaptive RTDP of P, R, and 'Y. P, il, and .y are the estimates maintained Notice that the action selection procedure (line 6) is left unspecified. Unlike RTDP, Adaptive RTDP can not always choose the greedy action. This is because it only has an e8timate of the system model on which to base its decisions, and the estimate could initially be quite inaccurate. Adaptive RTDP needs to explore, to choose actions that do not currently appear to be optimal, in order to ensure that the estimated model converges to the true model over time. 398 6 Steven Bradtke, Michael O. Duff Experiment: Routing to two heterogeneous servers Consider the queueing system shown in Figure 2. Arrivals are assumed to be Poisson with rate ).. Upon arrival, a customer must be routed to one of the two queues, whose servers have service times that are exponentially distributed with parameters J.l.1 and J.l.2 respectively. The goal is compute a policy that minimizes the objective function: e {foOO e-tJ t [c1n1(t) + C2n2(t)]dt}, where C1 and C2 are scalar cost factors, and n1(t) and n2(t) denote the number of customers in the respective queues at time t. The pair (n1(t), n2(t)) is the state of the system at time t; the state space for this problem is countably infinite. There are two actions available at every state: if an arrival occurs, route it to queue 1 or route it to queue 2. -.<-~ ___ -.J~ Figure 2: Routing to two queueing systems. It is known for this problem (and many like it [7]), that the optimal policy is a threshold policy; i.e., the set of states Sl for which it is optimal to route to the first queue is characterized by a monotonically nondecreasing threshold function F via Sl {(nl,n2)ln1 $ F(n2)}' For the case where C1 C2 1 and J.l.1 J.l.2, the policy is simply to join the shortest queue, and the theshold function is a line slicing diagnonally through the state space. = = = = We applied the SMDP version of Q-Iearning to this problem in an attempt to find the optimal policy for some subset of the state space. The system parameters were set to ). J.l.1 J.l.2 1, /3 0.1, and C1 C2 1. We used a feedforward neural network trained using backpropagation as a function approximator. = = = = = = Q-Iearning must take exploratory actions in order to adequately sample all of the available state transitions. At each decision time k, we selected the action aA: to be applied to state ZA: via the Boltzmann distribution where TA: is the "computational temperature." The temperature is initialized to a relatively high value, resulting in a uniform distribution for prospective actions. TA: is gradually lowered as computation proceeds, raising the probability of selecting actions with lower (and for this application, better) Q-values. In the limit, the action that is greedy with respect to the Q-function estimate is selected. The temperature and the learning rate erA: are decreased over time using a "search then converge" method [4]. Reinforcement Learning Methods for Continuous-Time Markov Decision Problems 399 Figure 3 shows the results obtained by Q-Iearning for this problem. Each square denotes a state visited, with nl(t) running along the z-axis, and n2(t) along the yaxis. The color of each square represents the probability of choosing action 1 (route arrivals to queue 1). Black represents probability 1, white represents probability o. An optimal policy would be black above the diagonal, white below the diagonal, and could have arbitrary colors along the diagonal. == == == == == = = ;;= == == == = = == ;;= == == == = = ? . ?? II II ?? ... !m il lUll it ? w? ? @@ ~d r2 E M m@ mi?1@ moo mllw ? ? ? ? ? ? ,.m @ll A ;;;;;; . .. ? ~2 III l1li 0 @oo @ %@Ii Ell =11111 m?? ? ?? IIIIJI III ?? ???????? @w liliiii w???????? ' @ m ? ? ? ? ? ? ? ? ? '. B II ?? mm oow mw m]lw mill lUll lM]lm ?? ? ? ? ? ? ? ? ? lIlIg!WI'm = wji n i?????????? II ...'.'........'. c Figure 3: Results of the Q-Iearning experiment. Panel A represents the policy after 50,000 total updates, Panel B represents the policy after 100,000 total updates, and Panel C represents the policy after 150,000 total updates. One unsatisfactory feature of the algorithm's performance is that convergence is rather slow, though the schedules governing the decrease of Boltzmann temperature TA: and learning rate 0A: involve design parameters whose tweakings may result in faster convergence. If it is known that the optimal policies are of theshold type, or that some other structural property holds, then it may be of extreme practical utility to make use of this fact by constraining the value-functions in some way or perhaps by representing them as a combination of appropriate basis vectors that implicity realize or enforce the given structural property. 7 Discussion In this paper we have proposed extending the applicability of well-known reinforcement learning methods developed for discrete-time MDP's to the continuous time domain. We derived semi-Markov versions of TD(O), Q-Iearning, RTDP, and Adaptive RTDP in a straightforward way from their discrete-time analogues. While we have not given any convergence proofs for these new algorithms, such proofs should not be difficult to obtain if we limit ourselves to problems with finite state spaces. (Proof of convergence for these new algorithms is complicated by the fact that, in general, the state spaces involved are infinite; convergence proofs for traditional reinforcement learning methods assume the state space is finite.) Ongoing work is directed toward applying these techniques to more complicated systems, examining distributed control issues, and investigating methods for incorporating prior 400 Steven Bradtke, Michael 0. Duff knowledge (such as structured function approximators). Acknowledgements Thanks to Professor Andrew Barto, Bob Crites, and to the members of the Adaptive Networks Laboratory. This work was supported by the National Science Foundation under Grant ECS-9214866 to Professor Barto. References [1] A. G. Barto, S. J. Bradtke, and S. P. Singh. Learning to act using real-time dynamic programming. Artificial Intelligence. Accepted. [2] D. P. Bertsekas. Dynamic Programming: Deterministic and Stochastic Models. Prentice Hall, Englewood Cliffs, NJ, 1987. [3] S. J. Bradtke. Incremental Dynamic Programming for On-line Adaptive Optimal Control. PhD thesis, University of Massachusetts, 1994. [4] C. Darken, J. Chang, and J. Moody. Learning rate schedules for faster stochastic gradient search. In Neural Networks for Signal Processing ~ - Proceedings of the 199~ IEEE Workshop. IEEE Press, 1992. [5] P. Dayan and T. J. Sejnowski. Td(A): Convergence with probability 1. Machine Learning, 1994. [6] E. V. Denardo. Contraction mappings in the theory underlying dynamic programming. SIAM Review, 9(2):165-177, April 1967. [7] B. Hajek. Optimal control of two interacting service stations. 29:491-499, 1984. IEEE-TAC, [8] T. Jaakkola, M. I. Jordan, and S. P. Singh. On the convergence of stochastic iterative dynamic programming algorithms. Neural Computation, 1994. [9] S. M. Ross. Applied Probability Models with Optimization Applications. HoldenDay, San Francisco, 1970. [10] R. S. Sutton. Learning to predict by the method of temporal differences. Machine Learning, 3:9-44, 1988. [11] J. N. Tsitsiklis. Asynchronous stochastic approximation and Q-Iearning. Technical Report LIDS-P-2172, Laboratory for Information and Decision Systems, MIT, Cambridge, MA, 1993. [12] C. J. C. H. Watkins. Learning from Delayed Rewards. PhD thesis, Cambridge University, Cambridge, England, 1989.
889 |@word version:4 contraction:1 uma:2 selecting:1 must:3 moo:1 realize:1 update:8 smdp:16 greedy:2 selected:2 intelligence:1 along:3 c2:3 direct:1 consists:1 expected:6 bellman:4 discounted:2 td:12 underlying:1 panel:3 unspecified:1 minimizes:2 rtdp:17 developed:3 finding:1 nj:1 temporal:2 every:1 act:2 usefully:2 iearning:13 control:5 unit:4 grant:1 appear:1 bertsekas:1 service:2 limit:2 sutton:3 era:1 cliff:1 black:2 limited:1 directed:1 practical:1 fot:1 backpropagation:1 procedure:1 selection:2 prentice:1 context:1 applying:2 deterministic:1 customer:2 straightforward:2 slicing:1 rule:5 exploratory:1 analogous:1 updated:1 suppose:1 programming:10 us:1 observed:2 steven:5 decrease:1 reward:9 dynamic:12 trained:1 singh:2 reviewing:1 upon:4 basis:1 zo:1 sejnowski:1 artificial:1 choosing:1 quite:1 whose:2 widely:1 solve:1 roo:2 nondecreasing:1 sequence:2 advantage:1 propose:1 interaction:1 convergence:10 etj:2 extending:2 incremental:1 converges:1 derive:1 oo:1 andrew:1 received:2 direction:1 stochastic:9 routing:4 ja:1 generalization:1 inspecting:1 tjt:1 mm:1 hold:1 hall:1 mapping:1 predict:1 lm:2 currently:1 visited:1 ross:1 proceedure:1 mit:1 always:1 rather:2 barto:4 jaakkola:1 derived:2 denardo:2 unsatisfactory:1 dayan:1 inaccurate:1 typically:1 initially:1 issue:1 among:2 priori:1 initialize:1 ell:1 having:1 identical:1 represents:7 nearly:1 report:1 ve:1 national:1 delayed:1 intended:1 consisting:1 ourselves:1 n1:2 attempt:1 englewood:1 extreme:1 nl:2 tj:4 chain:1 respective:1 initialized:1 re:4 desired:1 rtd:1 applicability:2 cost:2 subset:1 f3t:1 uniform:1 examining:1 rex:2 loo:1 thanks:1 amherst:2 siam:1 ie:3 michael:5 moody:1 thesis:2 choose:2 style:1 return:1 lull:2 satisfy:1 yv:2 start:1 complicated:2 il:2 square:2 bob:1 za:1 definition:1 involved:1 associated:1 mi:1 proof:4 sampled:4 massachusetts:3 color:2 knowledge:1 oow:1 schedule:2 hajek:1 ea:2 originally:2 dt:2 ta:3 april:1 zie:7 though:1 governing:1 until:2 incrementally:1 perhaps:1 mdp:6 iooo:1 fooo:3 true:2 adequately:1 analytically:2 iteratively:1 laboratory:2 white:2 ll:2 maintained:1 demonstrate:1 bradtke:7 temperature:4 recently:2 exponentially:1 discussed:1 l1li:1 cambridge:3 tac:1 similarly:1 lowered:1 base:1 route:4 server:3 approximators:1 wji:1 converge:2 shortest:1 monotonically:1 ale:3 semi:8 ii:10 signal:1 technical:1 faster:2 characterized:1 england:1 heterogeneous:2 circumstance:2 poisson:1 c1:3 decreased:1 unlike:1 member:1 jordan:1 structural:2 vw:3 mw:1 constraining:1 feedforward:1 identically:1 iii:2 idea:2 motivated:1 utility:1 effort:1 routed:1 queue:7 action:22 clear:1 involve:1 discount:4 fz:1 sl:2 notice:2 estimated:1 theshold:2 discrete:6 thereafter:1 threshold:2 queueing:4 named:1 decision:16 vex:1 followed:1 refine:1 adapted:1 speed:1 optimality:4 jex:1 relatively:1 department:2 structured:1 according:2 combination:1 wi:1 evolves:1 lid:1 gradually:1 tla:2 taken:1 equation:11 ln1:1 needed:2 available:2 rewritten:1 apply:2 observe:1 appropriate:2 enforce:1 assumes:1 denotes:1 ensure:1 running:1 build:2 objective:1 occurs:3 interacts:1 traditional:2 diagonal:3 mllw:1 gradient:1 dp:2 prospective:1 toward:1 modeled:1 difficult:2 design:1 countable:1 policy:17 boltzmann:2 fjt:2 perform:2 darken:1 markov:18 finite:3 defining:1 interacting:1 duff:5 station:1 arbitrary:1 pair:1 raising:1 qa:4 proceeds:1 below:1 max:4 analogue:1 event:1 rely:1 representing:1 axis:1 prior:1 review:1 acknowledgement:1 determining:1 asymptotic:1 approximator:1 foundation:1 repeat:1 supported:1 asynchronous:4 free:1 zle:1 pzy:2 tsitsiklis:1 distributed:2 transition:16 reinforcement:9 adaptive:14 san:1 ec:1 forever:1 countably:1 yaxis:1 investigating:1 conclude:2 assumed:1 francisco:1 continuous:8 search:2 iterative:1 tja:2 yex:2 domain:4 crites:1 arrival:4 n2:5 je:2 join:1 tl:1 slow:2 governed:1 watkins:4 lw:1 ilea:2 r2:1 pz:2 derives:1 incorporating:1 workshop:1 phd:2 horizon:1 mill:1 simply:1 explore:1 desire:1 scalar:1 chang:1 aa:1 corresponds:1 satisfies:4 ma:3 conditional:1 goal:1 professor:2 infinite:3 total:4 accepted:1 fja:1 select:1 ongoing:1 ex:2
7,123
89
127 Neural Network Implementation Approaches for the Connection Machine Nathan H. Brown, Jr. MRJlPerkin Elmer, 10467 White Granite Dr. (Suite 304), Oakton, Va. 22124 ABSlRACf The SIMD parallelism of the Connection Machine (eM) allows the construction of neural network simulations by the use of simple data and control structures. Two approaches are described which allow parallel computation of a model's nonlinear functions, parallel modification of a model's weights, and parallel propagation of a model's activation and error. Each approach also allows a model's interconnect structure to be physically dynamic. A Hopfield model is implemented with each approach at six sizes over the same number of CM processors to provide a performance comparison. INTRODUCflON Simulations of neural network models on digital computers perform various computations by applying linear or nonlinear functions, defined in a program, to weighted sums of integer or real numbers retrieved and stored by array reference. The numerical values are model dependent parameters like time averaged spiking frequency (activation), synaptic efficacy (weight), the error in error back propagation models, and computational temperature in thermodynamic models. The interconnect structure of a particular model is implied by indexing relationships between arrays defined in a program. On the Connection Machine (CM), these relationships are expressed in hardware processors interconnected by a 16-dimensional hypercube communication network. Mappings are constructed to defme higher dimensional interconnectivity between processors on top of the fundamental geometry of the communication network. Parallel transfers are defined over these mappings. These mappings may be dynamic. CM parallel operations transform array indexing from a temporal succession of references to memory to a single temporal reference to spatially distributed processors. Two alternative approaches to implementing neural network simulations on the CM are described. Both approaches use "data parallelism" 1 provided by the *Lisp virtual machine. Data and control structures associated with each approach and performance data for a Hopfield model implemented with each approach are presented. DATA STRUCTURES The functional components of a neural network model implemented in *Lisp are stored in a uniform parallel variable (pvar) data structure on the CM. The data structure may be viewed as columns of pvars. Columns are given to all CM virtual processors. Each CM physical processor may support 16 virtual processors. In the fust approach described, CM processors are used to represent the edge set of a models graph structure. In the second approach described, each processor can represent a unit, an outgoing link, or an incoming link in a model's structure. Movement of activation (or error) through a model's interconnect structure is simulated by moving numeric values ? American Institute of Physics 1988 128 over the eM's hypercube. Many such movements can result from the execution of a single CM macroinstruction. The CM transparently handles message buffering and collision resolution. However, some care is required on the part of the user to insure that message traffic is distributed over enough processors so that messages don't stack up at certain processors, forcing the CM to sequentially handle large numbers of buffered messages. Each approach requires serial transfers of model parameters and states over the communication channel between the host and the CM at certain times in a simulation. The first approach, "the edge list approach," distributes the edge list of a network graph to the eM, one edge per CM processor. Interconnect weights for each edge are stored in the memory of the processors. An array on the host machine stores the current activation for all units. This approach may be considered to represent abstract synapses on the eM. The interconnect structure of a model is described by product sets on an ordered pair of identification (id) numbers, rid and sid. The rid is the id of units receiving activation and sid the id of units sending activation. Each id is a unique integer. In a hierarchical network, the ids of input units are never in the set of rids and the ids of output units are never in the set of sids. Various set relations (e.g. inverse, reflexive, symmetric, etc.) defined over id ranges can be used as a high level representation of a network's interconnect structure. These relations can be translated into pvar columns. The limits to the interconnect complexity of a simulated model are the virtual processor memory limits of the CM configuration used and the stack space ~uired by functions used to compute the weighted sums of activation. Fig. 1 shows a R -> R2 -> R4 interconnect structure and its edge list representation on the CM. 6 7 8 :z eM PROCESSOR 0 1 2 3 4 9 3 5 6 7 8 9 1 0111213 f if :~ (~?,';)H f HHH ff SAcr ( 8j ) 1 2 3 1 2 3 4 5 4 5 4 5 4 5 Fig. 1. Edge List Representation of a R3_> R2 -> R4 Interconnect Structure This representation can use as few as six pvars for a model with Hebbian adaptation: rid (i), sid (j), interconnect weight (wij), ract (ai), sact (aj), and learn rate (11)? Error back propagation requires the addition of: error (ei), old interconnect weight (wij(t-l?, and the momentum term (ex). The receiver and sender unit identification pvars are described above. The interconnect weight pvar stores the weight for the interconnect. The activation pvar, sact, stores the current activation, aj' transfered to the unit specified by rid from the unit specified by sid. The activation pvar, ract, stores the current weighted activation ajwij- The error pvar stores the error for the unit specified by the sid. A variety of proclaims (e.g. integer, floating point, boolean, and field) exist in *Lisp to define the type and size ofpvars. Proclaims conserve memory and speed up execution. Using a small number of pvars limits the 129 amount of memory used in each CM processor so that maximum virtualization of the hardware processors can be realized. Any neural model can be specified in this fashion. Sigma-pi models require multiple input activation pvars be specified. Some edges may have a different number of input activation pvars than others. To maintain the uniform data structure of this approach a tag pvar has to be used to determine which input activation pvars are in use on a particular edge. The edge list approach allows the structure of a simulated model to "physically" change because edges may be added (up to the virtual processor limit), or deleted at any time without affecting the operation of the control structure. Edges may also be placed in any processor because the subselection (on rid or sid) operation performed before a particular update operation insures that all processors (edges) with the desired units are selected for the update. The second simulation approach, "the composite approach," uses a more complicated data structure where units, incoming links, and outgoing links are represented. Update routines for this approach use parallel segmented scans to form the weighted sum of input activation. Parallel segmented scans allow a MIMD like computation of the weighted sums for many units at once. Pvar columns have unique values for unit, incoming link, and outgoing link representations. The data structures for input units, hidden units, and output units are composed of sets of the three pvar column types. Fig. 2 shows the representation for the same model as in Fig. 1 implemented with the composite approach. 2 o1 3 5 4 6 7 8 9 2 3 4 5 6 7 8 9 101112 1314151617181920212223242526272829303132333435 rr, f~ ~\'~~Ii~ c - -?. c o c - -?. +----+ lol IO~ O.~~ ~ ~~+-t+*~ t -~. - ~ II I( II I( I Fig. 2. Composite Representation of a R3 -> R2 -> R4 Interconnect Structure In Fig. 2, CM processors acting as units, outgoing links, and incoming links are represented respectively by circles, triangles, and squares. CM cube address pointers used to direct the parallel transfer of activation are shown by arrows below the structure. These pointers defme the model interconnect mapping. Multiple sets of these pointers may be stored in seperate pvars. Segmented scans are represented by operation-arrow icons above the structure. A basic composite approach pvar set for a model with Hebbian adaptation is: forward B, forward A, forward transfer address, interconnect weight (Wij), act-l (ai), act-2 (aj), threshold, learn rate (Tl), current unit id (i), attached unit id U), level, and column type. Back progagation of error requires the addition of: backward B, backward A, backward transfer address, error (ei), previous interconnect weight (Wij(t-l?, and the momentum tenn (ex). The forward and backward boolean pvars control the segmented scanning operations over unit constructs. Pvar A of each type controls the plus scanning and pvar B of each type controls the copy scanning. The forward transfer pvar stores cube addresses for 130 forward (ascending cube address) parallel transfer of activation. The backward transfer pvar stores cube addresses for backward (descending cube address) parallel transfer of error. The interconnect weight, activation, and error pvars have the same functions as in the edge list approach. The current unit id stores the current unit's id number. The attached unit id stores the id number of an attached unit. This is the edge list of the network's graph. The contents of these pvars only have meaning in link pvar columns. The level pvar stores the level of a unit in a hierarchical network. The type pvar stores a unique arbitrary tag for the pvar column type. These last three pvars are used to subselect processor ranges to reduce the number of processors involved in an operation. Again, edges and units can be added or deleted. Processor memories for deleted units are zeroed out. A new structure can be placed in any unused processors. The level, column type, current unit id, and attached unit id values must be consistent with the desired model interconnectivity. The number of CM virtual processors required to represent a given model on the CM differs for each approach. Given N units and N(N-1) non-zero interconnects (e.g. a symmetric model), the edge list approach simply distributes N(N-1) edges to N(N-1) CM virtual processors. The composite approach requires two virtual processors for each interconnect and one virtual processor for each unit or N +2 N (N-1) CM virtual processors total. The difference between the number of processors required by the two approaches is N2. Table I shows the processor and CM virtualization requirements for each approach over a range of model sizes. TABLE I Model Sizes and CM Processors Required Run No. Grid Size Number of Units Edge List Quart CM Virt. Procs. Virt. LeveL N(N-1) 1 2 3 4 5 6 82 92 112 13 2 162 192 64 81 121 169 256 361 4032 6480 14520 28392 65280 129960 8192 8192 16384 32768 65536 131072 0 0 0 2 4 8 Run No. Grid Size Number of Units Composite Quart CM Virt. Procs. Virt. LeveL N+2N(N-1) 7 8 9 10 11 12 82 92 112 132 162 192 64 81 121 169 256 361 8128 13041 29161 56953 130816 260281 8192 16384 32768 65536 131072 262144 0 0 2 4 8 16 131 CONTROL STRUCTURES The control code for neural network simulations (in *Lisp or C*) is stored and executed sequentially on a host computer (e.g. Symbolics 36xx and V AX 86xx) connected to the CM by a high speed communication line. Neural network simulations executed in *Lisp use a small subset of the total instruction set: processor selection reset (*all), processor selection (*when), parallel content assignment (*set), global summation (*sum), parallel multiplication (*!! ), parallel summation (+! I), parallel exponentiation (exp! I), the parallel global memory references (*pset) and (pref! I), and the parallel segmented scans (copy!! and +!!). Selecting CM processors puts them in a "list of active processors" (loap) where their contents may be arithmetically manipulated in parallel. Copies of the list of active processors may be made and used at any time. A subset of the processors in the loap may be "subselected" at any time, reducing the loap contents. The processor selection reset clears the current selected set by setting all processors as selected. Parallel content assignment allows pvars in the currently selected processor set to be assinged allowed values in one step. Global summation executes a tree reduction sum across the CM processors by grid or cube address for particular pvars. Parallel multiplications and additions multiply and add pvars for all selected CM processors in one step. The parallel exponential applies the function, eX, to the contents of a specified pvar, x, over all selected processors. Parallel segmented scans apply two functions, copy!! and +!!, to subsets ofCM processors by scanning across grid or cube addresses. Scanning may be forward or backward (Le. by ascending or descending cube address order, respectively). Figs. 3 and 4 show the edge list approach kernels required for Hebbian learning for a R2 -> R2 model. The loop construct in Fig. 3 drives the activation update (1) operation. The usual loop to compute each weighted sum for a particular unit has been replaced by four parallel operations: a selection reset (*all), a subselection of all the processors for which the particular unit is a receiver of activation (*when (=!! rid (!! (1+ u??, a parallel multiplication (*!! weight sact), and a tree reduction sum (*sum ... ). Activation is spread for a particular unit, to all others it is connected to, by: storing the newly computed activation in an array on the host, then subselecting the processors where the particular unit is a sender of activation (*when (=!! sid (!! (1 + u??, and broadcasting the array value on the host to those processors. (dotimes (u 4) (*all (*when (=!! rid (!! (1+ u?) (setf (aref activation u) (some-nonlinearity (*sum (*!! weight sact?? (*set ract (!! (aref activation u?) (*all (*when (=!! sid (!! (1+ u?) (*set sact (!! (aref activation u??? Fig. 3. Activation Update Kernel for the Edge Lst Approach. Fig. 4 shows the Hebbian weight update kernel 132 (2) (*all (*set weight (*!! learn-rate ract sact?? Fig. 4. Hebbian Weight Modification Kernel for the Edge List Approach The edge list activation update kernel is essentially serial because the steps involved can only be applied to one unit at a time. The weight modification is parallel. For error back propagation a seperate loop for computing the errors for the units on each layer of a model is required. Activation update and error back propagation also require transfers to and from arrays on the host on every iteration step incurring a concomitant overhead. Other common computations used for neural networks can be computed in parallel using the edge list approach. Fig. 5 shows the code kernel for parallel computation of Lyapunov engergy equations (3) where i= 1 to number of units (N). (+ (* -.5 (*sum (*!! weight ract sact?) (*sum (*!! input sact?) Fig. 5. Kernel for Computation of the Lyapunov Energy Equation Although an input pvar, input, is defined for all edges, it is only non-zero for those edges associated with input units. Fig. 6 shows the pvar structure for parallel computation of a Hopfield weight prescription, with segmented scanning to produce the weights in one step, IJ -l:S r= I(2ar1?-I)(2arJ?-I) W? ? (4) where wii=O, Wij=Wjh and r=I to the number of patterns, S, to be stored. seg t n ract vII V2 1 ... V I 2 v22' .. sact weight n t n n VSI vII V2 I ... VSI .. . VS2 v13 v23 ... VS3 .. . wI2 w13 Fig. 6. Pvar Structure for Parallel Computation QfHopfield Weight Prescription Fig. 7 shows the *Lisp kernel used on the pvar structure in Fig. 6. (set weight (scan '+!! (*!! (-!! (*!! ract (!! 2? (!! 1? (-!! (*!! sact (!! 2? (!! 1?? :segment-pvar seg :inc1ude-self t) Fig. 7. Parallel Computation of Hopfield Weight Prescription 133 The inefficiencies of the edge list activation update are solved by the updating method used in the composite approach. Fig. 8 shows the *Lisp kernel for activation update using the composite approach. Fig. 9 shows the *Lisp kernel for the Hebbian learning operation in the composite approach. (*a1l (*when (=!! level (!! 1? (*set act (scan!! act-I 'copy!! :segment-pvar forwardb :include-self t? (*set act (*!! act-l weight? (*when (=!! type (!! 2? (*pset :overwrite act-l act-I ftransfer?) (*when (=!! level (!! 2? (*set act (scan!! act-l '+!! :segment-pvar forwarda :include-self t? (*when (=!! type (!! 1? (some-nonlinearity!! act-I?? Fig. 8. Activation Update Kernel for the Composite Approach (*all (*set act-l (scan!! act-I 'copy!! :segment-pvar forwardb :include-self t? (*when (=!! type (!! 2? (*set act-2 (pref!! act-I btransfer?) (*set weight (+!! weight (*!! learn-rate act-l act-2??) Fig. 9. Hebbian Weight Update Kernel for the Composite Approach It is immediately obvious that no looping is invloved. Any number of interconnects may be updated by the proper subselection. However, the more subselection is used the less efficient the computation becomes because less processors are invloved. COMPLEXITY ANALYSIS The performance results presented in the next section can be largely anticipated from an analysis of the space and time requirements of the CM implementation approaches. For simplicity I use a Rn -> Rn model with Hebbian adaptation. The oder of magnitude requirements for activation and weight updating are compared for both CM implementation approaches and a basic serial matrix arithmetic approach. F~r the given model the space requirements on a conventional serial machine are 2n+n locations or O(n 2). The growth of the space requirement is dominated by the nxn weight matrix. defining the system interconnect structure. The edge list appro~ch uses six pvars for each processor and uses nxn processors for the mapping, or 6n locations or O(n2). The composite approach uses 11 pvars. There are 2n processors for units and 2n2 proces~ors for interconnects in the given model. The composite approach uses 11(2n+2n ) locations or O(n2 ). The CM implementations take up roughly the same space as the serial implementation, but the space for the serial implementation is composed of passive memory whereas the space for the CM implementations is composed of interconnected processors with memory . The time analysis for the approaches compares the time order of magnitudes to compute the activation update (1) and the Hebbian weight update (2). On a serial 134 machine, the n weighted sums computed for the ac~vation update require n2 multiplicationsffd n(n-l) additions. There are 2n -n operations or time order of magnitude O(n ~ The time order of magnitude for the weight matrix update is O(n2) since there are n weight matrix elements. The edge list approach forms n weighted sums by performing a parallel product of all of the weights and activations in the model, (*!! weight sact), and then a tree reduction sum, (*sum ... ), of the products for the n uni~ (see Fig. 4). There are 1+n(nlog2n) operations or time order of magnitude O(n ). This is the same order of magnitude as obtained on a serial machine. Further, the performance of the activation update is a function of the number of interconnects to be processed. The composite approach forms n weighted sums in nine steps (see Fig. 8): five .selection operations; the segmented copy scan before the parallel multiplication; the parallel multiplication; the parallel transfer of the products; and the segmented plus scan, which forms the n sums in one step. This gives the composite activation update a time order of magnitude O( 1). Performance is independent of the number of interconnects processed. The next section shows that this is not quite true. The n2 weights in the model can be updated in three parallel steps using the edge list approach (see Fig. 4). The n2 weights in the model can be updated in eight parallel steps using the composite approach (see Fig. 9). In either case, the weight update operation has a time order of magnitude 0(1). The time complexity results obtained for the composite approach apply to computation of the Lyaponov energy equation (3) and the Hopfield weighting prescription (4), given that pvar structures which can be scanned (see Figs. 1 and 6) are used. The same operations performed serially are time order of magnitude 0(n2). The above operations all incur a one time overhead cost for generating the addresses in the pointer pvars, used for parallel transfers, and arranging the values in segments for scanning. What the above analysis shows is that time complexity is traded for space complexity. The goal of CM programming is to use as many processors as possible at every step. PERFORMANCE COMPARISON Simulations of a Hopfield spin-glass model2 were run for six different model sizes over the same number (16,384) of physical CM processors to provide a performance comparison between implementation approaches. The Hopfield network was chosen for the performance comparison because of its simple and well known convergence dynamics and because it uses a small set of pvars which allows a wide range of network sizes (degrees of virtualization) to be run. Twelve treaments are run. Six with the edge list approach and six with the composite approach. Table 3-1 shows the model sizes run for each treatment. Each treatment was run at the virtualization level just necessary to accomodate the number of processors required for each simulation. Two exemplar patterns are stored. Five test patterns are matched against the two exemplars. Two test patterns have their centers removed, two have a row and column removed, and one is a random pattern. Each exemplar was hand picked and tested to insure that it did not produce cross-talk. The number of rows and columns in the exemplars and patterns increase as the size of the networks for the treatments increases. 135 Since the performance of the CM is at issue, rather than the performance of the network model used, a simple model and a simple pattern set were chosen to minimize consideration of the influence of model dynamics on performance. Performance is presented by plotting execution speed versus model size. Size is measured by the number of interconnects in a model. The execution speed metric is interconnects updated per second, N*(N-l )/t, where N is the number of units in a model and t is the time used to update the activations for all of the units in a model. All of the units were updated three times for each pattern. Convergence was determined by the output activation remaining stable over the fmal two updates. The value of t for a treatment is the average of 15 samples of t. Fig. 10 shows the activation update cycle time for both approaches. Fig. 11 shows the interconnect update speed plots for both approaches. The edge list approach is plotted in black. The composite approach is plotted in white. The performance shown excludes overhead for interpretation of the *Lisp instructions. The model size categories for each plot correspond to the model sizes and levels of eM virtualization shown in Table I. Activation Update Cycle Time vs Model Size 1 .6 1 .4 1.2 ? sees O.B 0.6 0.4 ? ? o 0.2 ? OO___~~~__~O~__~O~__.O__~ 1 2 3 4 5 6 Model Size Fig. 10. Activation Update Cycle Times Interconnect Update Speed Comparison Edge Ust Approach vs. Composite Approach 2000000} 0 1500000 0 0 i.p.s. 1000000 0 0 0 500000t o?1 ? 2 ? ? 3 4 Model Size ? 5 ? 6 Fig. 11. Edge List Interconnect Update Speeds Fig. 11 shows an order of magnitude performance difference between the approaches and a roll off in performance for each approach as a function of the number of virtual processors supported by each physical processor. The performance tum around is at 4x virtualization for the edge list approach and 2x virtualization for the composite approach. 136 CONCLUSIONS Representing the interconnect structure of neural network models with mappings defined over the set of fine grain processors provided by the CM architecture provides good performance for a modest programming effort utilizing only a small subset of the instructions provided by *Lisp. Further, the perfonnance will continue to scale up linearly as long as not more than 2x virtualization is required. While the complexity analysis of the composite activation update suggests that its performance should be independent of the number of interconnects to be processed, the perfonnance results show that the performance is indirectly dependent on the number of interconnects to be processed because the level of virtualization required (after the physical processors are exhausted) is dependent on the number of interconnects to be processed and virtualization decreases performance linearly. The complexity analysis of the edge list activation update shows that its perfonnance should be roughly the same as serial implementations on comparable machines. The results suggest that the composite approach is to be prefered over the edge list approach but not be used at a virtualization level higher than 2x. The mechanism of the composite activation update suggest that hierarchical networks simulated in this fashion will compare in perfonnance to single layer networks because the parallel transfers provide a type of pipeline for activation for synchronously updated hierarchical networks while providing simultaneous activation transfers for asynchronously updated single layer networks. Researchers at Thinking Machines Corporation and the M.I.T. AI Laboratory in Cambridge Mass. use a similar approach for an implementation of NETtalk. Their approach overlaps the weights of connected units and simultaneously pipelines activation forward and error backward. 3 Perfonnance better than that presented can be gained by translation of the control code from interpreted *Lisp to PARIS and use of the CM2. In addition to not being interpreted, PARIS allows explicit control over important registers that aren't accessable through *Lisp. The CM2 will offer a number of new features which will enhance perfonnance of neural network simulations: a *Lisp compiler, larger processor memory (64K), and floating point processors. The complier and floating point processors will increase execution speeds while the larger processor memories will provide a larger number of virtual processors at the performance tum around points allowing higher perfonnance through higher CM utilization. REFERENCES 1. "Introduction to Data Level Parallelism," Thinking Machines Technical Report 86.14, (April 1986). 2. Hopfield, J. J., "Neural networks and physical systems with emergent collective computational abilities," Proc. Natl. Acad. Sci., Vol. 79, (April 1982), pp. 2554-2558. 3. Blelloch, G. and Rosenberg, C. Network Learning on the Connection Machine, M.I.T. Technical Report, 1987.
89 |@word instruction:3 cm2:2 simulation:10 pset:2 reduction:3 configuration:1 inefficiency:1 efficacy:1 selecting:1 current:8 activation:48 must:1 ust:1 grain:1 numerical:1 plot:2 update:30 v:2 tenn:1 selected:6 pointer:4 provides:1 location:3 five:2 constructed:1 direct:1 overhead:3 roughly:2 becomes:1 provided:3 xx:2 insure:2 matched:1 mass:1 what:1 cm:39 interpreted:2 corporation:1 suite:1 temporal:2 every:2 act:17 growth:1 control:10 unit:47 utilization:1 before:2 limit:4 io:1 acad:1 id:15 black:1 plus:2 r4:3 suggests:1 range:4 averaged:1 unique:3 differs:1 composite:24 suggest:2 uired:1 selection:5 put:1 applying:1 influence:1 descending:2 conventional:1 center:1 resolution:1 simplicity:1 immediately:1 array:7 utilizing:1 handle:2 overwrite:1 arranging:1 updated:7 construction:1 user:1 programming:2 us:6 element:1 conserve:1 updating:2 virt:4 solved:1 seg:2 v13:1 connected:3 cycle:3 movement:2 removed:2 decrease:1 prefered:1 complexity:7 dynamic:4 segment:5 incur:1 triangle:1 translated:1 model2:1 hopfield:8 lst:1 v22:1 various:2 represented:3 talk:1 emergent:1 seperate:2 pref:2 quite:1 larger:3 ability:1 transform:1 asynchronously:1 rr:1 interconnected:2 product:4 reset:3 adaptation:3 loop:3 convergence:2 requirement:5 produce:2 generating:1 ac:1 measured:1 exemplar:4 ij:1 implemented:4 lyapunov:2 virtual:12 implementing:1 a1l:1 require:3 blelloch:1 summation:3 around:2 considered:1 exp:1 mapping:6 traded:1 vsi:2 proc:1 currently:1 interconnectivity:2 weighted:9 rather:1 appro:1 rosenberg:1 ax:1 glass:1 dependent:3 interconnect:24 fust:1 hidden:1 relation:2 wij:5 issue:1 cube:8 field:1 simd:1 never:2 once:1 construct:2 procs:2 buffering:1 anticipated:1 thinking:2 others:2 report:2 few:1 composed:3 manipulated:1 simultaneously:1 floating:3 replaced:1 geometry:1 maintain:1 message:4 multiply:1 natl:1 edge:37 necessary:1 modest:1 perfonnance:7 tree:3 old:1 desired:2 circle:1 plotted:2 progagation:1 column:11 boolean:2 assignment:2 cost:1 reflexive:1 subset:4 uniform:2 stored:7 vation:1 scanning:7 wjh:1 fundamental:1 twelve:1 physic:1 receiving:1 off:1 enhance:1 transfered:1 again:1 dr:1 american:1 register:1 performed:2 picked:1 vs2:1 traffic:1 compiler:1 parallel:38 complicated:1 minimize:1 square:1 spin:1 roll:1 ract:7 largely:1 succession:1 correspond:1 identification:2 sid:8 drive:1 icon:1 researcher:1 processor:66 executes:1 simultaneous:1 synapsis:1 synaptic:1 vs3:1 against:1 energy:2 proces:1 frequency:1 involved:2 pp:1 obvious:1 associated:2 newly:1 treatment:4 routine:1 back:5 tum:2 higher:4 arj:1 april:2 fmal:1 just:1 accessable:1 hand:1 ei:2 nonlinear:2 propagation:5 aj:3 brown:1 true:1 spatially:1 symmetric:2 laboratory:1 white:2 nettalk:1 self:4 temperature:1 passive:1 meaning:1 consideration:1 common:1 functional:1 spiking:1 physical:5 attached:4 virtualization:11 interpretation:1 buffered:1 cambridge:1 ai:3 grid:4 nonlinearity:2 lol:1 moving:1 stable:1 etc:1 add:1 retrieved:1 forcing:1 store:11 certain:2 continue:1 care:1 determine:1 arithmetic:1 ii:3 thermodynamic:1 multiple:2 hebbian:9 segmented:9 technical:2 cross:1 w13:1 offer:1 long:1 prescription:4 host:6 serial:9 va:1 basic:2 essentially:1 metric:1 physically:2 iteration:1 represent:4 kernel:12 addition:5 affecting:1 whereas:1 fine:1 lisp:13 integer:3 unused:1 enough:1 variety:1 architecture:1 reduce:1 six:6 effort:1 nine:1 oder:1 collision:1 clear:1 quart:2 amount:1 subselection:4 hardware:2 processed:5 category:1 exist:1 transparently:1 per:2 vol:1 four:1 threshold:1 ar1:1 deleted:3 backward:8 graph:3 excludes:1 sum:18 run:7 inverse:1 exponentiation:1 comparable:1 layer:3 scanned:1 looping:1 tag:2 dominated:1 nathan:1 speed:8 performing:1 jr:1 across:2 em:6 modification:3 indexing:2 pipeline:2 equation:3 r3:1 mechanism:1 ascending:2 sending:1 granite:1 operation:16 incurring:1 wii:1 apply:2 eight:1 hierarchical:4 v2:2 indirectly:1 alternative:1 top:1 symbolics:1 include:3 remaining:1 hypercube:2 implied:1 added:2 realized:1 usual:1 link:9 simulated:4 sci:1 code:3 o1:1 relationship:2 mimd:1 concomitant:1 providing:1 executed:2 sigma:1 implementation:10 proper:1 collective:1 perform:1 allowing:1 defining:1 communication:4 rn:2 synchronously:1 stack:2 arbitrary:1 defme:2 required:9 pair:1 specified:6 connection:4 paris:2 address:11 parallelism:3 below:1 pattern:8 wi2:1 program:2 memory:11 overlap:1 serially:1 representing:1 interconnects:10 multiplication:5 nxn:2 versus:1 digital:1 degree:1 consistent:1 zeroed:1 plotting:1 storing:1 pi:1 translation:1 row:2 placed:2 last:1 copy:7 supported:1 allow:2 institute:1 wide:1 distributed:2 numeric:1 hhh:1 forward:8 made:1 uni:1 global:3 sequentially:2 incoming:4 active:2 rid:7 receiver:2 don:1 invloved:2 table:4 channel:1 transfer:14 learn:4 did:1 spread:1 linearly:2 arrow:2 n2:9 allowed:1 pvar:28 fig:32 tl:1 ff:1 fashion:2 momentum:2 explicit:1 exponential:1 weighting:1 list:25 r2:5 gained:1 magnitude:10 execution:5 accomodate:1 exhausted:1 vii:2 aren:1 broadcasting:1 simply:1 insures:1 sender:2 expressed:1 ordered:1 applies:1 ch:1 viewed:1 goal:1 content:6 change:1 determined:1 reducing:1 acting:1 distributes:2 total:2 support:1 scan:11 outgoing:4 tested:1 ex:3
7,124
890
A Model of the Neural Basis of the Rat's Sense of Direction William E. Skaggs James J. Knierim Hemant S. Kudrimoti bill@nsma.arizona. edu jim@nsma.arizona. edu hemant@nsma. arizona. edu Bruce L. McNaughton bruce@nsma. arizona. edu ARL Division of Neural Systems, Memory, And Aging 344 Life Sciences North, University of Arizona, 'IUcson AZ 85724 Abstract In the last decade the outlines of the neural structures subserving the sense of direction have begun to emerge. Several investigations have shed light on the effects of vestibular input and visual input on the head direction representation. In this paper, a model is formulated of the neural mechanisms underlying the head direction system. The model is built out of simple ingredients, depending on nothing more complicated than connectional specificity, attractor dynamics, Hebbian learning, and sigmoidal nonlinearities, but it behaves in a sophisticated way and is consistent with most of the observed properties ofreal head direction cells. In addition it makes a number of predictions that ought to be testable by reasonably straightforward experiments. 1 Head Direction Cells in the Rat There is quite a bit of behavioral evidence for an intrinsic sense of direction in many species of mammals, including rats and humans (e.g., Gallistel, 1990). The first specific information regarding the neural basis of this "sense" came with the discovery by Ranck (1984) of a population of "head direction" cells in the dorsal presubiculum (also known as the "postsubiculum") of the rat. A head direction cell 174 William Skaggs, James J. Knierim, Hemant S. Kudrimoti, Bruce L. McNaughton fires at a high rate if and only if the rat's head is oriented in a specific direction. Many things could potentially cause a cell to fire in a head-direction dependent manner: what made the postsubicular cells particularly interesting was that when their directionality was tested with the rat at different locations, the head directions corresponding to maximal firing were consistently parallel, within the experimental resolution. This is difficult to explain with a simple sensory-based mechanism; it implies something more sophisticated. 1 The postsubicular head direction cells were studied in depth by Taube et al. (1990a,b), and, more recently, head direction cells have also been found in other parts of the rat brain, in particular the anterior nuclei of the thalamus (Mizumori and Williams, 1993) and the retrosplenial (posterior cingulate) cortex (Chen et al., 1994a,b). Interestingly, all of these areas are intimately associated with the hippocampal formation, which in the rat contains large numbers of "place" cells. Thus, the brain contains separate but neighboring populations of cells coding for location and cells coding for direction, which taken together represent much of the information needed for navigation. Figure 1 shows directional tuning curves for three typical head direction cells from the anterior thalamus. In each of them the breadth of tuning is on the order of 90 degrees. This value is also typical for head direction cells in the postsubiculum and retrosplenial cortex, though in each of the three areas individual cells may show considerable variability. Figure 1: Polar plots of directional tuning (mean firing rate as a function of head direction) for three typical head direction cells from the anterior thalamus of a rat. Every study to date has indicated that the head direction cells constitute a unitary system, together with the place cells of the hippocampus . Whenever two head direction cells have been recorded simultaneously, any manipulation that caused one of them to shift its directional alignment caused the other to shift by the same amount; and when head direction cells have been recorded simultaneously with place cells , any manipulation that caused the head direction cells to realign either caused the hippocampal place fields to rotate correspondingly or to "remap" into a different pattern (Knierim et al., 1995). Head direction cells maintain their directional tuning for some time when the lights in the recording room are turned off, leaving an animal in complete darkness; the directionality tends to gradually drift, though, especially if the animal moves around (Mizumori and Williams, 1993). Directional tuning is preserved to some degree even 1 Sensitivity to the Earth's geomagnetic field has been ruled out as an explanation of head-directional firing . A Model of the Neural Basis of the Rat's Sense of Direction 175 if an animal is passively rotated in the dark, which indicates strongly that the head direction system receives information (possibly indirect) from the vestibular system. Visual input influences but does not dictate the behavior of head direction cells. The nature of this influence is quite interesting. In a recent series of experiments (Knierim et al., 1995), rats were trained to forage for food pellets in a gray cylinder with a single salient directional cue, a white card covering 90 degrees of the wall. During training, half of the rats were disoriented before being placed in the cylinder, in order to disrupt the relation between their internal sense of direction and the location of the cue card; the other half of the rats were not disoriented. Presumably, the rats that were not disoriented during training experienced the same initial relationship between their internal direction sense and the CUe card each time they were placed in the cylinder; this would not have been true of the disoriented rats. Head direction cells in the thalamus were subsequently recorded from both groups of rats as they moved in the cylinder. All rats were disoriented before each recording session. Under these conditions, the cue card had much weaker control over the head direction cells in the rats that had been disoriented during training than in the rats that had not been disoriented . For all rats the influence of the cue card upon the head direction system weakened gradually over the course of multiple recording sessions, and eventually they broke free, but this happened much sooner in the rats that had been disoriented during training. The authors concluded that a visual cue could only develop a strong influence upon the head direction system if the rat experienced it as stable. Figure 2 illustrates the shifts in alignment during a typical recording session. When the rat is initially placed in the cylinder, the cell's tuning curve is aligned to the west. Over the first few minutes of recording it gradually rotates to SSW, and there it stays. Note the "tail" of the curve. This comes from spikes belonging to another, neighboring head direction cell, which could not be perfectly isolated from the first. Note that, even though they come from different cells, both portions shift alignment synchronously. Figure 2: Shifts in alignment of a head direction cell over the course of a single recording session (one minute intervals). 2 The Model As reviewed above, the most important facts to be accounted for by any model of the head direction system are (1) the shape of the tuning curves for head direction cells, (2) the control of head direction cells by vestibular input, and (3) the stabilitydependent influence of visual cues on head direction cells. We introduce here a 176 William Skaggs, James J. Knierim, Hemant S. Kudrimoti, Bruce L McNaughton model that accounts for these facts. It is a refinement of a model proposed earlier by McNaughton et al. (1991), the main addition being a more specific account of neural connections and dynamics. The aim of this effort is to develop the simplest possible architecture consistent with the available data. The reality is sure to be more complicated than this model. Figure 3 schematically illustrates the architecture of the model. There are four groups of cells in the model: head direction cells, rotation cells (left and right), vestibular cells (left and right), and visual feature detectors. For expository purposes it is helpful to think of the network as a set of circular layers; this does not reflect the anatomical organization of the corresponding cells in the brain. V.oIIbul .. col (rtght) 00 ??00 0 ??0 0 ?0000 @ @ @ @ 00 @ ? @ ~ROt.lOftC~I~etQ 66 ?00---- Rotadon cell (rtght) ?0 0?0 0?0 ? 0 00 ? ? ? ? ? 00 0 ? ? 00000 ? @ @ @ ? ~H~ @ @ @ @ eM.ocUon coM Figure 3: Architecture of the head direction cell model. The head direction cell group has intrinsic connections that are stronger than any other connections in the model, and dominate their dynamics, so that other inputs only provide relatively small perturbations. The connections between them are set up so that the only possible stable state of the system is a single localized cluster of active cells, with all other cells virtually silent. This will occur if there are strong excitatory connections between neighboring cells, and strong inhibitory connections between distant cells. It is assumed that the network of interconnections has rotation and reflection symmetry. Small deviations from symmetry will not impair the model too much; large deviations may cause it to have strong attractors at a few points on the circle, which would cause problems. The crucial property of this network is the following . Suppose it is in a stable state, with a single cluster of activated cells at one point on the circle, and suppose an external input is applied that excites the cells selectively on one side (left or right) A Model of the Neural Basis of the Rat' s Sense of Direction 177 of the peak. Then the peak will rotate toward the side at which the input is applied, and the rate of rotation will increase with the strength of the input. This feature is exploited by the mechanisms for vestibular and visual control of the system. The vestibular mechanism operates via a layer of "rotation" cells, corresponding to the circle of head direction cells (Units with a similar role were referred to as "H x H'" cells in the McNaughton et al. (1991) model). There are two groups of rotation cells, for left and right rotations. Each rotation cell receives excitatory input from the head direction cell at the same point on the circle, and from the vestibular system. The activation function of the rotation cell is sigmoidal or threshold linear, so that the cell does not become active unless it receives input simultaneously from both sources. Each right rotation cell sends excitatory projections to head direction cells neighboring it on the right, but not to those that neighbor it on the left, and contrariwise for left rotation cells. It is easy to see how the mechanism works. When the animal turns to the right, the right vestibular cells are activated, and then the right rotation cells at the current peak of the head direction system are activated. These add to the excitation of the head direction cells to the right of the peak, thereby causing the peak to shift rightward . This in turn causes a new set of rotation cells to become active (and the old ones inactive), and thence a further shift of the peak, and so on. The peak will continue to move around the circle as long as the vestibular input is active, and the stronger the vestibular input, the more rapidly the peak will move. If the gain of this mechanism is correct (but weak compared to the gain of the intrinsic connections of the head direction cells), then the peak will move around the circle at the same rate that the animal turns, and the location of the peak will function as an allocentric compass. This can only be expected to work over a limited range of turning rates, but the firing rates of cells in the vestibular nuclei are linearly proportional to angular velocity over a surprisingly broad range, so there is no reason why the mechanism cannot perform adequately. Of course the mechanism is intrinsically error-prone, and without some sort of external correction, deviations are sure to build up over time. But this is an inevitable feature of any plausible model, and in any case does not conflict with the available data, which, while sketchy, suggests that passive rotation of animals in the dark can cause quite erratic behavior in head direction cells (E. J. Markus, J. J . Knierim, unpublished observations). The final ingredient of the model is a set of visual feature detectors, each of which responds if and only if a particular visual feature is located at a particular angle with respect to the axis of the rat's head. Thus, these cells are feature specific and direction specific, but direction specific in the head-centered frame, not in the world frame . It is assumed that each visual feature detector projects weakly to all of the head direction cells, and that these connections are modifiable according to a Hebbian rule, specifically, ~W = a(Wmaxt(Aposd - W)Apre, where W is the connection weight, W max is its maximum possible value, Apost is the firing rate of the postsynaptic cell, Apre is the firing rate of the presynaptic cell, and the function has the shape shown in figure 4. (Actually, the rule is modified slightly to prevent any of the weights from becoming negative.) The net effect of to 178 William Skaggs, James J. Knierim, Hemant S. Kudrimoti, Bruce L. McNaughton this rule is that the weight will only change when the presynaptic cell (the visual feature detector) is active, and the weight will increase if the postsynaptic cell is strongly active, but decrease ifit is weakly active or silent. Modification rules of this form have previously been proposed in theories of the development of topography in the neocortex (e.g., Bienenstock et al., 1982), and there is considerable evidence for such an effect in the control of LTP /LTD (Singer and Artola, 1994) . 1(rate) rate Figure 4: Dependence of synaptic weight change on postsynaptic firing rate for connections from visual feature detectors to head direction cells in the model. To understand how this works, suppose we have a feature detecting cell that responds to a cue card whenever the cue card is directly in front of the rat. Suppose the rat's motion is restricted to a small area, and the cue card is far away, so that it is always at approximately the same absolute bearing (say, 30 degrees), and suppose the rat's head direction system is working correctly, i.e., functioning as an absolute compass. Then the cell will only be active at moments when the head direction cells corresponding to 30 degrees are active, and the Hebbian learning process will cause the feature detecting cell to be linked by strong weights to these cells, but by vanishing weights to other head direction cells. If the absolute bearing of the cue card were more variable, then the connection strengths from the feature detecting cell would be weaker and more broadly dispersed . In the limit where the bearing of the cue card was completely random, all connections would be weak and equal. Thus the influence of a visual cue is determined by the amount of training and by the variability in its bearing (with respect to the head direction system). It can be seen that the model implements a competition between visual inputs and vestibular inputs for control of the head direction cells. If the visual cues are rotated while the rat is left stationary, then the head direction cells may either rotate to follow the visual cues, or stick with the inertial frame, depending on parameter values and, importantly, on the training regimen imposed on the network. Both of these outcomes have been observed in anterior thalamic head direction cells (McNaughton et al., 1993). 3 Discussion Do the necessary types of cells exist in the brain? Cells in the brainstem vestibular nuclei are known to have the properties required by the model (Precht, 1978). The "rotation" cells would be recognizable from the fact that they would fire only when A Model of the Neural Basis of the Rat's Sense of Direction 179 the rat is facing in a particular direction and turning in a particular direction, with rate at least roughly proportional to the speed ofturning. Cells with these properties have been recorded in the postsubiculum (Markus et al., 1990) and retrosplenial cortex (Chen et ai., 1994a). The visual cells would be recognizeable from the fact that they would respond to visual stimuli only when they come from a particular direction with respect to the animal's head axis. Cells with these properties have been recorded in the inferior parietal cortex, the internal medullary lamina of the thalamus, and the superior colliculus (e.g., Sparks, 1986). The superior colliculus also contains cells that respond in a direction-dependent manner to auditory inputs, thus allowing a possiblility of control of the head direction system by sound sources. There do not seem to be any strong direct projections from the superior colli cui us to the components of the head direction system, but there are numerous indirect pathways. The most general prediction of the model is that the influence of vestibular input upon head direction cells is not susceptible to experience-dependent modification, whereas the influence of visual input is plastic, and is enhanced by the duration of experience, the richness of the visual cue array, and the distance of visual cues from the rat's region of travel. The "rotation" cells should be responsive to stimulation of the vestibular system. It is possible to activate the vestibular system by applying hot or cold water to the ears: if this is done in the dark, and head direction cells are simultaneously recorded, the model predicts that they will show periodic bursts of activity, with a frequency related to the intensity of the stimulus. For another prediction, suppose we train two groups of rats to forage in a cylinder containing a single landmark. For one group, the landmark is placed at the edge of the cylinder; for the other group, the same landmark is placed halfway between the center and the edge. The model predicts that in both cases the landmark will influence the head direction sytem, but the influence will be stronger and more tightly focused when the landmark is at the edge. In some respects the model is flexible, and may be extended without compromising its essence. For example, there is no intrinsic necessity that the vestibular system be the sole input to the rotation cells (other than the head direction cells). The performance of the system might be improved in some ways by sending the rotation cells input about optokinetic flow, or certain types of motor efference copy. But there is as yet no clear evidence for these things. On a more abstract level, the mechanism used by the model for vestibular control may be thought of as a special case of a general-purpose method for integration with neurons. As such, it has significant advantages over some previously proposed neural integrators, in particular, better stability properties. It might be worth considering whether the method is applicable in other situations where integrators are known to exist, for example the control of eye position. 180 William Skaggs, James J. Knierim, Hemant S. Kudrimoti, Bruce L. McNaughton Supported by MH46823 and O.N .R. References Bienenstock, E. L., Cooper, L. N., and Munro, P. W. (1982). Theory for the development of neuron selectivity: orientation specificity and binocular interaction in visual cortex. J. Neurosci., 2:32-48. Chen, L. L., Lin, L., Green, E. J., Barnes, C. A., and McNaughton, B. L. (1994b). Head-direction cells in the rat posterior cortex. 1. Anatomical distribution and behavioral modulation. Exp. Brain Res., 101:8-23. Chen, L. L., Lin, L., Barnes, C. A., and McNaughton, B. L. (1994a). Head-direction cells in the rat posterior cortex. II. Contributions of visual and ideothetic information to the directional firing. Exp. Brain Res., 101:24-34. Gallistel, C. R. (1990). The Organization oj Learning. MIT Press, Cambridge, Massachusetts. Knierim, J. J ., Kudrimoti, H. S., and McNaughton, B. L. (1995). Place cells, head direction cells, and the learning of landmark stability. J. Neurosci. (in press). Markus, E. J., McNaughton, B. L., Barnes, C. A., Green, J. C., and Meltzer, J. (1990). Head direction cells in the dorsal presubiculum integrate both visual and angular velocity information. Soc. Neurosci. Abs., 16:44l. McNaughton, B. L., Chen, L. L., and Markus, E. J. (1991). ((Dead reckoning," landmark learning, and the sense of direction: A neurophysiological and computational hypothesis. J. Cognit. Neurosci ., 3: 190-202. McNaughton, B. L., Markus, E. J., Wilson, M. A., and Knierim, J . J. (1993). Familiar landmarks can correct for cumulative error in the inertially based dead-reckoning system. Soc. Neurosci. Abs., 19:795. Mizumori, S. J. and Williams, J. D. (1993). Directionally selective mnemonic properties of neurons in the lateral dorsal nucleus of the thalamus of rats. J. Neurosci., 13:4015-4028. Precht, W. (1978). Neuronal operations in the vestibular system. Springer, New York. Ranck, Jr ., J. B. (1984). Head direction cells in the deep cell layer of dorsal presubiculum in freely moving rats . Soc. Neurosci. Abs., 10:599. Singer, W . and Artola, A. (1994). Plasticity of the mature neocortex. In Selverston, A. I. and Ascher, P., editors, Cellular and molecular mechanisms underlying higher neural junctions, pages 49-69. Wiley. Sparks, D. L. (1986). Translation of sensory signals into commands for control of saccadic eye movements: role of primate superior colliculus. Physiol. Rev., 66:118-17l. Taube, J. S., Muller, R. V., and Ranck, Jr ., J. B. (1990a) . Head direction cells recorded from the postsubiculum in freely moving rats. I. Description and quantitative analysis . J. Neurosci., 10:420-435. Taube, J. S., Muller, R. V., and Ranck, Jr., J. B. (1990b) . Head direction cells recorded from the postsubiculum in freely moving rats. II. Effects of environmental manipulations. J. Neurosci., 10:436-447. PARTm LEARNiNGTHEORYANDDYNANUCS
890 |@word cingulate:1 hippocampus:1 stronger:3 mammal:1 thereby:1 moment:1 necessity:1 initial:1 contains:3 series:1 ranck:4 interestingly:1 current:1 com:1 anterior:4 activation:1 yet:1 physiol:1 distant:1 plasticity:1 shape:2 motor:1 plot:1 stationary:1 cue:17 half:2 vanishing:1 cognit:1 detecting:3 optokinetic:1 location:4 disoriented:8 sigmoidal:2 burst:1 direct:1 become:2 gallistel:2 retrosplenial:3 pathway:1 behavioral:2 recognizable:1 introduce:1 manner:2 expected:1 roughly:1 behavior:2 brain:6 integrator:2 food:1 considering:1 project:1 underlying:2 what:1 selverston:1 ought:1 quantitative:1 every:1 shed:1 stick:1 control:9 unit:1 before:2 postsubicular:2 tends:1 aging:1 limit:1 hemant:6 firing:8 becoming:1 approximately:1 modulation:1 might:2 studied:1 weakened:1 suggests:1 limited:1 range:2 implement:1 cold:1 area:3 thought:1 dictate:1 projection:2 specificity:2 cannot:1 influence:10 applying:1 darkness:1 bill:1 imposed:1 center:1 straightforward:1 williams:3 duration:1 focused:1 resolution:1 spark:2 rule:4 array:1 importantly:1 dominate:1 population:2 stability:2 mcnaughton:14 enhanced:1 suppose:6 hypothesis:1 velocity:2 particularly:1 located:1 predicts:2 observed:2 role:2 region:1 richness:1 decrease:1 movement:1 colli:1 dynamic:3 trained:1 weakly:2 upon:3 division:1 basis:5 completely:1 rightward:1 indirect:2 train:1 activate:1 mizumori:3 formation:1 outcome:1 quite:3 plausible:1 efference:1 say:1 interconnection:1 artola:2 think:1 final:1 directionally:1 advantage:1 net:1 interaction:1 maximal:1 neighboring:4 turned:1 aligned:1 causing:1 date:1 rapidly:1 description:1 moved:1 competition:1 az:1 cluster:2 rotated:2 lamina:1 depending:2 develop:2 excites:1 sole:1 strong:6 soc:3 arl:1 implies:1 come:3 direction:80 correct:2 compromising:1 subsequently:1 centered:1 human:1 broke:1 brainstem:1 wall:1 investigation:1 correction:1 around:3 exp:2 presumably:1 sytem:1 earth:1 purpose:2 polar:1 travel:1 applicable:1 pellet:1 kudrimoti:6 mit:1 always:1 aim:1 modified:1 command:1 wilson:1 consistently:1 indicates:1 sense:10 helpful:1 dependent:3 sketchy:1 initially:1 bienenstock:2 relation:1 selective:1 flexible:1 orientation:1 development:2 animal:7 special:1 integration:1 field:2 equal:1 broad:1 ascher:1 inevitable:1 stimulus:2 few:2 oriented:1 simultaneously:4 tightly:1 individual:1 familiar:1 fire:3 attractor:2 william:5 maintain:1 ab:3 cylinder:7 organization:2 circular:1 alignment:4 navigation:1 light:2 activated:3 edge:3 necessary:1 experience:2 unless:1 sooner:1 old:1 ruled:1 circle:6 re:2 isolated:1 earlier:1 compass:2 deviation:3 too:1 front:1 periodic:1 peak:10 sensitivity:1 stay:1 off:1 together:2 reflect:1 recorded:8 ear:1 containing:1 possibly:1 external:2 dead:2 account:2 nonlinearities:1 coding:2 north:1 caused:4 linked:1 precht:2 portion:1 sort:1 thalamic:1 complicated:2 parallel:1 bruce:6 contribution:1 reckoning:2 directional:8 weak:2 plastic:1 regimen:1 worth:1 explain:1 detector:5 whenever:2 synaptic:1 ofreal:1 frequency:1 james:5 associated:1 gain:2 auditory:1 begun:1 intrinsically:1 massachusetts:1 subserving:1 sophisticated:2 actually:1 inertial:1 higher:1 follow:1 improved:1 done:1 though:3 strongly:2 angular:2 binocular:1 working:1 receives:3 indicated:1 gray:1 effect:4 true:1 functioning:1 adequately:1 white:1 during:5 inferior:1 covering:1 essence:1 excitation:1 rat:39 hippocampal:2 outline:1 complete:1 motion:1 reflection:1 passive:1 recently:1 superior:4 rotation:17 behaves:1 stimulation:1 tail:1 significant:1 cambridge:1 ai:1 tuning:7 postsubiculum:5 session:4 had:4 rot:1 moving:3 stable:3 cortex:7 add:1 something:1 posterior:3 recent:1 manipulation:3 selectivity:1 certain:1 came:1 continue:1 life:1 exploited:1 muller:2 seen:1 freely:3 taube:3 signal:1 forage:2 ii:2 multiple:1 sound:1 thalamus:6 hebbian:3 long:1 lin:2 mnemonic:1 molecular:1 prediction:3 represent:1 cell:103 preserved:1 addition:2 schematically:1 whereas:1 interval:1 leaving:1 concluded:1 crucial:1 source:2 sends:1 sure:2 recording:6 virtually:1 ltp:1 thing:2 mature:1 flow:1 seem:1 unitary:1 easy:1 meltzer:1 skaggs:5 architecture:3 perfectly:1 silent:2 regarding:1 shift:7 inactive:1 whether:1 munro:1 remap:1 ltd:1 effort:1 york:1 cause:6 constitute:1 deep:1 clear:1 ifit:1 amount:2 dark:3 neocortex:2 simplest:1 exist:2 inhibitory:1 happened:1 correctly:1 modifiable:1 anatomical:2 broadly:1 group:7 salient:1 four:1 threshold:1 prevent:1 breadth:1 halfway:1 colliculus:3 angle:1 respond:2 place:5 bit:1 layer:3 arizona:5 barnes:3 activity:1 strength:2 occur:1 markus:5 speed:1 passively:1 relatively:1 according:1 expository:1 cui:1 belonging:1 jr:3 slightly:1 em:1 intimately:1 postsynaptic:3 rev:1 modification:2 primate:1 gradually:3 restricted:1 taken:1 previously:2 turn:3 eventually:1 mechanism:10 needed:1 singer:2 sending:1 available:2 operation:1 junction:1 away:1 responsive:1 allocentric:1 testable:1 especially:1 build:1 move:4 spike:1 saccadic:1 dependence:1 responds:2 distance:1 separate:1 card:10 rotates:1 lateral:1 landmark:8 presynaptic:2 cellular:1 toward:1 reason:1 water:1 relationship:1 difficult:1 susceptible:1 potentially:1 negative:1 perform:1 allowing:1 observation:1 neuron:3 thence:1 parietal:1 situation:1 extended:1 variability:2 head:68 jim:1 frame:3 synchronously:1 perturbation:1 knierim:10 drift:1 intensity:1 unpublished:1 required:1 connection:12 conflict:1 vestibular:19 impair:1 pattern:1 built:1 including:1 memory:1 explanation:1 erratic:1 max:1 hot:1 green:2 oj:1 turning:2 eye:2 numerous:1 axis:2 presubiculum:3 discovery:1 topography:1 interesting:2 proportional:2 facing:1 ingredient:2 localized:1 nucleus:4 integrate:1 degree:5 consistent:2 medullary:1 realign:1 editor:1 translation:1 prone:1 course:3 excitatory:3 accounted:1 placed:5 last:1 free:1 surprisingly:1 copy:1 supported:1 side:2 weaker:2 understand:1 neighbor:1 correspondingly:1 emerge:1 absolute:3 curve:4 depth:1 contrariwise:1 world:1 cumulative:1 sensory:2 author:1 nsma:4 made:1 refinement:1 far:1 active:9 inertially:1 assumed:2 disrupt:1 decade:1 why:1 reviewed:1 reality:1 nature:1 reasonably:1 symmetry:2 bearing:4 main:1 linearly:1 neurosci:9 nothing:1 neuronal:1 west:1 referred:1 cooper:1 wiley:1 experienced:2 position:1 col:1 minute:2 specific:6 evidence:3 intrinsic:4 ssw:1 illustrates:2 chen:5 neurophysiological:1 visual:23 springer:1 environmental:1 dispersed:1 formulated:1 room:1 considerable:2 change:2 directionality:2 typical:4 specifically:1 operates:1 determined:1 specie:1 connectional:1 experimental:1 selectively:1 internal:3 rotate:3 dorsal:4 tested:1
7,125
891
Morphogenesis of the Lateral Geniculate Nucleus: How Singularities Affect Global Structure Svilen Tzonev Beckman Institute University of Illinois Urbana, IL 61801 svilen@ks.uiuc.edu Klaus Schulten Beckman Institute University of Illinois Urbana, IL 61801 kschulte@ks.uiuc.edu Joseph G. Malpeli Psychology Department University of Illinois Champaign, IL 61820 jmalpeli@uiuc.edu Abstract The macaque lateral geniculate nucleus (LGN) exhibits an intricate lamination pattern, which changes midway through the nucleus at a point coincident with small gaps due to the blind spot in the retina. We present a three-dimensional model of morphogenesis in which local cell interactions cause a wave of development of neuronal receptive fields to propagate through the nucleus and establish two distinct lamination patterns. We examine the interactions between the wave and the localized singularities due to the gaps, and find that the gaps induce the change in lamination pattern. We explore critical factors which determine general LGN organization. 1 INTRODUCTION Each side of the mammalian brain contains a structure called the lateral geniculate nucleus (LGN), which receives visual input from both eyes and sends projections to 134 Svilen Tzonev, Klaus Schulten, Joseph G. Malpeli the primary visual cortex. In primates the LG N consists of several distinct layers of neurons separated by intervening layers ofaxons and dendrites. Each layer of neurons maps the opposite visual hemifield in a topographic fashion. The cells comprising these layers differ in terms of their type (magnocellular and parvocellular) , their input (from ipsilateral (same side) and contralateral (opposite side) eyes), and their receptive field organization (ON and OFF center polarity). Cells in one layer receive input from one eye only (Kaas et al., 1972), and in most parts of the nucleus have the same functional properties (Schiller & Malpeli, 1978). The maps are in register, i.e., representations of a point in the visual field are found in all layers, and lie in a narrow column roughly perpendicular to the layers (Figure 1). A prominent a projection column Figure 1: A slice along the plane of symmetry of the macaque LGN. Layers are numbered ventral to dorsal. Posterior is to the left, where foveal (central) parts of the retinas are mapped; peripheral visual fields are mapped anteriorly (right). Cells in different layers have different morphology and functional properties: 6-P /C/ON; 5-P /I/ON; 4-P /C/OFF; 3-P /I/OFF; 2-M/I/ON&OFF; 1-M/C/ON&OFF, where P is parvocellular, M is magnocellular, C is contralateral, I is ipsilateral, ON and OFF refer to polarities of the receptive-field centers. The gaps in layers 6, 4, and 1 are images of the blind spot in the contralateral eye. Cells in columns perpendicular to the layers receive input from the same point in the visual field. feature in this laminar organization is the presence of cell-free gaps in some layers. These gaps are representations of the blind spot (the hole in the retina where the optic nerve exits) of the opposite retina. In the LGN of the rhesus macaque monkey (Macaca mulatta) the pattern of laminar organization drastically changes at the position of the gaps - foveal to the gaps there are six distinct layers, peripheral to the gaps there are four layers. The layers are extended two-dimensional structures whereas the gaps are essentially localized. However, the laminar transition occurs in a surface that extends far beyond the gaps, cutting completely across the main axis of the LGN (Malpeli & Baker., 1975). We propose a developmental model of LGN laminar morphogenesis. In particular, we investigate the role of the blind-spot gaps in the laminar pattern transition, and their extended influence over the global organization of the nucleus. In this model a wave of development caused by local cell interactions sweeps through the system (Figure 2). Strict enforcement of retinotopy maintains and propagates an initially localized foveal pattern. At the position of the gaps, the system is in a metastable Morphogenesis of the Lateral Geniculate Nucleus 135 wavefront maturing cells . immature cells ~;r , F ...... ...... . I? ? ? ? ~ 1 ? I I I .. -->- : I .,Y ~/!.. .".,. ....~ . ... .- X blind spot gap (no cells) Figure 2: Top view of a single layer. As a wave of development sweeps through the LGN the foveal part matures first and the more peripheral parts develop later. The shape of the developmental wave front is shown schematically by lines of "equal development" . state, and the perturbation in retinotopy caused by the gaps is sufficient to change the state of the system to its preferred four-layered pattern. We study the critical factors in this model, and make some predictions about LGN morphogenesis. 2 MODEL OF LGN MORPHOGENESIS We will consider only the upper four (parvocellular) layers since the laminar transition does not involve the other two layers. This transition results simply from a reordering of the four parvocellular strata (Figure 1). Foveal to the gaps, the strata form four morphologically distinct layers (6, 5, 4 and 3) because adjacent strata receive inputs from opposite eyes, which "repel" one another. Peripheral to the gaps, the reordering of strata reduces the number of parvocellular eye alternations to one, resulting in two parvocellular layers (6+4 and 5+3). 2.1 GEOMETRY AND VARIABLES LGN cells Ci are labeled by indices i = 1,2, ... , N. The cells have fixed, quazi-random and uniformly distributed locations ri E V e R3, where V = {(x,y,z) 10 < x < Sz,O < y < Sy,O < z < Sz}, and belong to one projection column Cab, a = 1,2, ... ,A and b = 1,2, ... ,B, (Figure 3). Functional properties of the neurons change in time (denoted by T), and are described by eye specificity and receptive-field polarity, ei(T), and Pi(T), respectively: ei (T), pdT) E [-1,1] C R, i = 1,2, ... ,N, T = 0, 1, ... , T maz ? The values of eye specificity and polarity represent the proportions of synapses from competing types of retinal ganglion cells (there are four type of ganglion cells from different eyes and with ON or OFF polarity). ei = -1 (ei = 1) denotes that the i-th cell is receiving input solely from the opposite (same side) retina. Similarly, Pi = -1 (Pi = 1) denotes that the cell input is pure ON (OFF) center. Intermediate values of ei and Pi imply that the cell does not have pure properties (it receives 136 Svilen Tzonev, Klaus Schulten, Joseph G. Malpeli N) Figure 3: Geometry of the model. LGN cells Ci (i = 1,2, ... , have fixed random, and uniformly-distributed locations Ti within a volume VcR , and belong to one projection column Cab. input from retinal ganglion cells of both eyes and with different polarities). Initially, at r = 0, all LGN cells are characterized by ei, Pi = O. This corresponds to two possibilities: no retinal ganglion cells synapse on any LGN cell, or proportions of synapses from different ganglion cells on all LGN neurons are equal, i.e., neurons possess completely undetermined functionality because of competing inputs of equal strength. As the neurons mature and acquire functional properties, their eye specificity and polarity reach their asymptotic values, ?l. Even when cells are not completely mature, we will refer to them as being of four different types, depending on the signs of their functional properties. Following accepted anatomical notation, we will label them as 6, 5, 4, and 3. We denote eye specificity of cell types 6 and 4 as negative, and cell types 5 and 3 as positive. Polarity of cell types 6 and 5 is negative, while polarity of types 4 and 3 is positive. Cell functional properties are subject to the dynamics described in the following section. The process of LGN development starts from its foveal part, since in the retina it is the fovea that matures first. As more peripheral parts of the retina mature, their ganglion cells start to compete to establish permanent synapses on LGN cells. In this sorting process, each LGN cell gradually emerges with permanent synapses that connect only to several neighboring ganglions of the same type. A wave of gradual development of functionality sweeps through the nucleus. The driving force for this maturation process is described by localized cell interactions modulated by external influences. The particular pattern of the foveal lamination is shaped by external forces, and later serves as a starting point for a "propagation of sameness" of cell properties. Such a sameness propagation produces clustering of similar cells and formation of layers. It should be stressed that cells do not move, only their characteristics change. 2.2 DYNAMICS The variables describing cell functional properties are subject to the following dynamics edr + 1) Pi (r + 1) (r) + ~ei (r) + 1Je Pi (r) + ~Pi (r) + 1Jp, ei - 1,2, ... ,N. (1) Morphogenesis of the Lateral Geniculate Nucleus 137 In Eq. (1), there are two contributions to the change of the intermediate variables and Pi (T). The first is deterministic, given by ei (T) 6.e;( r) = ,,(r,) [ 6.p,(r) - ,,(r,) (t. e; (r)! (Ie; - r;ll) + E", (r,)] (1- c; (r)) f3;. [(t,p;(r)!(lr,_r;ll) +P", (r,)] (I-p;(r)) f3;?. (2) The second is a stochastic contribution corresponding to fluctuations in the growth of the synapses between retinal ganglion cells and LGN neurons. This noise in synaptic growth plays both a driving and a stabilizing role to be explained below. We explain the meaning of the variables in Eq. (2) only for the eye specificity variable ei. The corresponding parameters for polarity Pi have similar interpretations. The parameter a (ri) is the rate of cell development. This rate is the same for eye specificity and polarity. It depends on the position ri of the cell in order to allow for spatially non-uniform development. The functional form of a (ri) is given in the Appendix. The term Eint (ri) = 2:7=1 ejl (h - rjl) is effectively a cell force field. This field influences the development of nearby cells and promotes clustering of same type of cells. It depends on the maturity of the generating cells and on the distance between cells through the interaction function 1(8). We chose for 1(8) a Gaussian form, i.e., 1 (8) = exp ( _6 2 / (T2), with characteristic interaction distance (T. The external influences on cell development are incorporated in the term for the external field Eext(ri). This external field plays two roles: it launches a particular laminar configuration of the system (in the foveal part of the LGN), and determines its peripheral pattern. It has, thus, two contributions Eext (ri) = E!xt (ri) + E~xt (ri). The exact forms of E!xt(ri) and E~xt(ri) are provided in the Appendix. The nonlinear term (1 - e~) in Eq. (2) ensures that ?1 are the only stable fixed points of the dynamics. The neuronal properties gradually converge to either of these fixed points capturing the maturation process. This term also stabilizes the dynamic variables and prevents them from diverging. The last term f3~b (T) reflects the strict columnar organization of the maps. At each step of the development the proportion of all four types of LGN cells is calculated within a single column Cab, and f3~b(T) for different types t is adjusted such that all types are equally represented. Without this term, the cell organization degenerates to a non-laminar pattern (the system tries to minimize the surfaces between cell clusters of different type). The exact form of f3!b(T) is given in the Appendix. At each stage of LGN development, cells receive input from retinal ganglion cells of particular types. This means that eye specificity and polarity of LGN cells are not independent variables. In fact, they are tightly coupled in the sense that lei (T) I = IPi ( T) I should hold for all cells at all times. This gives rise to coupled dynamics described by 138 Svilen Tzonev, Klaus Schulten, Joseph G. Malpeli mi(7+1) ei (7 + 1) Pi (7 + 1) = - min ( Iei (7 + 1) I, IPi (7 + 1) I) mi (7 + 1) sgn ( ei (7 + 1) ) md7+ 1) sgn(Pi(7+ 1)), i = 1,2, ... ,N. (3) The blind spot gaps are modeled by not allowing cells in certain columns to acquire types of functionality for which retinal projections do not exist, e.g., from the blind spot of the opposite eye. Accordingly, ei is not allowed to become negative. Thus, some cells never reach a pure state ei, Pi = ? 1. It is assumed that in reality such cells die out. Of all quantities and parameters, only variables describing the neuronal receptive fields (ei and Pi) are time-dependent. 3 RESULTS We simulated the dynamics described by Eqs. (1, 2, 3), typically for 100,000 time steps. Depending on the rate of cell development, mature states were reached in about 10,000 steps. The maximum value of Q: was 0.0001. We used an interaction function with u = 1. First, we considered a two-dimensional LGN, V = {(x,z)IO < x < Sx,O < z < Sz} with Sx = 10 and Sz = 6. There were ten projection columns (with equal size) along the x axis. An initial pattern was started in the foveal part by the external field. The size of the gaps 9 measured in terms of the interaction distance u was crucial for pattern development. When the developmental wave reached the gaps, layer 6 could" jump" its gap and continued to spread peripherally if the gap was sufficiently narrow (g/u < 1.5). If its gap was not too narrow (g/u > 0.5), layer 4 completely stopped (since cells in the gaps were not allowed to acquire negative eye specificity) and so layers 5 and 3 were able to merge. Cells of type 4 reappeared after the gaps (Figure 4, right side, shows behavior similar to the two-dimensional model) because of the required equal representation of all cell types in the projection columns, and because of noise in cell development. Energetically, the most favorable position of cell type 4 would be on top of type 6, which is inconsistent with experimental observations. Therefore, one must assume the existence of an external field in the peripheral part that will drive the system away from its otherwise preferred state. If the gaps were too large (g/u > 1.5), cells of type 6 and 4 reappeared after the gaps in a more or less random vertical position and caused transitions of irregular nature. On the other hand, if the gaps were too narrow (g/u < 0.5), both layers 6 and 4 could continue to grow past their gaps, and no transition between laminar patterns occurred at all. When g/u was close to the above limits, the pattern after the gaps differed from trial to trial. For the two-dimensional system, a realistic peripheral pattern always occurred for 0.7 < g/u < 1.2. = = = We simulated a three-dimensional system with size Sx 10, Sy 10, and Sz 6, and projection columns ordered in a 10 by 10 grid. The topology of the system is different in two and three dimensions: in two dimensions the gaps interrupt the layers completely and, thus, induce perturbations which cannot be by-passed. In three dimensions the gaps are just holes in a plane and generate localized perturbations: the layers can, in principle, grow around the gaps maintaining the initial laminar pattern. Nevertheless, in the three-dimensional case, an extended transi- Morphogenesis of the Lateral Geniculate Nue/ells 139 6 5 4~~UIIii 3 Figure 4: Left: Mature state of the macaque LG~ - result of the three-dimensional model with 4,800 cells. Spheres with different shades represent cells with different properties. Gaps in strata 6 and 4 (this gap is not visible) are coded by the darkest color, and coincide with the transition surface between 4- and 2-layered patterns. Right: A cut of the three-dimensional structure along its plane of symmetry. A two-dimensional system exhibits similar organization. Compare with upper layers in Figure 1. Spatial segregation between layers is not modeled explicitly. tion was triggered by the gaps. The transition surface, which passed through the gaps and was oriented roughly perpendicularly to the x axis, cut completely across the nucleus (Figure 4). Several factors were critical for the general behavior of the system. As in two dimensions, the size of the gap~ must be within certain limits: typically 0.5 < g / (J" < 1.0. These limits depend on the curvature of the wavefront. The gaps must lie in a certain "inducing" interval along the x axis. If they were too close to the origin, the foveal pattern was still more stable, so no transition could be induced there. However, a spontaneous transition might occur downstream. If the gaps were too far from the origin, a ~pontaneou~ transition might occur before them. The occurrence and location of a spontaneous transition, (therefore, the limits of the "inducing" interval) depended on the external-field parameters. A realistic transition was observed only when the front of the developmental wave had sufficient curvature when it reached the gaps. Underlying anatomical reasons for a sufficiently curved front along the main axis could be the curvature of the nucleus, differences in layer thickness, or differences in ganglion-cell densities in the retinas. Propagation of the developmental wave away from the gaps was quite stable. Before and after the gaps, the wave simply propagated the already established patterns. In a system without gaps, transitions of variable shape and location occurred when the peripheral contribution to the external fields was sufficiently large; a weaker contribution allowed the foveal pattern to propagate through the entire nucleus. 4 SUMMARY \Ve present a model that successfully captures the most important features of macaque LG~ morphogenesis. It produces realistic laminar patterns and supports 140 Svilen Tzonev, Klaus Schulten, Joseph G. Malpeli the hypothesis (Lee & Malpeli, 1994) that the blind spot gaps trigger the transition between patterns. It predicts that critical factors in LGN development are the size and location of the gaps, cell interaction distances, and shape of the front of the developmental wave. The model may be general enough to incorporate the LGN organizations of other primates. Small singularities, similar to the blind spot gaps, may have an extended influence on global organization of other biological systems. Acknowledgements This work has been supported by a Beckman Institute Research Assistantship, and by grants PHS 2P41 RR05969 and NIH EY02695. References J.H. Kaas, R.W. Guillery & J.M. Allman. (1972) Some principles of organization in the dorsal lateral geniculate nucleus, Brain Behav. Evol., 6: 253-299. D.Lee & J.G.Malpeli. (1994) Global Form and Singularity: Modeling the Blind Spot's Role in Lateral Geniculate Morphogenesis, Science, 263: 1292-1294. J.G. Malpeli & F.H. Baker. (1975) The representation of the visual field in the lateral geniculate nucleus of Macaca mulatta, 1. Comp. Neural., 161: 569-594. P.H. Schiller & J.G. Malpeli. (1978) Functional specificity ofLGN ofrhesus monkey, 1. Neurophysiol., 41: 788-797. APPENDIX The form of 0 (x, y, z) (with o(x,y,z) 00 = 0.0001) was chosen as = 00(0.1+exp(-(y-Sy/2)2)). (4) Foveal external fields of the following form were used: E!:z:t(x,y,z) = lO[O(z-d)-20(z-2d)+20(z-3d)-O(d-z)] exp(-x) p!:z:t(x,y,z) = 10[20(z-2d)-1]exp(-x), (5) where d = Sz/4 is the layers' thickness and the "theta" function is defined as O( x) = 1, x > 0 and O( x) = 0, x < O. Peripheral external fields (in fact they are present everywhere but determine the pattern in the peripheral part only) were chosen as E::z:t(x,y,z) = 5[20(z-2d)-1] P::z:t(x,y,z) = 5 [O(z-d)-20(z-2d)+20(z-3d)-O(d-z)]. (6) f3!b (1') was calculated in the following way: at any given time 1', within the column Cab, we counted the number N~b( 1') of cells, that could be classified as one of the four types t = 3,4,5,6. Cells with ei( 1') or Pi( 1') exactly zero were not counted. The total number of classified cells is then Nab( 1') = 2:~=3 N~b( 1'). If there were no classified cells (Nab( 1') = 0), then f3~b( 1') was set to one for all t. Otherwise the ratio of different types was calculated: n~b = N~b(1')/Nab(1'). In this way we calculated f3~b (1') = 4 - 12 nab, t = 3,4,5,6. (7) If f3~b (1') was negative it was replaced by zero.
891 |@word trial:2 maz:1 proportion:3 rhesus:1 propagate:2 gradual:1 initial:2 configuration:1 contains:1 foveal:12 past:1 must:3 visible:1 realistic:3 midway:1 shape:3 reappeared:2 accordingly:1 plane:3 lr:1 location:5 ipi:2 along:5 become:1 maturity:1 consists:1 intricate:1 behavior:2 roughly:2 examine:1 uiuc:3 morphology:1 brain:2 provided:1 baker:2 notation:1 underlying:1 monkey:2 ti:1 growth:2 exactly:1 grant:1 positive:2 before:2 local:2 limit:4 io:1 depended:1 solely:1 fluctuation:1 merge:1 might:2 chose:1 k:2 perpendicular:2 wavefront:2 vcr:1 spot:10 projection:8 induce:2 numbered:1 specificity:9 cannot:1 close:2 layered:2 influence:5 map:3 deterministic:1 center:3 starting:1 stabilizing:1 evol:1 pure:3 continued:1 spontaneous:2 play:2 trigger:1 exact:2 hypothesis:1 origin:2 mammalian:1 cut:2 predicts:1 labeled:1 observed:1 role:4 capture:1 ensures:1 developmental:6 parvocellular:6 dynamic:7 depend:1 exit:1 completely:6 neurophysiol:1 represented:1 separated:1 distinct:4 klaus:5 formation:1 quite:1 otherwise:2 topographic:1 nab:4 triggered:1 propose:1 interaction:9 neighboring:1 malpeli:11 degenerate:1 intervening:1 inducing:2 macaca:2 cluster:1 produce:2 generating:1 depending:2 develop:1 measured:1 eq:4 launch:1 differ:1 functionality:3 stochastic:1 sgn:2 biological:1 singularity:4 adjusted:1 hold:1 sufficiently:3 considered:1 around:1 exp:4 stabilizes:1 driving:2 ventral:1 favorable:1 geniculate:9 beckman:3 label:1 successfully:1 assistantship:1 reflects:1 gaussian:1 always:1 interrupt:1 sense:1 dependent:1 typically:2 entire:1 initially:2 lgn:26 comprising:1 morphologically:1 denoted:1 development:16 spatial:1 ell:1 field:19 equal:5 f3:9 shaped:1 never:1 t2:1 perpendicularly:1 retina:8 oriented:1 tightly:1 ve:1 replaced:1 geometry:2 organization:11 investigate:1 possibility:1 stopped:1 column:11 modeling:1 contralateral:3 undetermined:1 uniform:1 front:4 too:5 connect:1 thickness:2 density:1 stratum:5 ie:1 lee:2 off:8 receiving:1 hemifield:1 central:1 external:11 iei:1 retinal:6 permanent:2 explicitly:1 register:1 caused:3 blind:10 depends:2 tion:1 later:2 view:1 try:1 kaas:2 reached:3 wave:11 start:2 maintains:1 contribution:5 minimize:1 il:3 characteristic:2 sy:3 comp:1 drive:1 classified:3 explain:1 synapsis:5 reach:2 synaptic:1 mi:2 propagated:1 color:1 emerges:1 nerve:1 matures:2 maturation:2 synapse:1 just:1 stage:1 hand:1 receives:2 ei:16 nonlinear:1 propagation:3 lei:1 spatially:1 adjacent:1 ll:2 die:1 prominent:1 ofaxons:1 image:1 meaning:1 nih:1 mulatta:2 functional:9 jp:1 volume:1 belong:2 interpretation:1 ejl:1 occurred:3 refer:2 grid:1 similarly:1 illinois:3 had:1 stable:3 cortex:1 surface:4 curvature:3 posterior:1 certain:3 continue:1 immature:1 alternation:1 tzonev:5 determine:2 converge:1 reduces:1 champaign:1 characterized:1 sphere:1 equally:1 promotes:1 coded:1 prediction:1 essentially:1 represent:2 transi:1 cell:70 irregular:1 receive:4 whereas:1 schematically:1 interval:2 grow:2 sends:1 crucial:1 posse:1 strict:2 subject:2 induced:1 mature:5 inconsistent:1 allman:1 presence:1 intermediate:2 enough:1 affect:1 psychology:1 competing:2 opposite:6 topology:1 six:1 passed:2 energetically:1 cause:1 behav:1 involve:1 ten:1 ph:1 generate:1 exist:1 sign:1 ipsilateral:2 anatomical:2 four:9 nevertheless:1 peripherally:1 downstream:1 compete:1 everywhere:1 extends:1 appendix:4 capturing:1 layer:31 laminar:11 strength:1 occur:2 optic:1 ri:11 nearby:1 min:1 department:1 metastable:1 peripheral:11 across:2 joseph:5 primate:2 explained:1 gradually:2 segregation:1 describing:2 r3:1 enforcement:1 serf:1 pdt:1 away:2 occurrence:1 darkest:1 existence:1 top:2 denotes:2 clustering:2 maintaining:1 maturing:1 establish:2 magnocellular:2 sweep:3 move:1 already:1 quantity:1 occurs:1 receptive:5 primary:1 rjl:1 exhibit:2 fovea:1 distance:4 schiller:2 mapped:2 lateral:9 simulated:2 reason:1 index:1 polarity:12 modeled:2 ratio:1 acquire:3 lg:3 negative:5 rise:1 allowing:1 upper:2 vertical:1 neuron:7 observation:1 urbana:2 coincident:1 curved:1 extended:4 incorporated:1 perturbation:3 morphogenesis:10 required:1 repel:1 narrow:4 established:1 macaque:5 beyond:1 able:1 below:1 pattern:23 critical:4 force:3 theta:1 eye:17 imply:1 axis:5 started:1 coupled:2 acknowledgement:1 asymptotic:1 reordering:2 localized:5 nucleus:15 sufficient:2 propagates:1 principle:2 pi:15 lo:1 summary:1 supported:1 last:1 free:1 drastically:1 side:5 allow:1 weaker:1 institute:3 distributed:2 slice:1 calculated:4 dimension:4 transition:15 jump:1 coincide:1 counted:2 far:2 cutting:1 preferred:2 sz:6 global:4 assumed:1 reality:1 nature:1 symmetry:2 dendrite:1 main:2 spread:1 noise:2 allowed:3 neuronal:3 je:1 fashion:1 differed:1 position:5 schulten:5 lie:2 shade:1 xt:4 effectively:1 ci:2 cab:4 p41:1 hole:2 sx:3 gap:47 sorting:1 columnar:1 simply:2 explore:1 ganglion:10 visual:7 prevents:1 ordered:1 corresponds:1 determines:1 retinotopy:2 change:7 uniformly:2 edr:1 called:1 total:1 accepted:1 experimental:1 diverging:1 support:1 modulated:1 dorsal:2 stressed:1 incorporate:1
7,126
892
Glove-TalkII: Mapping Hand Gestures to Speech Using Neural Networks S. Sidney Fels Department of Computer Science University of Toronto Toronto, ON, M5S lA4 ssfels@ai.toronto.edu Geoffrey Hinton Department of Computer Science University of Toronto Toronto, ON, M5S lA4 hinton@ai.toronto.edu Abstract Glove-TaikII is a system which translates hand gestures to speech through an adaptive interface. Hand gestures are mapped continuously to 10 control parameters of a parallel formant speech synthesizer. The mapping allows the hand to act as an artificial vocal tract that produces speech in real time. This gives an unlimited vocabulary in addition to direct control of fundamental frequency and volume. Currently, the best version of Glove-TalkII uses several input devices (including a CyberGlove, a ContactGlove, a 3space tracker, and a foot-pedal), a parallel formant speech synthesizer and 3 neural networks. The gesture-to-speech task is divided into vowel and consonant production by using a gating network to weight the outputs of a vowel and a consonant neural network. The gating network and the consonant network are trained with examples from the user. The vowel network implements a fixed, user-defined relationship between hand-position and vowel sound and does not require any training examples from the user. Volume, fundamental frequency and stop consonants are produced with a fixed mapping from the input devices. One subject has trained to speak intelligibly with Glove-TalkII. He speaks slowly with speech quality similar to a text-to-speech synthesizer but with far more natural-sounding pitch variations. 844 1 S. Sidney Fe Is, Geoffrey Hinton Introduction There are many different possible schemes for converting hand gestures to speech. The choice of scheme depends on the granularity of the speech that you want to produce. Figure 1 identifies a spectrum defined by possible divisions of speech based on the duration of the sound for each granularity. What is interesting is that in general, the coarser the division of speech, the smaller the bandwidth necessary for the user. In contrast, where the granularity of speech is on the order of articulatory muscle movements (i.e. the artificial vocal tract [AVT]) high bandwidth control is necessary for good speech. Devices which implement this model of speech production are like musical instruments which produce speech sounds. The user must control the timing of sounds to produce speech much as a musician plays notes to produce music. The AVT allows unlimited vocabulary, control of pitch and non-verbal sounds. Glove-TalkII is an adaptive interface that implements an AVT. Translating gestures to speech using an AVT model has a long history beginning in the late 1700's. Systems developed include a bellows-driven hand-varied resonator tube with auxiliary controls (1790's [9]), a rubber-moulded skull with actuators for manipulating tongue and jaw position (1880's [1]) and a keyboard-footpedal interface controlling a set of linearly spaced bandpass frequency generators called the Yoder (1940 [3]). The Yoder was demonstrated at the World's Fair in 1939 by operators who had trained continuously for one year to learn to speak with the system. This suggests that the task of speaking with a gestural interface is very difficult and the training times could be significantly decreased with a better interface. GloveTalkII is implemented with neural networks which allows the system to learn the user's interpretation of an articulatory model of speaking. This paper begins with an overview of the whole Glove-TalkII system. Then, each neural network is described along with its training and test results. Finally, a qualitative analysis is provided of the speech produced by a single subject after 100 hours of speaking with Glove-TalkI!. Artificial Vocal Tract (AVT) Phoneme Generator Finger Spelling Syllable Generator Word Generator I 10-30 > 100 130 200 600 Approximate time per gesture (msec) Figure 1: Spectrum of gesture-to-speech mappings based on the granularity of speech. 2 Overview of Glove-TalkII The Glove-TalkIl system converts hand gestures to speech , based on a gesture-toformant model. The gesture vocabulary is based on a vocal-articulator model of the hand. By dividing the mapping tasks into independent subtasks , a substantial reduction in network size and training time is possible (see [4]). Figure 2 illustrates the whole Glove-TalkIl system. Important features include the Glove- Talkll 845 Rlabt Hand data ll.y.:r: roll.l!itch. yaw every 1/60 IIOCUld ~~ ~, k-.--'- 10 flell analea 4 abduction anclea thumb and pinkie rotation wrist pitch and yaw every 1/100 aecond )))S~ Fllled Pitch Mappina VIC Doc:ision Network Vowel Network I---~ Combinina Function Ccnsonant Network Fllled SlOp Mappina Figure 2: Block diagram of Glove-TalkII: input from the user is measured by the Cyberglove, polhemus, keyboard and foot pedal, then mapped using neural networks and fixed functions to formant parameters which drive the parallel formant synthesizer [8]. three neural networks labeled vowel/consonant decision (V /C), vowel, and consonant. The V /C network is trained on data collected from the user to decide whether he wants to produce a vowel or a consonant sound . Likewise, the consonant network is trained to produce consonant sounds based on user-generated examples based on an initial gesture vocabulary. In contrast, the vowel network implements a fixed mapping between hand-positions and vowel phonemes defined by the user. Nine contact points measured on the user's left hand by a ContactGlove designate the nine stop consonants (B, D, G, J, P, T, K, CH, NG), because the dynamics of such sounds proved too fast to be controlled by the user. The foot pedal provides a volume control by adjusting the speech amplitude and this mapping is fixed. The fundamental frequency, which is related to the pitch of the speech, is determined by a fixed mapping from the user's hand height. The output of the system drives 10 control parameters of a parallel formant speech synthesizer every 10 msec. The 10 control parameters are: nasal amplitude (ALF), first, second and third formant frequency and amplitude (F1, A1, F2, A2, F3, A3), high frequency amplitude (AHF), degree of voicing (V) and fundamental frequency (FO). Each of the control parameters is quantized to 6 bits. Once trained, Glove-Talk II can be used as follows: to initiate speech, the user forms the hand shape of the first sound she intends to produce. She depresses the foot pedal and the sound comes out of the synthesizer. Vowels and consonants of various qualities are produced in a continuous fashion through the appropriate co-ordination of hand and foot motions. Words are formed by making the correct motions; for example, to say "hello" the user forms the "h" sound, depresses the foot pedal and quickly moves her hand to produce the "e" sound, then the "I" sound and finally the "0" sound. The user has complete control of the timing and quality of the individual sounds. The articulatory mapping between gestures and speech 846 S. Sidney Fe/s, Geoffrey Hinton ... .-),~O""':--+--"'--7-... -~-~c- - - - : - - Figure 3: Hand-position to Vowel Sound Mapping. The coordinates are specified relative to the origin at the sound A. The X and Y coordinates form a horizontal plane parallel to the floor when the user is sitting. The 11 cardinal phoneme targets are determined with the text-to-speech synthesizer. u - ... --!----"~ Y(cm) .. I is decided a priori. The mapping is based on a simplistic articulatory phonetic description of speech (5]. The X,Y coordinates (measured by the polhemus) are mapped to something like tongue position and height l producing vowels when the user's hand is in an open configuration (see figure 2 for the correspondence and table 1 for a typical vowel configuration). Manner and place of articulation for non-stop consonants are determined by opposition of the thumb with the index and middle fingers as described in table 1. The ring finger controls voicing. Only static articulatory configurations are used as training points for the neural networks, and the interpolation between them is a result of the learning but is not explicitly trained. Ideally, the transitions should also be learned, but in the text-to-speech formant data we use for training [6] these transitions are poor, and it is very hard to extract formant trajectories from real speech accurately. 2.1 The Vowel/Consonant (VIC) Network The VIC network decides, on the basis of the current configuration of the user's hand, to emit a vowel or a consonant sound. For the quantitative results reported here, we used a 10-5-1 feed-forward network with sigmoid activations [7]. The 10 inputs are ten scaled hand parameters measured with a Cyberglove: 8 flex angles (knuckle and middle joints of the thumb, index, middle and ring fingers), thumb abduction angle and thumb rotation angle. The output is a single number representing the probability that the hand configuration indicates a vowel. The output of the VIC network is used to gate the outputs of the vowel and consonant networks, which then produce a mixture of vowel and consonant formant parameters. The training data available includes only user-produced vowel or consonant sounds. The network interpolates between hand configurations to create a smooth but fairly rapid transition between vowels and consonants. For quantitative analysis, typical training data consists of 2600 examples of consonant configurations (350 approximants, 1510 fricatives [and aspirant], and 740 nasals) and 700 examples of vowel configurations. The consonant examples were obtained from training data collected for the consonant network by an expert user. The vowel examples were collected from the user by requiring him to move his hand in vowel configurations for a specified amount of time. This procedure was performed in several sessions. The test set consists of 1614 examples (1380 consonants and 234 vowels). After training,2 the mean squared error on the training and test lIn reality, the XY coordinates map more closely to changes in the first two formants, FI and F2 of vowels. From the user's perspective though, the link to tongue movement is useful. 2The VIe network, the vowel network and the consonant network are trained using Glove-Talkll 847 ~.. '.~ ;J(il. ~ ? ??:$??? , ~ ~ '.:~:~:::. DH F H 7?:. .I.!:.:? ~ ~. N R S .:::~::. .. ~ ....... ~ ' '.>' V :<: ? ~:;?::. " ~':" ':: , - - ~ . ..;- ., .... ... L M ''':;;1.. ' ...to:.?.?. ~: ). ":"~"'" SH TH . ~ ~ ~ Z ZH vowel .~ ":'s (, ~~. W , ' .? ~.,,> ? 0:' :~: ????'<1.?? Table 1: Static Gesture-to-Consonant Mapping for all phonemes. Note, each gesture corresponds to a static non-stop consonant phoneme generated by the text-to-speech synthesizer. set was less than 10- 4 . During normal speaking neither network made perceptual errors. The decision boundary feels quite sharp, and provides very predictable, quick transitions from vowels to consonants and back. Also, vowel sounds are produced when the user hyperextends his hand. Any unusual configurations that would intuitively be expected to produce consonant sounds do indeed produce consonant sounds. 2.2 The Vowel Network The vowel network is a 2-11-8 feed forward network . The 11 hidden units are normalized radial basis functions (RBFs) [2] which are centered to respond to one of 11 cardinal vowels. The outputs are sigmoid units representing 8 synthesizer control parameters (ALF, F1, AI, F2, A2, F3, A3, AHF). The radial basis function used is: L(Wji-O.)~ (1) where OJ is the (un-normalized) output of the RBF unit, Wji is the weight from unit i to unit j, 0i is the output of input unit i, and (1/ is the variance of the RBF. The normalization used is: O? nj = L J (2) oj=e- <l'j2 mEpom where nj is the normalized output of unit j and the summation is over all the units in the group of normalized RBF units. The centres of the RBF units are fixed conjugate gradient descent and a line search. 848 S. Sidney Fels, Geoffrey Hinton according to the X and Y values of each of the 11 vowels in the predefined mapping (see figure 2). The variances of the 11 RBF's are set to 0.025. The weights from the RBF units to the output units are trained. For the training data, 100 identical examples of each vowel are generated from their corresponding X and Y positions in the user-defined mapping, providing 1100 examples. Noise is then added to the scaled X and Y coordinates for each example. The added noise is uniformly distributed in the range -0.025 to 0.025. In terms of unscaled ranges, these correspond to an X range of approximately ? 0.5 cm and a Y range of ? 0.26 cm. Three different test sets were created. Each test set had 50 examples of each vowel for a total of 550 examples. The first test set used additive uniform noise in the interval ? 0.025. The second and third test sets used additive uniform noise in the interval ? 0.05 and ? 0.1 respectively. The mean squared error on the training set was 0.0016. The MSE on the additive noise test sets (noise = ? 0.025, 0.05 and 0.01) was 0.0018, 0.0038, 0.0120 which corresponds to expected errors of 1.1 %, 3.1% and 5.5% in the formant parameters, respectively. This network performs well perceptually. The key feature is the normalization of the RBF units. Often, when speaking, the user will overshoot cardinal vowel positions (especially when she is producing dipthongs) and all the RBF units will be quite suppressed. However, the normalization magnifies any slight difference between the activities of the units and the sound produced will be dominated by the cardinal vowel corresponding to the one whose centre is closest in hand space. 2.3 The Consonant Network The consonant network is a 10-14-9 feed-forward network . The 14 hidden units are normalized RBF units. Each RBF is centred at a hand configuration determined from training data collected from the user corresponding to one of 14 static consonant phonemes. The target consonants are created with a text-to-speech synthesizer. Figure 1 defines the initial mapping for each of the 14 consonants. The 9 sigmoid output units represent 9 control parameters of the formant synthesizer (ALF, F1, AI, F2, A2, F3, A3, AHF, V). The voicing parameter is required since consonant sounds have different degrees of voicing. The inputs are the same as for the manager V Ie network. Training and test data for the consonant network is obtained from the user . Target data is created for each of the 14 consonant sounds using the text-to-speech synthesizer . The scheme to collect data for a single consonant is: 1. The target consonant is played for 100 msec through the speech synthesizer; 2. the user forms a hand configuration corresponding to the consonant; 3. the user depresses the foot pedal to begin recording; the start of recording is indicated by the appearance of a green square; 4. 10-15 time steps of hand data are collected and stored with the corresponding formant parameter targets and phoneme identifier; the end of data collection is indicated by turning the green square red; 5. the user chooses whether to save the data to a file, and whether to redo the current target or move to the next one. Glove- Talkll 849 Using this procedure 350 approximants , 1510 fricatives and 700 nasals were collected and scaled for the training data. The hand data were averaged for each consonant sound to form the RBF centres. For the test data, 255 approximants, 960 fricatives and 165 nasals were collected and scaled . The RBF variances were set to 0.05. The mean square error on the training set was 0.005 and on the testing set was 0.01 corresponding to expected errors of 3.3% and 4.7% in the formant parameters, respectively. Listening to the output of the network reveals that each sound is produced reasonably well when the user's hand is held in a fixed position. The only difficulty is that the Rand L sounds are very sensitive to motion of the index finger. 3 Qualitative Performance of Glove-TalkII One subject, who is an accomplished pianist, has been trained extensively to speak with Glove-TalkII. We expected that his pianistic skill in forming finger patterns and his musical training would help him learn to speak with Glove-TalkII. After 100 hours of training, his speech with Glove-TalklI is intelligible and somewhat natural-sounding. He still finds it difficult to speak quickly, pronounce polysyllabic words, and speak spontaneously. During his training, Glove-TalkII also adapted to suit changes required by the subject . Initially, good performance of the VIC network is critical for the user to learn to speak. If the VIe network performs poorly the user hears a mixture of vowel and consonant sounds making it difficult to adjust his hand configurations to say different utterances . For this reason, it is important to have the user comfortable with the initial mapping so that the training data collected leads to the VIC network performing well. In the 100 hours of practice, Glove-Talk II was retrained about 10 times . Four significant changes were made from the original system analysed here for the new subject. First, the NG sound was added to the non-stop consonant list by adding an additional hand shape, namely the user touches his pinkie to his thumb on his right hand. To accomodate this change, the consonant and VIC network had two inputs added to represent the two flex angles of the pinkie. Also, the consonant network has an extra hidden unit for the NG sound. Second, the consonant network was trained to allow the RBF centres to change. After the hidden-to-output weights were trained until little improvement was seen, the input-to-hidden weights (i.e. the RBF centres) were also allowed to adapt . This noticeably improved performance for the user . Third, the vowel mapping was altered so that the I was moved closer to the EE sound and the entire mapping was reduced to 75% of its size. Fourth, for this subject, the VIC network needed was a 10-10-1 feed-forward sigmoid unit network. Understanding the interaction between the user's adaptation and Glove-TalkII's adaptation remains an interesting research pursuit. 4 Summary The initial mapping is loosely based on an articulatory model of speech. An open configuration of the hand corresponds to an unobstructed vocal tract, which in turn generates vowel sounds. Different vowel sounds are produced by movements of the hand in a horizontal X-Y plane that corresponds to movements of the first two formants which are roughly related to tongue position . Consonants other than stops are produced by closing the index, middle, or ring fingers or flexing the thumb, representing constrictions in the vocal tract . Stop consonants are produced by 850 S. Sidney Fels, Geoffrey Hinton contact switches worn on the user's left hand. FO is controlled by hand height and speaking intensity by foot pedal depression. Glove-TaikII learns the user's interpretation of this initial mapping. The VIC network and the consonant network learn the mapping from examples generated by the user during phases of training. The vowel network is trained on examples computed from the user-defined mapping between hand-position and vowels. The FO and volume mappings are non-adaptive. One subject was trained to use Glove-TalkII. After 100 hours of practice he is able to speak intelligibly. His speech is fairly slow (1.5 to 3 times slower than normal speech) and somewhat robotic. It sounds similar to speech produced with a text-tospeech synthesizer but has a more natural intonation contour which greatly improves the intelligibility and naturalness of the speech. Reading novel passages intelligibly usually requires several attempts, especially with polysyllabic words. Intelligible spontaneous speech is possible but difficult. Acknowledgements We thank Peter Dayan, Sageev Oore and Mike Revow for their contributions. This research was funded by the Institute for Robotics and Intelligent Systems and NSERC. Geoffrey Hinton is the Noranda fellow of the Canadian Institute for Advanced Research. References [1] A. G. Bell. Making a talking-machine. In Beinn Bhreagh Recorder, pages 61-72, November 1909~ [2] D. Broomhead and D. Lowe. Multivariable functional interpolation and adaptive networks. Complex Systems, 2:321-355, 1988. [3] Homer Dudley, R. R. Riesz, and S. S. A. Watkins. A synthetic speaker. Journal of the Franklin Institute, 227(6):739-764, June 1939. [4] S. S. Fels. Building adaptive interfaces using neural networks: The Glove-Talk pilot study. Technical Report CRG-TR-90-1, University of Toronto, 1990. [5] P. Ladefoged. A course in Phonetics (2 ed.). Harcourt Brace Javanovich, New York, 1982. [6] E. Lewis. A 'C' implementation of the JSRU text-to-speech system. Technical report, Computer Science Dept., University of Bristol, 1989. [7] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Learning internal representations by back-propagating errors. Nature, 323:533-536, 1986. [8] J. M. Rye and J. N. Holmes. A versatile software parallel-formant speech synthesizer. Technical Report JSRU-RR-1016, Joint Speech Research Unit, Malvern, UK, 1982. [9] Wolfgang Ritter von Kempelen. Mechanismus der menschlichen Sprache nebst Beschreibungeiner sprechenden Maschine. Mit einer Einleitung vonHerbert E. Brekle und Wolfgang Wild. Stuttgart-Bad Cannstatt F. Frommann, Stuttgart, 1970.
892 |@word middle:4 version:1 open:2 tr:1 versatile:1 reduction:1 initial:5 configuration:14 franklin:1 current:2 analysed:1 activation:1 synthesizer:15 must:1 additive:3 shape:2 device:3 plane:2 beginning:1 flexing:1 provides:2 quantized:1 toronto:7 height:3 along:1 direct:1 qualitative:2 consists:2 sidney:5 wild:1 manner:1 speaks:1 indeed:1 expected:4 rapid:1 magnifies:1 roughly:1 formants:2 manager:1 little:1 begin:2 provided:1 what:1 cm:3 developed:1 homer:1 nj:2 quantitative:2 every:3 fellow:1 act:1 scaled:4 uk:1 control:14 unit:21 producing:2 comfortable:1 vie:2 timing:2 interpolation:2 approximately:1 suggests:1 collect:1 co:1 range:4 averaged:1 decided:1 pronounce:1 spontaneously:1 wrist:1 flex:2 testing:1 block:1 implement:4 practice:2 procedure:2 bell:1 significantly:1 word:4 radial:2 vocal:6 gestural:1 operator:1 map:1 demonstrated:1 quick:1 musician:1 worn:1 williams:1 duration:1 holmes:1 his:11 variation:1 coordinate:5 feel:1 controlling:1 play:1 target:6 user:42 speak:8 spontaneous:1 us:1 origin:1 rumelhart:1 coarser:1 labeled:1 mike:1 intends:1 movement:4 tospeech:1 substantial:1 predictable:1 und:1 ideally:1 dynamic:1 trained:14 overshoot:1 division:2 f2:4 basis:3 joint:2 various:1 finger:7 talk:3 fast:1 artificial:3 quite:2 whose:1 say:2 formant:14 la4:2 rr:1 interaction:1 adaptation:2 j2:1 poorly:1 description:1 moved:1 produce:12 tract:5 ring:3 help:1 itch:1 propagating:1 measured:4 dividing:1 auxiliary:1 implemented:1 come:1 riesz:1 foot:8 closely:1 correct:1 centered:1 translating:1 noticeably:1 require:1 f1:3 summation:1 designate:1 crg:1 tracker:1 normal:2 mapping:23 alf:3 a2:3 currently:1 sensitive:1 him:2 create:1 mit:1 hello:1 fricative:3 june:1 she:3 articulator:1 improvement:1 indicates:1 abduction:2 greatly:1 contrast:2 dayan:1 entire:1 initially:1 her:1 hidden:5 manipulating:1 einer:1 priori:1 fairly:2 once:1 f3:3 polysyllabic:2 ng:3 identical:1 yaw:2 report:3 intelligent:1 cardinal:4 individual:1 phase:1 vowel:44 suit:1 attempt:1 adjust:1 mixture:2 sh:1 held:1 articulatory:6 predefined:1 emit:1 closer:1 necessary:2 xy:1 loosely:1 tongue:4 approximants:3 uniform:2 too:1 reported:1 stored:1 constriction:1 synthetic:1 chooses:1 fundamental:4 ie:1 ritter:1 continuously:2 quickly:2 squared:2 von:1 tube:1 slowly:1 expert:1 centred:1 includes:1 explicitly:1 depends:1 performed:1 lowe:1 wolfgang:2 red:1 start:1 parallel:6 depresses:3 rbfs:1 contribution:1 square:3 formed:1 il:1 roll:1 musical:2 who:2 phoneme:7 likewise:1 spaced:1 sitting:1 variance:3 correspond:1 thumb:7 accurately:1 produced:11 trajectory:1 drive:2 m5s:2 bristol:1 history:1 fo:3 ed:1 frequency:7 static:4 stop:7 pilot:1 proved:1 adjusting:1 broomhead:1 improves:1 amplitude:4 back:2 feed:4 improved:1 rand:1 stuttgart:2 though:1 until:1 hand:38 horizontal:2 harcourt:1 touch:1 defines:1 quality:3 indicated:2 building:1 requiring:1 normalized:5 ll:1 during:3 speaker:1 multivariable:1 complete:1 performs:2 motion:3 interface:6 passage:1 phonetics:1 novel:1 fi:1 sigmoid:4 rotation:2 unobstructed:1 functional:1 overview:2 volume:4 he:4 interpretation:2 slight:1 significant:1 ai:4 session:1 closing:1 centre:5 had:3 funded:1 something:1 closest:1 perspective:1 driven:1 keyboard:2 phonetic:1 accomplished:1 der:1 muscle:1 wji:2 seen:1 additional:1 somewhat:2 floor:1 converting:1 ii:2 sound:35 smooth:1 technical:3 adapt:1 gesture:15 long:1 lin:1 divided:1 recorder:1 a1:1 controlled:2 oore:1 pitch:5 sageev:1 simplistic:1 normalization:3 represent:2 robotics:1 addition:1 want:2 decreased:1 interval:2 diagram:1 fllled:2 extra:1 brace:1 file:1 subject:7 sounding:2 recording:2 slop:1 ee:1 granularity:4 canadian:1 switch:1 bandwidth:2 translates:1 listening:1 whether:3 redo:1 peter:1 speech:45 speaking:6 nine:2 interpolates:1 york:1 depression:1 useful:1 nasal:4 amount:1 ten:1 extensively:1 reduced:1 per:1 group:1 key:1 four:1 neither:1 year:1 convert:1 angle:4 you:1 respond:1 fourth:1 place:1 decide:1 doc:1 decision:2 bit:1 opposition:1 syllable:1 played:1 correspondence:1 activity:1 adapted:1 software:1 unlimited:2 dominated:1 generates:1 aspirant:1 performing:1 jaw:1 department:2 according:1 poor:1 conjugate:1 smaller:1 suppressed:1 skull:1 making:3 intuitively:1 rubber:1 remains:1 turn:1 needed:1 avt:5 initiate:1 instrument:1 end:1 unusual:1 available:1 pursuit:1 naturalness:1 actuator:1 voicing:4 appropriate:1 intelligibility:1 dudley:1 save:1 gate:1 slower:1 original:1 include:2 music:1 especially:2 contact:2 move:3 added:4 spelling:1 gradient:1 link:1 mapped:3 thank:1 pedal:7 collected:8 reason:1 ahf:3 index:4 relationship:1 providing:1 difficult:4 fe:2 implementation:1 descent:1 november:1 hinton:8 varied:1 sharp:1 retrained:1 subtasks:1 intensity:1 namely:1 required:2 specified:2 learned:1 hour:4 able:1 usually:1 pattern:1 articulation:1 reading:1 including:1 oj:2 green:2 critical:1 natural:3 difficulty:1 turning:1 advanced:1 representing:3 scheme:3 altered:1 vic:9 identifies:1 created:3 extract:1 utterance:1 hears:1 pianist:1 text:8 understanding:1 acknowledgement:1 zh:1 relative:1 interesting:2 geoffrey:6 generator:4 degree:2 unscaled:1 production:2 polhemus:2 jsru:2 course:1 summary:1 verbal:1 allow:1 institute:3 distributed:1 boundary:1 vocabulary:4 world:1 transition:4 contour:1 forward:4 made:2 adaptive:5 collection:1 far:1 approximate:1 skill:1 decides:1 reveals:1 robotic:1 consonant:48 noranda:1 spectrum:2 continuous:1 un:1 search:1 table:3 reality:1 learn:5 reasonably:1 nature:1 mse:1 complex:1 linearly:1 whole:2 noise:6 intelligible:2 identifier:1 fair:1 resonator:1 allowed:1 malvern:1 fashion:1 slow:1 position:10 msec:3 bandpass:1 intonation:1 perceptual:1 watkins:1 late:1 third:3 learns:1 bad:1 gating:2 list:1 a3:3 adding:1 perceptually:1 illustrates:1 accomodate:1 appearance:1 forming:1 nserc:1 talking:1 ch:1 corresponds:4 lewis:1 dh:1 rbf:14 revow:1 hard:1 change:5 glove:25 determined:4 typical:2 uniformly:1 bellow:1 called:1 total:1 internal:1 dept:1
7,127
893
A Growing Neural Gas Network Learns Topologies Bernd Fritzke Institut fur Neuroinformatik Ruhr-Universitat Bochum D-44 780 Bochum Germany Abstract An incremental network model is introduced which is able to learn the important topological relations in a given set of input vectors by means of a simple Hebb-like learning rule. In contrast to previous approaches like the "neural gas" method of Martinetz and Schulten (1991, 1994), this model has no parameters which change over time and is able to continue learning, adding units and connections, until a performance criterion has been met. Applications of the model include vector quantization, clustering, and interpolation. 1 INTRODUCTION In unsupervised learning settings only input data is available but no information on the desired output. What can the goal of learning be in this situation? One possible objective is dimensionality reduction: finding a low-dimensional subspace of the input vector space containing most or all of the input data. Linear subspaces with this property can be computed directly by principal component analysis or iteratively with a number of network models (Sanger, 1989; Oja, 1982). The Kohonen feature map (Kohonen, 1982) and the "growing cell structures" (Fritzke, 1994b) allow projection onto non-linear, discretely sampled subspaces of a dimensionality which has to be chosen a priori. Depending on the relation between inherent data dimensionality and dimensionality of the target space, some information on the topological arrangement of the input data may be lost in the process. 626 Bernd Fritzke This is not astonishing since a reversible mapping from high-dimensional data to lower-dimensional spaces (or structures) does not exist in general. Asking how structures must look like to allow reversible mappings directly leads to another possible objective of unsupervised learning which can be described as topology learning: Given some high-dimensional data distributionp(e), find a topological structure which closely reflects the topology of the data distribution. An elegant method to construct such structures is "competitive Hebbian learning" (CHL) (Martinetz, 1993). CHL requires the use of some vector quantization method. Martinetz and Schulten propose the "neural gas" (NG) method for this purpose (Martinetz and Schulten, 1991). We will briefly introduce and discuss the approach of Martinetz and Schulten. Then we propose a new network model which also makes use of CHL. In contrast to the above-mentioned CHL/NG combination, this model is incremental and has only constant parameters. This leads to a number of advantages over the previous approach. 2 COMPETITIVE HEBBIAN LEARNING AND NEURAL GAS CHL (Martinetz, 1993) assumes a number of centers in R n and successively inserts topological connections among them by evaluating input signals drawn from a data distribution p(e). The principle of this method is: For each input signal x connect the two closest centers (measured by Euclidean distance) by an edge. The resulting graph is a subgraph of the Delaunay triangulation (fig. 1a) corresponding to the set of centers. This subgraph (fig. 1b), which is called the "induced Delaunay triangulation", is limited to those areas of the input space R n where p(e? O. The "induced Delaunay triangulation" has been shown to optimally preserve topology in a very general sense (Martinetz, 1993). Only centers lying on the input data submanifold or in its vicinity actually develop any edges. The others are useless for the purpose of topology learning and are often called dead units. To make use of all centers they have to be placed in those regions of R n where P (e) differs from zero. This could be done by any vector quantization (VQ) procedure. Martinetz and Schulten have proposed a particular kind of VQ method, the mentioned NG method (Martinetz and Schulten, 1991). The main principle of NG is the following: For each input signal x adapt the k nearest centers whereby k is decreasing from a large initial to a small final value. A large initial value of k causes adaptation (movement towards the input signal) of a large fraction of the centers. Then k (the adaptation range) is decreased until finally only the nearest center for each input signal is adapted. The adaptation strength underlies a similar decay schedule. To realize the parameter decay one has to define the total number of adaptation steps for the NG method in advance. A Growing Neural Gas Network Learns Topologies a) Delaunay triangulation 627 b) induced Delaunay triangulation Figure 1: Two ways of defining closeness among a set of points. a) The Delaunay triangulation (thick lines) connects points having neighboring Voronoi polygons (thin lines). Basically this reduces to points having small Euclidean distance w.r.t. the given set of points. b) The induced Delaunay triangulation (thick lines) is obtained by masking the original Delaunay triangulation with a data distribution P(~) (shaded) . Two centers are only connected if the common border of their Voronoi polygons lies at least partially in a region where P(~? 0 (closely adapted from Martinetz and Schulten, 1994) For a given data distribution one could now first run the NG algorithm to distribute a certain number of centers and then use CHL to generate the topology. It is, however, also possible to apply both techniques concurrently (Martinetz and Schulten, 1991). In this case a method for removing obsolete edges is required since the motion of the centers may make edges invalid which have been generated earlier. Martinetz and Schulten use an edge aging scheme for this purpose. One should note that the CHL algorithm does not influence the outcome of the NG method in any way since the adaptations in NG are based only on distance in input space and not on the network topology. On the other hand NG does influence the topology generated by CHL since it moves the centers around. The combination of NG and CHL described above is an effective method for topology learning. A problem in practical applications, however, may be to determine a priori a suitable number of centers. Depending on the complexity of the data distribution which one wants to model, very different numbers of centers may be appropriate. The nature of the NG algorithm requires a decision in advance and, if the result is not satisfying, one or several new simulations have to be performed from scratch. In the following we propose a method which overcomes this problem and offers a number of other advantages through a flexible scheme for center insertion. Bernd Fritzke 628 3 THE GROWING NEURAL GAS ALGORITHM In the following we consider networks consisting of ? a set A of units (or nodes). Each unit c E A has an associated reference vector We E Rn. The reference vectors can be regarded as positions in input space of the corresponding units. ? a set N of connections (or edges) among pairs of units. These connections are not weighted. Their sole purpose is the definition of topological structure. Moreover, there is a (possibly infinite) number of n-dimensional input signals obeying some unknown probability density function P(~). The main idea of the method is to successively add new units to an initially small network by evaluating local statistical measures gathered during previous adaptation steps. This is the same approach as used in the "growing cell structures" model (Fritzke, 1994b) which, however, has a topology with a fixed dimensionality (e.g., two or three). In the approach described here, the network topology is generated incrementally by CHL and has a dimensionality which depends on the input data and may vary locally. The complete algorithm for our model which we call "growing neural gas" is given by the following: o. Start with two units a and b at random positions Wa and Wb in Rn. 1. Generate an input signal ~ according to P(~). 2. Find the nearest unit 81 and the second-nearest unit 82. 3. Increment the age of all edges emanating from 81. 4. Add the squared distance between the input signal and the nearest unit in input space to a local counter variable: Aerror(8t} = IIWSl - ell 2 5. Move 81 and its direct topological neighbors1 towards ~ by fractions Eb and En, respectively, of the total distance: AW s1 = Eb(e - AWn = En(~ - w n ) W S1 ) for all direct neighbors n of 81 6. If 81 and 82 are connected by an edge, set the age of this edge to zero. If such an edge does not exist, create it. 2 7. Remove edges with an age larger than a maz ? If this results in points having no emanating edges, remove them as well. IThroughout this paper the term neighbors denotes units which are topological neighbors in the graph (as opposed to units within a small Euclidean distance of each other in input space). 2This step is Hebbian in its spirit since correlated activity is used to decide upon insertions. A Growing Neural Gas Network Learns Topologies 629 8. If the number of input signals generated so far is an integer multiple of a parameter A, insert a new unit as follows: ? Determine the unit q with the maximum accumulated error. ? Insert a new unit r halfway between q and its neighbor f with the largest error variable: Wr = 0.5 (w q + wf)' ? Insert edges connecting the new unit r with units q and f, and remove the original edge between q and f. ? Decrease the error variables of q and f by multiplying them with a constant 0:. Initialize the error variable of r with the new value of the error variable of q. 9. Decrease all error variables by multiplying them with a constant d. 10. If a stopping criterion (e.g., net size or some performance measure) is not yet fulfilled go to step 1. How does the described method work? The adaptation steps towards the input signals (5.) lead to a general movement of all units towards those areas of the input space where signals come from (P(~) > 0). The insertion of edges (6.) between the nearest and the second-nearest unit with respect to an input signal generates a single connection of the "induced Delaunay triangulation" (see fig. 1b) with respect to the current position of all units. The removal of edges (7.) is necessary to get rid of those edges which are no longer part of the "induced Delaunay triangulation" because their ending points have moved. This is achieved by local edge aging (3.) around the nearest unit combined with age re-setting of those edges (6.) which already exist between nearest and second-nearest units. With insertion and removal of edges the model tries to construct and then track the "induced Delaunay triangulation" which is a slowly moving target due to the adaptation of the reference vectors. The accumulation of squared distances (4.) during the adaptation helps to identify units lying in areas of the input space where the mapping from signals to units causes much error. To reduce this error, new units are inserted in such regions. 4 SIMULATION RESULTS We will now give some simulation results to demonstrate the general behavior of our model. The probability distribution in fig. 2 has been proposed by Martinetz and Schulten (1991) to demonstrate the non-incremental "neural gas" model. It can be seen that our model quickly learns the important topological relations in this rather complicated distribution by forming structures of different dimensionalities. The second example (fig. 3) illustrates the differences between the proposed model and the original NG network. Although the final topology is rather similar for both models, intermediate stages are quite different. Both models are able to identify the clusters in the given distribution. Only the "growing neural gas" model, however, Bernd Fritzke 630 Figure 2: The "growing neural gas" network adapts to a signal distribution which has different dimensionalities in different areas of the input space. Shown are the initial network consisting of two randomly placed units and the networks after 600, 1800, 5000, 15000 and 20000 input signals have been applied. The last network shown is not the necessarily the "final" one since the growth process could in principle be continued indefinitely. The parameters for this simulation were: A = 100, Eb = 0.2, En = 0.006, a = 0.5, a maz = 50, d = 0.995. could continue to grow to discover still smaller clusters (which are not present in this particular example, though). 5 DISCUSSION The "growing neural gas" network presented here is able to make explicit the important topological relations in a given distribution of input signals. An advantage over the NG method of Martinetz and Schulten is the incremental character of the model which eliminates the need to pre-specify a network size. Instead, the growth process can be continued until a user-defined performance criterion or network size is met. All parameters are constant over time in contrast to other models which heavily rely on decaying parameters (such as the NG method or the Kohonen feature map). pee) It should be noted that the topology generated by CHL is not an optional feature A Growing Neural Gas Network Learns Topologies "neural gas" and "competitive Hebbian learning" o 0 0 0 00 631 "growing neural gas" (uses "competitive Hebbian learning") 0 o~oo ?~~r~:;-.;. {]J'. 000 090- o V o 0 ~ 0 o j 00 co 00 000 8 0 Figure 3: The NG/CHL network of Martinetz and Schulten (1991) and the author's "growing neural gas" model adapt to a clustered probability distribution. Shown are the respective initial states (top row) and a number of intermediate stages. Both the number of units in the NG model and the final number of units in the "growing neural gas" model are 100. The bottom row shows the distribution of centers after 10000 adaptation steps (the edges are as in the previous row but not shown). The center distribution is rather similar for both models although the intermediate stages differ significantly. 632 Bernd Fritzke of our method (as it is for the NG model) but an essential component since it is used to direct the (completely local) adaptation as well as insertion of centers. It is probably the proper initialization of new units by interpolation from existing ones which makes it possible to have only constant parameters and local adaptations. Possible applications of our model are clustering (as shown) and vector quantization. The network should perform particularly well in situations where the neighborhood information (in the edges) is used to implement interpolation schemes between neighboring units. By using the error occuring in early phases it can be determined where to insert new units to generate a topological look-up table of different density and different dimensionality in particular areas of the input data space. Another promising direction of research is the combination with supervised learning. This has been done earlier with the "growing cell structures" (Fritzke, 1994c) and recently also with the "growing neural gas" described in this paper (Fritzke, 1994a). A crucial property for this kind of application is the possibility to choose an arbitrary insertion criterion. This is a feature not present, e.g., in the original "growing neural gas". The first results of this new supervised network model, an incremental radial basis function network, are very promising and we are further investigating this currently. References Fritzke, B. (1994a). Fast learning with incremental rbf networks. Neural Processing Letters, 1(1):2-5. Fritzke, B. (1994b). Growing cell structures - a self-organizing network for unsupervised and supervised learning. Neural Networks, 7(9):1441-1460. Fritzke, B. (1994c). Supervised learning with growing cell structures. In Cowan, J., Tesauro, G., and Alspector, J., editors, Advances in Neural Information Processing Systems 6, pages 255-262. Morgan Kaufmann Publishers, San Mateo, CA. Kohonen, T. (1982). Self-organized formation of topologically correct feature maps. Biological Cybernetics, 43:59-69. Martinetz, T. M. (1993). Competitive Hebbian learning rule forms perfectly topology preserving maps. In ICANN'93: International Conference on Artificial Neural Networks, pages 427-434, Amsterdam. Springer. Martinetz, T. M. and Schulten, K J. (1991). A "neural-gas" network learns topologies. In Kohonen, T., Makisara, K, Simula, 0., and Kangas, J., editors, Artificial Neural Networks, pages 397-402. North-Holland, Amsterdam. Martinetz, T. M. and Schulten, K J. (1994). Topology representing networks. Neural Networks, 7(3):507-522. Oja, E. (1982). A simplified neuron model as a principal component analyzer. Journal of Mathematical Biology, 15:267-273. Sanger, T. D. (1989). An optimality principle for unsupervised learning. In Touretzky, D. S., editor, Advances in Neural Information Processing Systems 1, pages 11-19. Morgan Kaufmann, San Mateo, CA.
893 |@word maz:2 briefly:1 ruhr:1 simulation:4 reduction:1 initial:4 neighbors1:1 existing:1 current:1 yet:1 must:1 realize:1 remove:3 obsolete:1 indefinitely:1 node:1 mathematical:1 direct:3 introduce:1 alspector:1 behavior:1 growing:19 decreasing:1 discover:1 moreover:1 what:1 kind:2 finding:1 growth:2 unit:32 local:5 aging:2 interpolation:3 eb:3 initialization:1 mateo:2 shaded:1 co:1 limited:1 range:1 practical:1 lost:1 implement:1 differs:1 procedure:1 area:5 significantly:1 projection:1 pre:1 radial:1 bochum:2 get:1 onto:1 influence:2 accumulation:1 map:4 center:18 go:1 rule:2 continued:2 regarded:1 increment:1 target:2 heavily:1 user:1 us:1 simula:1 satisfying:1 particularly:1 bottom:1 inserted:1 region:3 connected:2 movement:2 counter:1 decrease:2 mentioned:2 complexity:1 insertion:6 astonishing:1 upon:1 completely:1 basis:1 polygon:2 fast:1 effective:1 emanating:2 artificial:2 formation:1 outcome:1 neighborhood:1 ithroughout:1 quite:1 larger:1 final:4 advantage:3 net:1 propose:3 adaptation:12 kohonen:5 neighboring:2 organizing:1 subgraph:2 adapts:1 moved:1 chl:12 cluster:2 incremental:6 help:1 depending:2 develop:1 oo:1 measured:1 nearest:10 sole:1 come:1 met:2 differ:1 direction:1 thick:2 closely:2 correct:1 clustered:1 biological:1 insert:5 lying:2 around:2 mapping:3 vary:1 early:1 purpose:4 pee:1 currently:1 largest:1 create:1 reflects:1 weighted:1 concurrently:1 rather:3 fur:1 contrast:3 sense:1 wf:1 voronoi:2 stopping:1 accumulated:1 initially:1 relation:4 germany:1 among:3 flexible:1 priori:2 ell:1 initialize:1 construct:2 having:3 ng:17 makisara:1 biology:1 look:2 unsupervised:4 thin:1 others:1 inherent:1 randomly:1 oja:2 preserve:1 phase:1 connects:1 consisting:2 possibility:1 edge:22 necessary:1 respective:1 institut:1 euclidean:3 desired:1 re:1 earlier:2 wb:1 asking:1 submanifold:1 universitat:1 optimally:1 connect:1 aw:1 combined:1 density:2 international:1 connecting:1 quickly:1 squared:2 successively:2 containing:1 opposed:1 possibly:1 slowly:1 choose:1 dead:1 distribute:1 north:1 depends:1 performed:1 try:1 competitive:5 start:1 decaying:1 complicated:1 masking:1 kaufmann:2 gathered:1 identify:2 basically:1 multiplying:2 cybernetics:1 touretzky:1 definition:1 associated:1 sampled:1 dimensionality:9 organized:1 schedule:1 actually:1 supervised:4 specify:1 done:2 though:1 stage:3 until:3 hand:1 reversible:2 incrementally:1 vicinity:1 iteratively:1 during:2 self:2 whereby:1 noted:1 criterion:4 complete:1 demonstrate:2 occuring:1 motion:1 recently:1 common:1 analyzer:1 moving:1 longer:1 add:2 delaunay:11 closest:1 triangulation:11 tesauro:1 certain:1 continue:2 seen:1 morgan:2 preserving:1 determine:2 signal:16 multiple:1 reduces:1 hebbian:6 adapt:2 offer:1 underlies:1 achieved:1 cell:5 want:1 decreased:1 grow:1 crucial:1 publisher:1 eliminates:1 martinetz:18 probably:1 induced:7 elegant:1 cowan:1 spirit:1 call:1 integer:1 fritzke:12 intermediate:3 topology:19 perfectly:1 reduce:1 idea:1 cause:2 locally:1 generate:3 exist:3 fulfilled:1 wr:1 track:1 drawn:1 graph:2 fraction:2 halfway:1 run:1 letter:1 topologically:1 decide:1 decision:1 topological:10 discretely:1 activity:1 adapted:2 strength:1 awn:1 generates:1 optimality:1 according:1 combination:3 smaller:1 character:1 s1:2 vq:2 discus:1 available:1 apply:1 appropriate:1 original:4 denotes:1 top:1 assumes:1 clustering:2 include:1 sanger:2 objective:2 move:2 arrangement:1 already:1 subspace:3 distance:7 useless:1 proper:1 unknown:1 perform:1 neuron:1 gas:20 optional:1 situation:2 defining:1 rn:2 kangas:1 arbitrary:1 neuroinformatik:1 introduced:1 bernd:5 required:1 pair:1 connection:5 able:4 suitable:1 rely:1 representing:1 scheme:3 removal:2 age:4 principle:4 editor:3 row:3 placed:2 last:1 allow:2 neighbor:4 evaluating:2 ending:1 author:1 san:2 simplified:1 far:1 overcomes:1 investigating:1 rid:1 table:1 distributionp:1 promising:2 learn:1 nature:1 ca:2 necessarily:1 icann:1 main:2 border:1 fig:5 en:3 hebb:1 position:3 schulten:14 explicit:1 obeying:1 lie:1 learns:6 removing:1 decay:2 closeness:1 essential:1 quantization:4 adding:1 illustrates:1 forming:1 amsterdam:2 partially:1 holland:1 springer:1 goal:1 invalid:1 towards:4 rbf:1 change:1 infinite:1 determined:1 principal:2 called:2 total:2 scratch:1 correlated:1
7,128
894
JPMAX: Learning to Recognize Moving Objects as a Model-fitting Problem Suzanna Becker Department of Psychology, McMaster University Hamilton, Onto L8S 4K1 Abstract Unsupervised learning procedures have been successful at low-level feature extraction and preprocessing of raw sensor data. So far, however, they have had limited success in learning higher-order representations, e.g., of objects in visual images. A promising approach is to maximize some measure of agreement between the outputs of two groups of units which receive inputs physically separated in space, time or modality, as in (Becker and Hinton, 1992; Becker, 1993; de Sa, 1993). Using the same approach, a much simpler learning procedure is proposed here which discovers features in a single-layer network consisting of several populations of units, and can be applied to multi-layer networks trained one layer at a time. When trained with this algorithm on image sequences of moving geometric objects a two-layer network can learn to perform accurate position-invariant object classification. 1 LEARNING COHERENT CLASSIFICATIONS A powerful constraint in sensory data is coherence over time, in space, and across different sensory modalities. An unsupervised learning procedure which can capitalize on these constraints may be able to explain much of perceptual self-organization in the mammalian brain. The problem is to derive an appropriate cost function for unsupervised learning which will capture coherence constraints in sensory signals; we would also like it to be applicable to multi-layer nets to train hidden as well as output layers. Our ultimate goal is for the network to discover natural object classes based on these coherence assumptions. 934 1.1 Suzanna Becker PREVIOUS WORK Successive images in continuous visual input are usually views of the same object; thus, although the image pixels may change considerably from frame to frame, the image usually can be described by a small set of consistent object descriptors, or lower-level feature descriptors. We refer to this type of continuity as temporal coherence. This sort of structure is ubiquitous in sensory signals, from vision as well as other senses, and can be used by a neural network to derive temporally coherent classifications. This idea has been used, for example, in temporal versions of the Hebbian learning rule to associate items over time (Weinshall, Edelman and Biilt hoff, 1990; FOldiak, 1991). To capitalize on temporal coherence for higher-order feature extraction and classification, we need a more powerful learning principle. A promising approach is to maximize some measure of agreement between the outputs of two groups of units which receive inputs physically separated in space, time or modality, as in (Becker and Hinton, 1992; Becker, 1993; de Sa, 1993). This forces the units to extract features which are coherent across the different input sources. Becker and Hinton's (1992) Imax algorithm maximizes the mutual information between the outputs of two modules, y~ and Yb, connected to different parts of the input, a and b. Becker (1993) extended this idea to the problem of classifying temporally varying patterns by applying the discrete case of the mutual information cost function to the outputs of a single module at successive time steps, y-;'(t) and y-;'(t + 1). However, the success of this method relied upon the back-propagation of derivatives to train the hidden layer and it was found to be extremely susceptible to local optima. de Sa's method (1993) is closely related, and minimizes the probability of disagreement between output classifications, y~(t) and y1(t), produced by two modules having different inputs, e.g., from different sensory modalities. The success of this method hinges upon bootstrapping the first layer by initializing the weights to randomly selected training patterns, so this method too is susceptible to the problem of local optima. If we had a more flexible cost function that could be applied to a multi-layer network, first to each hidden layer in turn, and finally to the~utput layer for classification, so that the two layers could discover genuinely different structure, we might be able to overcome the problem of getting trapped in local optima, yielding a more powerful and efficient learning procedure. We can analyze the optimal solutions for both de Sa's and Becker's cost functions (see Figure 1 a) and see that both cost functions are maximized by having perfect one-to-one agreement between the two groups of units over all cases, using a one-of-n encoding, i.e., having only a single output unit on for each case. A major limitation of these methods is that they strive for perfect classifications by the units. While this is desirable at the top layer of a network, it is an unsuitable goal for training intermediate layers to detect low-level features. For example, features like oriented edges would not be perfect predictors across spatially or temporally nearby image patches in images of translating and rotating objects. Instead, we might expect that an oriented edge at one location would predict a small range of similar orientations at nearby locations. So we would prefer a cost function whose optimal solution was more like those shown in Figure 1 b) or c). This would allow a feature i in group a to agree with any of several nearby features, e.g. i - 1, i, or i + 1 in group b. JPMAX 935 b) a) ?? === ?.!. iii === ;;; ? ?11 II , ???? =???- ? I!!!!! ..... . c) , ?? == -== ?? ?? ? .1. II' ill!! II ? 11,? ? ? ????? ? .i. I' I' ??? = ? ? ? = ? 11 I!!!!! I ?? ;;; ? .. ?.'. .. ??.i. ??? ? II ? II II' I? ? 1' ? J[. , ? ??? .. :== I!!III'J!!'? ? ii ??? II "'-= , n ;;; . == ? Ilii Figure 1: Three possible joint distributions for the probability that the i th and j th units in two sets of m classification units are both on. White is high density, black is low density. The optimal joint distribution for Becker's and de Sa's algorithms is a matrix with all its density either in the diagonal as in a), or any subset of the diagonal entries for de Sa's method, or a permutation of the diagonal matrix for Becker's algorithm. Alternative distributions are shown in b) and c). 1.2 THE JPMAX ALGORITHM One way to achieve an arbitrary configuration of agreement over time between two groups of units (as in Figure 1 b) or c? is to treat the desired configuration as a prior joint probability distribution over their outputs. We can obtain the actual distribution by observing the temporal correlations between pairs of units' outputs in the two groups over an ensemble of patterns. We can then optimize the actual distribution to fit the prior. We now derive two different cost functions which achieve this result. Interestingly, they result in very similar learning rules. Suppose we have two groups of m units as shown in Figure 2 a), receiving inputs, x"7t and xi" from the same or nearby parts of the input image. Let Ca(t) and Cb(t) be the classifications of the two input patches produced by the network at time step t; the outputs of the two groups of units, y"7t(t) and yi,(t), represent these classification probabilities: enetoi(t) Yai(t) - P(Ca(t) = i) = Lj eneto;(t) Ybi(t) = enetbi (t) P(Cb(t) = i) = Lj enetb;(t) (1) (the usual "soft max" output function) where netai(t) and netbj(t) are the weighted net inputs to units. We could now observe the expected joint probability distribution qij = E [Yai(t)Ybj(t + 1)]t = E (P(Ca(t) = i, Cb(t + 1) = j)]t by computing the temporal covariances between the classification probabilities, averaged over the ensemble of training patterns; this joint probability is an m 2 -valued random variable. Given the above statistics, one possible cost function we could minimize is the -log probability of the observed temporal covariance between the two sets of units' outputs under some prior distribution (e.g. Figure 1 b) or c?. If we knew the actual frequency counts for each (joint) classification k = kll ,?? ., kIm, k21 , ... ,kmm, 936 Suzanna Becker b) At b'~-#--#----------"'-""""-"''''''''''''''''''''':iIo. <t (I) t Figure 2: a) Two groups of 15 units receive inputs from a 2D retina. The groups are able to observe each other's outputs across lateral links with unit time delays. b) A second layer of two groups of 3 units is added to the architecture in a). rather than just the observed joint probabilities, qij = E [~] , then given our prior model, pu, ... ,Pmm, we could compute the probability of the observations under a multinomial distribution: (2) Using the de Moivre-Laplace approximation leads to the following: P(k) ~ 1 exp v(27rn)m 2 -1 IL,j Pij (_! L 2 (k ij i,j - n Pij )2) (3) npij Taking the derivative of the - log probability wrt k ij leads to a very simple learning rule which depends only on the observed probabilities qij and priors Pij: fJ -logP(k) fJk ij ~ npij - k ij npij Pij - qij (4) Pij %!i: To obtain the final weight update rule, we just multiply this by n l . One problem with the above formulation is that the priors Pij must not be too close to zero for the de Moivre-Laplace approximation to hold. In practice, this cost function works well if we simply ignore the derivative terms where the priors are zero. An alternative cost function (as suggested by Peter Dayan) which works equally well is the Kullback-Liebler divergence or G-error between the desired joint probabilities Pij and the observed probabilities qij: G(p,q) = - LL j (Pij logpij - Pij lOgqij) (5) JPMAX 937 Figure 3: 10 of the 1500 tmining patterns: geometric objects centered in 36 possible locations on a 12-by-12 pixel grid. Object location varied mndomly between patterns. The derivative of G wrt qij is: aG Pij aqij qij (6) subject to Llij qij = 1 (enforced by the softmax output function). Note the similarity between the learning rules given by equations 4 and 6. 2 EXPERIMENTS The network shown in Figure 2 a) was trained to minimize equation 5 on an ensemble of pattern trajectories of circles, squares and triangles (see Figure 3) for five runs starting from different random initial weights, using a gradient-based learning method. For ten successive frames, the same object would appear, but with the centre varying randomly within the central six-by-six patch of pixels. In the last two frames, another randomly selected object would appear, so that object trajectories overlapped by two frames. These images are meant to be a crude approximation to what a moving observer would see while watching multiple moving objects in a scene; at any given time a single object would be approximately centered on the retina but its exact location would always be jittering from moment to moment. In these simulations, the prior distribution for the temporal covariances between the two groups of units' outputs was a block-diagonal configuration as in Figure 1 c), but with three five-by-five blocks along the diagonal. Our choice of a blockdiagonal prior distribution with three blocks encodes the constraint that units in a given block in one group should try to agree with units in the same block in the other group; so each group should discover three classes of features. The number of units within a block was varied in preliminary experiments, and five units was found to be a reasonable number to capture all instances of each object class (although the performance of the algorithm seemed to be robust with respect to the number of units per block). The learning took about 3000 iterations of steepest descent with 938 Suzanna Becker .... .. - - .. .. . ... ... . ... .... ?.... ... .. .. -. -.... .....-,..?. .......... .. . . -i?" .. . . . . . . .? ....?? ~~;. ? ? .. ? 1 ??. . . ? .. .. . . . .. .. . .... ...?? . . I'....... ..?. .... .? -1-?... . -.1 .... :1" . ....... ..... .' . ..... . ,'.=::::.: . ... ;~. :~:" i1 ? ? ? 1: . . . . . ' ~:. I' 'a_" _" ?? .. ... ? I .. ? 1 0; .... . ?? .. ? ??? _ _ _ ,. .. . . . ?? .. ? .a . . 1 ' 1 ? .. :" ~ . . ~i~" ._ , ' . ..? ::: ") ;: .. ?? ':. ?? " __ . ??? ? ?? ' .. .... .. . ? . ?1 ?????????? ~ .??..????? :.. ? ....... . ' .. .. ' e o_ .1. .... .... ?? . ? ? . ..... . .. ...... :: .. ? ?i . . . . . . . . . . . ???...? ~.~.: i ??? . _ _ .. ... .. I - ?1 ~::.... .; . ....... ????? II ........ ?????? ? . .. i " ' ;' :': ? ? ?:... a ? . :- .~:~ji----.. .........?:.......... ::: .;~ I .. . .; .. , ...... ?..... .... .. ?? a ???? .... ... . ..... .. ? .. a ? ? .... i??. .. .?. ??.. I?i....... I .. " ? . ?? :::' ":: . \~. ..... ... .. . ...... ... .. .. ......... . ....... .... .... ..... ? ? ? . .... .. ? .. ? ....... . Figure 4: Weights learned by one of the two groups of 15 units in the first layer. White weights are positive and black are negative. momentum to converge, but after about 1000 iterations only very small weight changes were made. Weights learned on a typical run for one of the two groups of fifteen units are shown in Figure 4. The units' weights are displayed in three rows corresponding to units in the three blocks in the block-diagonal joint prior matrix. Units within the same block each learned different instances of the same pattern class. For example, on this run units in the first block learned to detect circles in specific positions. Units in the second block tended to learn combinations of either horizontal or vertical lines, or sometimes mixtures of the two. In the third block, units learned blurred, roughly triangular shape detectors, which for this training set were adequate to respond specifically to triangles. In all five runs the network converged to equivalent solutions (only the groups' particular shape preferences varied across runs). Varying the number of units per block from three to five (Le. three three-by-three blocks versus three five-by-five blocks of units) produced similar results, except that with fewer units per block, each unit tended to capture multiple instances of a particular object class in different positions. A second layer of two groups of three units was added to the network, as shown in Figure 2 b). While keeping the first layer of weights frozen, this network was trained using exactly the same cost function as the first layer for about 30 iterations using a gradient-based learning method. This time the prior joint distribution for the two classifications was a three-by-three matrix with 80% of the density along the diagonal and 20% evenly distributed across the remainder of the distribution. Units in this layer learned to discriminate fairly well between the three object classes, as shown in Figure 5 a). On a test set with the ambiguous patterns removed (Le., patterns containing multiple objects), units in the second layer achieved very JPMAX 939 Loyer 2 Unit Responses on Trai ning Set a) I 1. ~. Circles Squares Triangles 0.80 .~ 0. 50 c ~ 1 0.40 '0 c i? 0.20 0.00 Loyer 2 Unit Responses on Test Set (no overlap s) b) 1.00 c . 0 . :;; 0.80 0 .; 0.60 i '0 0. 40 c 0 0 1?,20 0.00 LIllI ? ? UM3 .... ----. Unit 6 Figure 5: Response probabilities for the six output units to each of the three shapes. accurate object discrimination as shown in Figure 5 b). On ambiguous patterns containing multiple objects, the network's performance was disappointing. The output units would sometimes produce the "correct" response, i.e., all the units representing the shapes present in the image would be partially active. Most often, however, only one of the correct shapes would be detected, and occasionally the network's response indicated the wrong shape altogether. It was hoped? that the diagonally dominant prior mixed with a uniform density would allow units to occasionally disagree, and they would therefore be able to represent cases of multiple objects. It may have helped to use a similar prior for the hidden layer; however, this would increase the complexity of the learning considerably. 3 DISCUSSION We have shown that the algorithm can learn 2D-translation-invariant shape recognition, but it should handle equally well other types of transformations, such as rotation, scaling or even non-linear transformations. In principle, the algorithm should be applicable to real moving images; this is currently being investigated. Although we have focused here on the temporal coherence constraint, the algorithm could be applied equally well using other types of coherence, such as coherence 940 Suzanna Becker across space or across different sensory modalities. Note that the units in the first layer of the network did not learn anything about the geometric transformations between translated versions of the same object; they simply learned to associate different views together. In this respect, the representation learned at the hidden layer is similar to that predicted by the "privileged views" theory of viewpoint-invariant object recognition advocated by Weinshall et al. (1990) (and others). Their algorithm learns a similar representation in a single layer of competing units with temporal Hebbian learning applied to the lateral connections between these units. However, the algorithm proposed here goes further in that it can be applied to subsequent stages of learning to discover higher-order object classes. Yuille et al. (1994) have previously proposed an algorithm based on similar principles, which also involves maximizing the log probability of the network outputs under a prior; in one special case it is equivalent to Becker and Hinton's Imax algorithm. The algorithm proposed here differs substantially, in that we are dealing with the ensemble-averaged joint probabilities of two populations of units, and fitting this quantity to a prior; further, Yuille et aI's scheme employs back-propagation. One challenge for future work is to train a network with smaller receptive fields for the first layer units, on images of objects with common low-level features, such as squares and rectangles. At least three layers of weights would be required to solve this task: units in the first layer would have to learn local object parts such as corners, while units in the next layer could group parts into viewpoint-specific whole objects and in the top layer viewpoint-invariance, in principle, could be achieved. Acknowledgements Helpful comments from Geoff Hinton, Peter Dayan, Tony Plate and Chris Williams are gratefully acknowledged. References Becker, S. (1993). Learning to categorize objects using temporal coherence. In Advances in Neural Information Processing Systems 5, pages 361- 368. Morgan Kaufmann. Becker, S. and Hinton, G. E. (1992). A self-organizing neural network that discovers surfaces in random-dot stereograms. Nature, 355: 161-163. de Sa, V. R. (1993). Minimizing disagreement for self-supervised classification. In Proceedings of the 1993 Connectionist Models Summer School, pages 300-307. Lawrence Erlbaum associates. F51diak, P. (1991). Learning invariance from transformation sequences. Neural Computation, 3(2):194-200. Weinshall, D., Edelman, S., and Biilthoff, H. H. (1990). A self-organizing multiple-view representation of 3D objects. In Advances in Neural Information Processing Systems 2, pages 274-282. Morgan Kaufmann. Yuille, A. L., Stelios, M. S., and Xu, L. (1994). Bayesian Self-Organization. Technical Report No. 92-10, Harvard Robotics Laboratory.
894 |@word version:2 simulation:1 covariance:3 fifteen:1 moment:2 initial:1 configuration:3 interestingly:1 must:1 subsequent:1 shape:7 update:1 ilii:1 discrimination:1 selected:2 fewer:1 item:1 steepest:1 location:5 successive:3 preference:1 simpler:1 five:8 along:2 edelman:2 qij:8 fitting:2 expected:1 roughly:1 multi:3 brain:1 actual:3 discover:4 maximizes:1 what:1 weinshall:3 minimizes:1 substantially:1 ag:1 transformation:4 bootstrapping:1 temporal:10 exactly:1 wrong:1 unit:53 appear:2 hamilton:1 positive:1 local:4 treat:1 encoding:1 approximately:1 might:2 black:2 limited:1 range:1 averaged:2 practice:1 block:17 differs:1 procedure:4 onto:1 close:1 applying:1 optimize:1 equivalent:2 maximizing:1 go:1 williams:1 starting:1 focused:1 suzanna:5 rule:5 imax:2 population:2 handle:1 laplace:2 suppose:1 exact:1 agreement:4 associate:3 overlapped:1 harvard:1 recognition:2 genuinely:1 mammalian:1 observed:4 module:3 initializing:1 capture:3 pmm:1 connected:1 removed:1 complexity:1 stereograms:1 jittering:1 trained:4 yuille:3 upon:2 triangle:3 translated:1 joint:11 geoff:1 train:3 separated:2 detected:1 whose:1 valued:1 solve:1 triangular:1 statistic:1 final:1 sequence:2 frozen:1 net:2 took:1 remainder:1 organizing:2 achieve:2 getting:1 optimum:3 produce:1 perfect:3 object:30 derive:3 ij:4 school:1 advocated:1 sa:7 predicted:1 involves:1 ning:1 closely:1 iio:1 correct:2 centered:2 translating:1 yai:2 preliminary:1 hold:1 exp:1 cb:3 lawrence:1 predict:1 major:1 loyer:2 applicable:2 currently:1 weighted:1 sensor:1 always:1 rather:1 varying:3 kim:1 detect:2 helpful:1 dayan:2 lj:2 hidden:5 i1:1 pixel:3 classification:14 flexible:1 ill:1 orientation:1 softmax:1 fairly:1 hoff:1 mutual:2 special:1 field:1 extraction:2 having:3 capitalize:2 unsupervised:3 future:1 others:1 connectionist:1 report:1 employ:1 retina:2 randomly:3 oriented:2 recognize:1 divergence:1 consisting:1 kll:1 organization:2 multiply:1 mixture:1 yielding:1 sens:1 accurate:2 edge:2 rotating:1 desired:2 circle:3 instance:3 soft:1 logp:1 cost:11 subset:1 entry:1 predictor:1 uniform:1 delay:1 successful:1 erlbaum:1 too:2 considerably:2 density:5 receiving:1 together:1 central:1 containing:2 watching:1 corner:1 mcmaster:1 derivative:4 strive:1 de:9 blurred:1 depends:1 view:4 observer:1 try:1 helped:1 observing:1 analyze:1 sort:1 relied:1 moivre:2 minimize:2 il:1 square:3 descriptor:2 kaufmann:2 maximized:1 ensemble:4 ybi:1 raw:1 bayesian:1 produced:3 trajectory:2 liebler:1 converged:1 explain:1 detector:1 tended:2 jpmax:5 frequency:1 ubiquitous:1 back:2 higher:3 supervised:1 response:5 yb:1 formulation:1 just:2 stage:1 correlation:1 horizontal:1 propagation:2 continuity:1 indicated:1 spatially:1 laboratory:1 white:2 ll:1 self:5 ambiguous:2 anything:1 plate:1 fj:1 image:12 discovers:2 common:1 rotation:1 multinomial:1 ji:1 refer:1 ai:1 grid:1 centre:1 gratefully:1 had:2 dot:1 moving:5 similarity:1 surface:1 pu:1 dominant:1 foldiak:1 disappointing:1 occasionally:2 success:3 yi:1 morgan:2 converge:1 maximize:2 signal:2 ii:9 multiple:6 desirable:1 hebbian:2 technical:1 equally:3 privileged:1 vision:1 physically:2 iteration:3 represent:2 sometimes:2 robotics:1 achieved:2 receive:3 source:1 modality:5 comment:1 subject:1 intermediate:1 iii:2 fit:1 psychology:1 architecture:1 competing:1 idea:2 six:3 ultimate:1 becker:17 peter:2 adequate:1 ten:1 trapped:1 per:3 discrete:1 group:21 acknowledged:1 rectangle:1 enforced:1 run:5 powerful:3 respond:1 reasonable:1 patch:3 coherence:9 prefer:1 scaling:1 layer:30 summer:1 constraint:5 scene:1 encodes:1 stelios:1 nearby:4 extremely:1 department:1 o_:1 tmining:1 combination:1 across:8 smaller:1 kmm:1 invariant:3 equation:2 agree:2 previously:1 turn:1 count:1 wrt:2 observe:2 appropriate:1 disagreement:2 alternative:2 altogether:1 top:2 tony:1 hinge:1 unsuitable:1 k1:1 added:2 quantity:1 receptive:1 usual:1 diagonal:7 gradient:2 link:1 lateral:2 evenly:1 chris:1 minimizing:1 susceptible:2 negative:1 perform:1 disagree:1 vertical:1 observation:1 descent:1 displayed:1 hinton:6 extended:1 frame:5 y1:1 rn:1 varied:3 arbitrary:1 pair:1 required:1 connection:1 coherent:3 learned:8 able:4 suggested:1 usually:2 pattern:11 challenge:1 max:1 overlap:1 natural:1 force:1 representing:1 scheme:1 temporally:3 extract:1 fjk:1 prior:15 geometric:3 acknowledgement:1 blockdiagonal:1 expect:1 permutation:1 mixed:1 limitation:1 versus:1 pij:10 consistent:1 principle:4 viewpoint:3 classifying:1 translation:1 row:1 diagonally:1 last:1 keeping:1 allow:2 taking:1 distributed:1 overcome:1 seemed:1 sensory:6 made:1 preprocessing:1 far:1 biilthoff:1 ignore:1 kullback:1 dealing:1 llij:1 active:1 a_:1 knew:1 xi:1 continuous:1 promising:2 learn:5 nature:1 robust:1 ca:3 investigated:1 utput:1 did:1 ybj:1 whole:1 xu:1 position:3 momentum:1 crude:1 perceptual:1 third:1 learns:1 specific:2 k21:1 trai:1 l8s:1 hoped:1 simply:2 visual:2 partially:1 goal:2 lilli:1 change:2 typical:1 specifically:1 except:1 discriminate:1 invariance:2 meant:1 categorize:1
7,129
895
Factorial Learning and the EM Algorithm Zoubin Ghahramani zoubin@psyche.mit.edu Department of Brain & Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 Abstract Many real world learning problems are best characterized by an interaction of multiple independent causes or factors. Discovering such causal structure from the data is the focus of this paper. Based on Zemel and Hinton's cooperative vector quantizer (CVQ) architecture, an unsupervised learning algorithm is derived from the Expectation-Maximization (EM) framework. Due to the combinatorial nature of the data generation process, the exact E-step is computationally intractable. Two alternative methods for computing the E-step are proposed: Gibbs sampling and mean-field approximation, and some promising empirical results are presented. 1 Introduction Many unsupervised learning problems fall under the rubric of factorial learning-that is, the goal of the learning algorithm is to discover multiple independent causes, or factors, that can well characterize the observed data (Barlow, 1989; Redlich, 1993; Hinton and Zemel, 1994; Saund, 1995). Such learning problems often arise naturally in response to the actual process by which the data have been generated. For instance, images may be generated by combining multiple objects , or varying colors, locations, and poses, with different light sources. Similarly, speech signals may result from an interaction of factors such as the tongue position, lip aperture, glottal state, communication line, and background noises. The goal of factorial learning is to invert this data generation process, discovering a representation that will both parsimoniously describe the data and reflect its underlying causes. A recent approach to factorial learning uses the Minimum Description Length (MDL) principle (Rissanen, 1989) to extract a compact representation of the input (Zemel, 1993; Hinton and Zemel, 1994). This has resulted in a learning architecture 618 Zoubin Ghahramani called Cooperative Vector Quantization (CVQ), in which a set of vector quantizers cooperates to reproduce the input . Within each vector quantizer a competitive learning mechanism operates to select an appropriate vector code to describe the input. The CVQ is related to algorithms based on mixture models, such as soft competitive clustering, mixtures of experts (Jordan and Jacobs, 1994), and hidden Markov models (Baum et al., 1970), in that each vector quantizer in the CVQ is itself a mixture model. However, it generalizes this notion by allowing the mixture models to cooperate in describing features in the data set, thereby creating a distributed representations of the mixture components. The learning algorithm for the CVQ uses MDL to derive a cost function composed of a reconstruction cost (e.g. sum squared error), representation cost (negative entropy of the vector code), and model complexity (description length of the network weights), which is minimized by gradient descent. In this paper we first formulate the factorial learning problem in the framework of statistical physics (section 2). Through this formalism, we derive a novel learning algorithm for the CVQ based on the Expectation-Maximization (EM) algorithm (Dempster et al., 1977) (section 3). The exact EM algorithm is intractable for this and related factorial learning problems-however, a tractable mean-field approximation can be derived. Empirical results on Gibbs sampling and the mean-field approximation are presented in section 4. 2 Statistical Physics Formulation The CVQ architecture, shown in Figure 1, is composed of hidden and observable units, where the observable units, y, are real-valued, and the hidden units are discrete and organized into vectors Si, i = 1, ... , d. The network models a data generation process which is assumed to proceed in two stages. First, a factor is independently sampled from each hidden unit vector, Sj, according to its prior probability density, ?ri. Within each vector the factors are mutually exclusive, i.e. if Sij = 1 for some j, then Sik = 0, Vk -# j. The observable is then generated from a Gaussian distribution with mean 2:1=1 WiSi. Notation: 0 0 0 0 0 VOl 51 0 0 0 0 0 52 ??? V0 2 0 0 0 0 0 d k p N Sd Sij Si Wi VOd Y number of vectors number of hidden units per vector number of outputs number of patterns hidden unit j in vector i vector i of units (Si = [Si1, ... , Sik]) weight matrix from Si to output network output (observable) Figure 1. The factorial learning architecture. Defining the energy of a particular configuration of hidden states and outputs as 1 H(s, y) = "2 lly - d L i=l d W i s ill 2 - k LL i=l j=l Sij log 7rij, (1) Factorial Learning and the EM Algorithm 619 the Boltzmann distribution p(s, y) 1 = -free Z exp{-11.(s,y)}, (2) exactly recovers the probability model for the CVQ. The causes or factors are represented in the multinomial variables Si and the observable in the multivariate Gaussian y. The undamped partition function, Zjree, can be evaluated by summing and integrating over all the possible configurations of the system to obtain Zjree =~ $ 1 exp{ -11.(s, y)}dy = (21l")P/2, (3) Y which is constant, independent of the weights. This constant partition function results in desirable properties, such as the lack of a Boltzmann machine-like sleep phase (Neal, 1992), which we will exploit in the learning algorithm. The system described by equation (1)1 can be thought of as a special form of the Boltzmann machine (Ackley et al., 1985). Expanding out the quadratic term we see that there are pairwise interaction terms between every unit. The evaluation of the partition function (3) tells us that when y is unclamped the quadratic term can be integrated out and therefore all Si are independent. However, when y is clamped all the Si become dependent. 3 The EM Algorithm Given a set of observable vectors, the goal of the unsupervised learning algorithm is to find weight matrices such that the network is most likely to have generated the data. If the hidden causes for each observable where known, then the weight matrices could be easily estimated . However, the hidden causes cannot be inferred unless these weight matrices are known. This chicken-and-egg problem can be solved by iterating between computing the expectation of the hidden causes given the current weights and maximizing the likelihood of the weights given these expected causes-the two steps forming the basis of the Expectation-Maximization (EM) algorithm (Dempster et al., 1977). Formally, from (2) we obtain the expected log likelihood of the parameters ?/: = (-11.(s,y) -logZjree)c,q, where ? denotes the current parameters, ? = {Wi}?=1, and (-)c,q, Q(?,?/) (4) denotes expectation given ? and the damped observables. The E-step of EM consists of computing this expected log likelihood. As the only random variables are the hidden causes, this simplifies to computing the (Si)c and (SiS])c terms appearing in the quadratic expansion of 11.. Once these terms have been computed, the M-step consists of maximizing Q with respect to the parameters. Setting the derivatives to zero we obtain a linear system, 1 For the remainder of the paper we will ignore the second term in (1), thereby assuming equal priors on the hidden states. Relaxing this assumption and estimating priors from the data is straightforward. 620 Zoubin Ghahramani which can be solved via the normal equations, where s is the vector of concatenated Si and the subscripts denote matrix size. For models in which the observable is a monotonic differentiable function ofLi WiSi, i.e. generalized linear models, least squares estimates of the weights for the M-step can be obtained iteratively by the method of scoring (McCullagh and NeIder, 1989). 3.1 E-step: Exact The difficulty arises in the E-step of the algorithm. The expectation of hidden unit j in vector i given pattern y is: P(Sij = 11Y; W) Ii; <X Ii; <x P(ylSij = 1;W)1l"ij Ii; L?? L .. LP(ylsij = 1, slh = 1, .. . ,Sdjd = 1; W)1l"ij jl=l ih,ti=l jd=l To compute this expectation it is necessary to sum over all possible configurations of the other hidden units. If each vector quantizer has k hidden units, each expectation has time complexity of O( k d - l ), i.e. O( N k d ) for a full E-step. The exponential time is due inherently to the cooperative nature of the model-the setting of one vector only determines the observable if all the other vectors are fixed. 3.2 E-step: Gibbs sampling Rather than summing over all possible hidden unit patterns to compute the exact expectations, a natural approach is to approximate them through a Monte Carlo method. As with Boltzmann machines, the CVQ architecture lends itself well to Gibbs sampling (Geman and Geman, 1984). Starting from a clamped observable y and a random setting of the hidden units {Sj}, the setting of each vector is updated in turn stochastically according to its conditional distribution Si '" p( sdY, {Sj h;ti; W). Each conditional distribution calculation requires k forward passes through the network, one for each possible state of the vector being updated, and k Gaussian distance calculations between the resulting predicted and clamped observables. If all the probabilities are bounded away from zero this process is guaranteed to converge to the equilibrium distribution of the hidden units given the observable. The first and second-order statistics, for (Si)c and (SiS])c respectively, can be collected using the Sij'S visited and p( Si Iy, {Sj h;ti; W) calculated during this sampling process. These estimated expectations are then used in the E-step. 3.3 E-step: Mean-field approximation Although Gibbs sampling is generally much more efficient than exact calculations, it too can be computationally demanding . A more promising approach is to approximate the intractable system with a tractable mean-field approximation (Parisi , 1988), and perform the E-step calculation on this approximation . We can write the Factorial Learning and the EM Algorithm 621 negative log likelihood minimized by the original system as a difference between the damped and undamped free energies: Cost -logp(y/W) = -log ~p(y, s/W) s -log~exp{-'J-l(y,s)} + log~ s Fcl - FJree s [ exp{-'J-l(y,s)}dy Jy The mean-field approximation allows us to replace each free energy in this cost with an upper bound approximation Cost M F F:t F - Ffrt'e. Unfortunately, a difference of two upper bounds is not generally an upper bound, and therefore minimizing Cost M F in, for example, mean-field Boltzmann machines does not guarantee that we are minimizing an upper bound on Cost. However, for the factorial learning architectures described in this paper we have the property that FJree is constant, and therefore the mean-field approximation of the cost is an upper bound on the exact cost. = The mean-field approximation can be obtained by approximating the probability density given by (1) and (2) by a completely factorized probability density: p(s, y) = (21r~P/2 exp{ -~/IY -1L/12} nm:y ',J In this approximation all units are independent: the observables are Gaussian distributed with mean IL and each hidden unit is binomially distributed with mean mij. To obtain the mean-field approximation we solve for the mean values that minimize the Kullback-Leibler divergence KC(p,p) == Ep[logp] - Ep[logp]. Noting that: Ep[Sij] = mij, Ep[S[j] = mij, Ep[SijSkd = mijmkl, and Ep[SijSik] = 0, we obtain the mean-field fixed point equations (5) where y == l:i WiIDi. The softmax function is the exponential normalized over the k hidden units in each IDi vector. The first term inside the softmax has an intuitive interpretation as the projection of the error in the observable onto the weights of the hidden unit vector i. The more a hidden unit can reduce this error, the higher its mean. The second term arises from the fact that Ep[s[j] = mij and not Ep[s[j] m[j. The means obtained by iterating equation (5) are used in the E-step by substituting mi for (Si)c and IDiID] for (sisJk = 4 Empirical Results Two methods, Gibbs sampling and mean-field, have been provided for computing the E-step of the factorial learning algorithm. There is a key empirical question that needs to be answered to determine the efficiency and accuracy of each method. For Gibbs sampling it is important to know how many samples will provide robust estimates of the expectations required for the E-step. It is well known that for stochastic Boltzmann machines the number of samples needed to obtain good 622 Zoubin Ghahramani estimates of the gradients is generally large and renders the learning algorithm prohibitively slow. Will this architecture suffer from the same problem? For mean-field it is important to know the loss incurred by approximating the true likelihood. We explore these questions by presenting empirical results on two small unsupervised learning problems. The first benchmark problem consists of a data set of 4 x 4 greyscale images generated by a combination of two factors: one producing a single horizontal line and the other, a vertical line (Figure 2a; cf. Zemel, 1993). Using a network with 2 vectors of 4 hidden units each, both the Gibbs sampling and mean-field EM algorithms converge on a solution after about a dozen steps (Figure 2b). The solutions found resemble the generative model of the data (Figure 2c & d). b) a) 70 .0 5o :'i 0 ." ~ 40 ~ l O ; i 2 0 _= s ~.r-r--- _--- ~-="-~- 10 oo-~_~_~_~_~ 10 J'i 20 Iteration c) d) Figure 2. Lines Problem. a) Complete data set of 160 patterns. b) Learning curves for Gibbs (solid) and mean-field (dashed) forms of the algorithm. c) A sample output weight matrix after learning (MSE=1.20). The top vector of hidden units has come to represent horizontal lines, and the bottom, vertical lines. d) Another typical output weight matrix (MSE=1.78). The second problem consists of a data set of 6 x 6 images generated by a combination of three shapes-a cross, a diagonal line, and an empty square-each of which can appear in one of 16 locations (Figure 3a). The data set of 300 out of 4096 possible images was presented to a network with the architecture shown in Figure 3b. After 30 steps of EM, each consisting of 5 Gibbs samples of each hidden unit, the network reconstructed a representation that approximated the three underlying causes of the data-dedicating one vector mostly to diagonal lines, one to hollow squares, and one to crosses (Figure 3c). To assess how many Gibbs samples are required to obtain accurate estimates of the expectations for the E-step we repeated the lines problem varying the number of samples. Clearly, as the number of samples becomes large the Gibbs E-step becomes exact. Therefore we expect performance to asymptote at the performance of the exact E-step. The results indicate that, for this problem, 3 samples are sufficient to achieve ceiling performance (Figure 4). Surprisingly, a single iteration of the Factorial Learning and the EM Algorithm 623 a) b) ~_- lSi- -~ , -:JI t rIl 1f --.~ ? .p- rj- ~_- ~~ 36 ;. ~+ .,. a:,. :'I,," D.... If :m' ~1r- ~.t ti .q. at- ..:.p If !ill1I- . ........ :11:1 ti+ "t:I:I .ttt~ ... 16 16 oo 16 0 0 c) Figure 3. Shapes Problem. a) Sample images from the data set. b) Learning architecture used. c) Output weight matrix after learning. mean-field equations also performs quite well. 5 Discussion The factorial learning problem for cooperative vector quantizers has been formulated in the EM framework, and two learning algorithms, based on Gibbs sampling and mean-field approximation, have been derived. Unlike the Boltzmann machine, Gibbs sampling for this architecture seems to require very few samples for adequate performance. This may be due to the fact that, whereas the Boltzmann machine relies on differences of noisy estimates for its weight changes, due to the constant partition function the factorial learning algorithm does not. The mean-field approximation also seems to perform quite well on all problems tested to date . This may also be a consequence of the constant partition function which guarantees that the mean-field cost is an upper bound on the exact cost. The framework can be extended to hidden Markov models (HMMs), showing that simple HMMs are a special case of dynamical CVQs, with the general case corresponding to parallel, factorial HMMs. The two principal advantages of such architectures are (1) unlike the traditional HMM, the state space can be represented as a combination of features, and (2) time series generated by multiple sources can be modeled. Simulation results on the Gibbs and mean-field EM algorithms for factorial HMMs are also promising (Ghahramani, 1995). 624 Zoubin Ghahramani 2.8 1 2 3 4 Gibbs samples or mean-field lIerarions Figure 4. Comparison of the Gibbs and mean-field EM algorithms for the lines data. Each data point shows the mean squared training error averaged over 10 runs of 20 EM steps, with standard error bars. For the Gibbs curve the abscissa is the number of samples per vector of hidden units; for the mean-field curve it is the number of iterations of equation (5). Acknowledgements The author wishes to thank Lawrence Saul and Michael Jordan for invaluable discussions. This project was supported in part by a grant from the McDonnell-Pew Foundation, by a grant from ATR Human Information Processing Research Laboratories, by a grant from Siemens Corporation, and by grant N00014-94-1-0777 from the Office of Naval Research. References Ackley, D., Hinton, G., and Sejnowski, T. (1985). A learning algorithm for Boltzmann machines. Cognitive Science, 9:147-169. Barlow, H. (1989). Unsupervised learning. Neural Computation, 1:295-31l. Baum, L., Petrie, T., Soules, G., and Weiss, N. (1970). A maximization technique occurring in the statistical analysis of probabilistic functions of Markov chains. The Annals of Mathematical Statistics, 41:164-17l. Dempster, A., Laird, N., and Rubin, D. (1977). Maximum likelihood from incomplete data via the EM algorithm. J. Royal Statistical Society Series B, 39:1-38. Geman, S. and Geman, D. (1984). Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Transactions on Pattern Analysis and Machine Intelligence,6:721-74l. Ghahramani, Z. (1995). Factorial learning and the EM algorithm. MIT Computational Cognitive Science TR 9501. Hinton, G. and Zemel, R. (1994) . Autoencoders, minimum description length, and Helmholtz free energy. In Cowan, J., Tesauro, G., and Alspector, J., editors, Advances in Neural Information Processing Systems 6. Morgan Kaufmanm Publishers, San Francisco, CA. Jordan, M. and Jacobs, R. (1994). Hierarchical mixtures of experts and the EM algorithm. Neural Computation, 6:181-214. McCullagh, P ..and NeIder, J. (1989). Generalized Linear Models. Chapman & Hall, London. Neal, R. (1992). Connectionist learning of belief networks. Artificial Intelligence, 56:71113. Parisi, G. (1988). Statistical Field Theory. Addison-Wesley, Redwood City, CA. Redlich, A. (1993). Supervised factorial learning. Neural Computation, 5:750-766. Rissanen, J. (1989). Stochastic Complexity in Statistical Inquiry. World Scientific, Singapore. Saund, E. (1995). A multiple cause mixture model for unsupervised learning. Neural Computation,7(1):51-7l. Zemel, R. (1993). A minimum description length framework for unsupervised learning. Ph.D. Thesis, Dept. of Computer Science, University of Toronto, Toronto, Canada.
895 |@word seems:2 simulation:1 jacob:2 thereby:2 tr:1 solid:1 configuration:3 series:2 current:2 soules:1 si:15 partition:5 shape:2 asymptote:1 generative:1 discovering:2 intelligence:2 slh:1 quantizer:4 idi:1 location:2 toronto:2 si1:1 mathematical:1 become:1 consists:4 inside:1 pairwise:1 expected:3 alspector:1 abscissa:1 brain:1 actual:1 becomes:2 provided:1 discover:1 underlying:2 notation:1 estimating:1 bounded:1 factorized:1 project:1 corporation:1 guarantee:2 every:1 ti:5 exactly:1 prohibitively:1 unit:23 grant:4 appear:1 producing:1 sd:1 consequence:1 subscript:1 cooperates:1 relaxing:1 hmms:4 averaged:1 empirical:5 thought:1 projection:1 integrating:1 zoubin:6 cannot:1 onto:1 baum:2 maximizing:2 straightforward:1 starting:1 independently:1 formulate:1 notion:1 updated:2 annals:1 exact:9 us:2 helmholtz:1 approximated:1 geman:4 cooperative:4 observed:1 ackley:2 ep:8 bottom:1 rij:1 solved:2 dempster:3 complexity:3 efficiency:1 basis:1 observables:3 completely:1 easily:1 represented:2 describe:2 london:1 monte:1 sejnowski:1 artificial:1 zemel:7 tell:1 quite:2 valued:1 solve:1 statistic:2 itself:2 noisy:1 laird:1 advantage:1 differentiable:1 parisi:2 reconstruction:1 interaction:3 remainder:1 combining:1 date:1 achieve:1 description:4 intuitive:1 empty:1 object:1 derive:2 oo:2 pose:1 ij:2 predicted:1 resemble:1 come:1 indicate:1 stochastic:3 human:1 require:1 hall:1 normal:1 exp:5 equilibrium:1 lawrence:1 substituting:1 combinatorial:1 visited:1 city:1 mit:2 clearly:1 gaussian:4 rather:1 varying:2 office:1 derived:3 focus:1 unclamped:1 naval:1 vk:1 likelihood:6 dependent:1 integrated:1 glottal:1 hidden:27 kc:1 reproduce:1 ill:1 special:2 softmax:2 field:24 once:1 equal:1 sampling:11 chapman:1 wisi:2 unsupervised:7 minimized:2 connectionist:1 few:1 composed:2 resulted:1 divergence:1 parsimoniously:1 phase:1 consisting:1 evaluation:1 mdl:2 mixture:7 light:1 damped:2 chain:1 accurate:1 necessary:1 unless:1 incomplete:1 causal:1 tongue:1 instance:1 formalism:1 soft:1 logp:3 restoration:1 maximization:4 cost:12 too:1 characterize:1 density:3 probabilistic:1 physic:2 michael:1 iy:2 squared:2 reflect:1 nm:1 thesis:1 cognitive:3 creating:1 expert:2 derivative:1 stochastically:1 saund:2 neider:2 competitive:2 parallel:1 ttt:1 minimize:1 square:3 il:1 accuracy:1 ass:1 bayesian:1 carlo:1 inquiry:1 energy:4 naturally:1 mi:1 recovers:1 sampled:1 massachusetts:1 color:1 organized:1 wesley:1 higher:1 supervised:1 response:1 wei:1 formulation:1 evaluated:1 stage:1 autoencoders:1 horizontal:2 lack:1 scientific:1 normalized:1 true:1 barlow:2 ril:1 iteratively:1 leibler:1 laboratory:1 neal:2 ll:1 during:1 generalized:2 presenting:1 complete:1 performs:1 invaluable:1 cooperate:1 image:6 novel:1 petrie:1 multinomial:1 ji:1 jl:1 interpretation:1 cambridge:1 gibbs:19 pew:1 similarly:1 v0:1 multivariate:1 recent:1 tesauro:1 n00014:1 scoring:1 morgan:1 minimum:3 converge:2 determine:1 dashed:1 signal:1 ii:3 multiple:5 desirable:1 full:1 rj:1 characterized:1 calculation:4 cross:2 jy:1 expectation:12 lly:1 iteration:3 represent:1 invert:1 chicken:1 dedicating:1 background:1 whereas:1 source:2 publisher:1 unlike:2 pass:1 cowan:1 jordan:3 noting:1 architecture:11 reduce:1 simplifies:1 render:1 suffer:1 speech:1 proceed:1 cause:11 adequate:1 generally:3 iterating:2 factorial:18 ph:1 lsi:1 singapore:1 estimated:2 per:2 discrete:1 write:1 vol:1 key:1 rissanen:2 relaxation:1 sum:2 run:1 dy:2 bound:6 guaranteed:1 sleep:1 quadratic:3 ri:1 answered:1 department:1 according:2 combination:3 mcdonnell:1 em:19 psyche:1 wi:2 lp:1 sij:6 ceiling:1 computationally:2 equation:6 mutually:1 describing:1 turn:1 mechanism:1 needed:1 know:2 addison:1 tractable:2 rubric:1 generalizes:1 hierarchical:1 away:1 appropriate:1 appearing:1 alternative:1 jd:1 original:1 denotes:2 clustering:1 cf:1 top:1 exploit:1 concatenated:1 ghahramani:7 approximating:2 society:1 question:2 exclusive:1 diagonal:2 traditional:1 gradient:2 lends:1 distance:1 thank:1 atr:1 hmm:1 collected:1 assuming:1 length:4 code:2 modeled:1 minimizing:2 unfortunately:1 mostly:1 quantizers:2 greyscale:1 negative:2 binomially:1 boltzmann:9 perform:2 allowing:1 upper:6 vertical:2 markov:3 benchmark:1 descent:1 defining:1 hinton:5 communication:1 extended:1 redwood:1 canada:1 inferred:1 required:2 bar:1 dynamical:1 pattern:5 royal:1 belief:1 demanding:1 difficulty:1 natural:1 technology:1 cvqs:1 extract:1 prior:3 acknowledgement:1 loss:1 expect:1 generation:3 undamped:2 foundation:1 incurred:1 sufficient:1 rubin:1 principle:1 sik:2 editor:1 surprisingly:1 supported:1 free:4 institute:1 fall:1 saul:1 distributed:3 curve:3 calculated:1 world:2 forward:1 author:1 san:1 transaction:1 sj:4 approximate:2 compact:1 observable:12 ignore:1 aperture:1 kullback:1 reconstructed:1 summing:2 assumed:1 francisco:1 lip:1 promising:3 nature:2 robust:1 expanding:1 inherently:1 ca:2 expansion:1 mse:2 noise:1 arise:1 repeated:1 redlich:2 egg:1 slow:1 position:1 wish:1 exponential:2 clamped:3 dozen:1 showing:1 intractable:3 quantization:1 ih:1 occurring:1 entropy:1 likely:1 explore:1 forming:1 monotonic:1 mij:4 determines:1 relies:1 ma:1 conditional:2 goal:3 formulated:1 replace:1 change:1 mccullagh:2 typical:1 operates:1 cvq:9 principal:1 called:1 siemens:1 select:1 formally:1 arises:2 hollow:1 dept:1 tested:1
7,130
896
Learning Local Error Bars for Nonlinear Regression David A.Nix Department of Computer Science and Institute of Cognitive Science University of Colorado Boulder, CO 80309-0430 dnix@cs.colorado.edu Andreas S. Weigend Department of Computer Science and Institute of Cognitive Science University of Colorado Boulder, CO 80309-0430 andreas@cs.colorado.edu? Abstract We present a new method for obtaining local error bars for nonlinear regression, i.e., estimates of the confidence in predicted values that depend on the input. We approach this problem by applying a maximumlikelihood framework to an assumed distribution of errors. We demonstrate our method first on computer-generated data with locally varying, normally distributed target noise. We then apply it to laser data from the Santa Fe Time Series Competition where the underlying system noise is known quantization error and the error bars give local estimates of model misspecification. In both cases, the method also provides a weightedregression effect that improves generalization performance. 1 Learning Local Error Bars Using a Maximum Likelihood Framework: Motivation, Concept, and Mechanics Feed-forward artificial neural networks used for nonlinear regression can be interpreted as predicting the mean of the target distribution as a function of (conditioned on) the input pattern (e.g., Buntine & Weigend, 1991; Bishop, 1994), typically using one linear output unit per output variable. If parameterized, this conditional target distribution (CID) may also be ?http://www.cs.colorado.edu/~andreas/Home.html. This paper is available with figures in colors as ftp://ftp.cs.colorado.edu/pub/ Time-Series/MyPapers/nix.weigenCLnips7.ps.Z . 490 David A. Nix, Andreas S. Weigend viewed as an error model (Rumelhart et al., 1995). Here, we present a simple method that provides higher-order information about the cm than simply the mean. Such additional information could come from attempting to estimate the entire cm with connectionist methods (e.g., "Mixture Density Networks," Bishop, 1994; "fractional binning, "Srivastava & Weigend, 1994) or with non-connectionist methods such as a Monte Carlo on a hidden Markov model (Fraser & Dimitriadis, 1994). While non-parametric estimates of the shape of a C1D require large quantities of data, our less data-hungry method (Weigend & Nix, 1994) assumes a specific parameterized form of the C1D (e.g., Gaussian) and gives us the value of the error bar (e.g., the width of the Gaussian) by finding those parameters which maximize the likelihood that the target data was generated by a particular network model. In this paper we derive the specific update rules for the Gaussian case. We would like to emphasize, however, that any parameterized unimodal distribution can be used for the em in the method presented here. j------------, I------T-------, I I A o y(x) I A2,. ) 0 cr IX I I /\ i O'OOh k : , '. \ I I I I I I I : ,-----------_.l Figure 1: Architecture of the network for estimating error bars using an auxiliary output unit. All weight layers have full connectivity. This architecture allows the conditional variance ~2 -unit access to both information in the input pattern itself and in the hidden unit representation formed while learning the conditional mean, y(x). We model the desired observed target value d as d(x) = y(x) + n(x), where y(x) is the underlying function we wish to approximate and n(x) is noise drawn from the assumed cm. Just as the conditional mean of this cm, y(x), is a function of the input, the variance (j2 of the em, the noise level, may also vary as a function of the input x (noise heterogeneity). Therefore, not only do we want the network to learn a function y(x) that estimates the conditional mean y(x) of the cm, but we also want it to learn a function a- 2 (x) that estimates the conditional variance (j2(x). We simply add an auxiliary output unit, the a- 2-unit, to compute our estimate of (j2(x). Since (j2(x) must be positive, we choose an exponential activation function to naturally impose this bound: a-2 (x) = exp [Lk Wq2khk (x) + ,8], where,8 is the offset (or "bias"), and Wq2k is the weight between hidden unit k and the a- 2 -unit. The particular connectivity of our architecture (Figure 1), in which the a- 2-unit has a hidden layer of its own that receives connections from both the y-unit's hidden layer and the input pattern itself, allows great flexibility in learning a- 2 (x). In contrast, if the a- 2 -unit has no hidden layer of its own, the a- 2 -unit is constrained to approximate (j2 (x) using only the exponential of a linear combination of basis functions (hidden units) already tailored to represent y(x) (since learning the conditional variance a- 2 (x) before learning the conditional mean y(x) is troublesome at best). Such limited connectivity can be too constraining on the functional forms for a- 2 ( x) and, in our experience, I The case of a single Gaussian to represent a unimodal distribution can also been generalized to a mixture of several Gaussians that allows the modeling of multimodal distributions (Bishop, 1994). Learning Local Error Bars for Nonlinear Regression 491 produce inferior results. This is a significant difference compared to Bishop's (1994) Gaussian mixture approach in which all output units are directly connected to one set of hidden units. The other extreme would be not to share any hidden units at all, i.e., to employ two completely separate sets of hidden units, one to the y(x)-unit, the other one to the a- 2(x)-unit. This is the right thing to do if there is indeed no overlap in the mapping from the inputs to y and from the inputs to cr2 ? The two examples discussed in this paper are between these two extremes; this justifies the mixed architecture we use. Further discussion on shared vs. separate hidden units for the second example of the laser data is given by Kazlas & Weigend (1995, this volume). For one of our network outputs, the y-unit, the target is easily available-it is simply given by d. But what is the target for the a- 2-unit? By maximizing the likelihood of our network ' model N given the data, P(Nlx, d), a target is "invented" as follows. Applying Bayes' rule and assuming statistical independence of the errors, we equivalently do gradient descent in the negative log likelihood of the targets d given the inputs and the network model, summed over all patterns i (see Rumelhart et at., 1995): C = - Li In P(dilxi, N). Traditionally, the resulting form of this cost function involves only the estimate Y(Xi) of the conditional mean; the variance of the CID is assumed to be constant for all Xi, and the constant terms drop out after differentiation. In contrast, we allow the conditional variance to depend on x and explicitly keep these terms in C, approximating the conditional variance for Xi by a- 2(Xi). Given any network architecture and any parametric form for the ern (Le., any error model), the appropriate weight-update equations for gradient decent learning can be straightforwardly derived. Assuming normally distributed errors around y(x) corresponds to a em density function of P(dilxj) = [27rcr2(Xi)t 1/ 2 exp {- d~:Y~.) Using the network output Y(Xi) ~ y(Xi) to estimate the conditional mean and using the auxiliary output a- 2(Xi) ~ cr2(xd to estimate the conditional variance, we obtain the monotonically related negative log . over all patterns likelib00, d - I n P(di IXi, nAf\J -- 2"IIn 2 7rcrA2() Xi + [di-y(Xi)]2 2".2(X.) . Summatlon 2}. gives the total cost: C = ! ,,{ [di =-2 y(xd] 2 + Ina-2(Xi) + In27r} 2~ , cr (Xi) (1) To write explicit weight-update equations, we must specify the network unit transfer functions. Here we choose a linear activation function for the y-unit, tanh functions for the hidden units, and an exponential function for the a- 2 -unit. We can then take derivatives of the cost C with respect to the network weights. To update weights connected to the Yand a- 2-units we have: 11 a-2~i) [di - 11 2a-2~Xi) {[di - (2) Y(Xi)] hj(Xi) y(Xi)f - a-2 (Xi) } hk (Xi) (3) where 11 is the learning rate. For weights not connected to the output, the weight-update equations are derived using the chain rule in the same way as in standard backpropagation. Note that Eq. (3) is equivalent to training a separate function-approximation network for a- 2(x) where the targets are the squared errors [d i - y(Xi)]2]. Note also that if a- 2(Xj) is 492 David A. Nix, Andreas S. Weigend constant, Eqs. (1)-(2) reduce to their familiar forms for standard backpropagation with a sum-squared error cost function. The 1/&2(X) term in Eqs. (2)-(3) can be interpreted as a form of "weighted regression," increasing the effective learning rate in low-noise regions and reducing it in high-noise regions. As a result, the network emphasizes obtaining small errors on those patterns where it can (low &2); it discounts learning patterns for which the expected error is going to be large anyway (large &2). This weighted-regression term can itself be highly beneficial where outliers (i.e., samples from high-noise regions) would ordinarily pull network resources away from fitting low-noise regions which would otherwise be well approximated. For simplicity, we use simple gradient descent learning for training. Other nonlinear minimization techniques could be applied, however, but only if the following problem is avoided. If the weighted-regression term described above is allowed a significant influence early in learning, local minima frequently result. This is because input patterns for which low errors are initially obtained are interpreted as "low noise" in Eqs. (2)-(3) and overemphasized in learning. Conversely, patterns for which large errors are initially obtained (because significant learning of y has not yet taken place) are erroneously discounted as being in "high-noise" regions and little subsequent learning takes place for these patterns, leading to highly-suboptimal solutions. This problem can be avoided if we separate training into the following three phases: Phase I (Initial estimate of the conditional mean): Randomly split the available data into equal halves, sets A and 8. Assuming u 2 (x) is constant, learn the estimate of the conditional mean y(x) using set A as the training set. This corresponds to "traditional" training using gradient descent on a simple squared-error cost function, i.e., Eqs. (1)-(2) without the 1/&2(X) terms. To reduce overfitting, training is considered complete at the minimum of the squared error on the cross-validation set 8, monitored at the end of each complete pass through the training data. Phase II (Initial estimate of the conditional variance): Attach a layer of hidden units connected to both the inputs and the hidden units of the network from Phase I (see Figure 1). Freeze the weights trained in Phase I, and train the &2-unit to predict the squared errors (see Eq. (3?, again using simple gradient descent as in Phase I. The training set for this phase is set 8, with set A used for cross-validation. If set A were used as the training set in this phase as well, any overfitting in Phase I could result in seriously underestimating u 2 (x). To avoid this risk, we interchange the data sets. The initial value for the offset (3 of the &2-unit is the natural logarithm of the mean squared error (from Phase I) of set 8. Phase II stops when the squared error on set A levels off or starts to increase. Phase ill (Weighted regression): Re-split the available data into two new halves, A' and 8'. Unfreeze all weights and train all network parameters to minimize the full cost function C on set A'. Training is considered complete when C has reached its minimum on set 8'. 2 Examples Example #1: To demonstrate this method, we construct a one-dimensional example problem where y(x) and u 2 (x) are known. We take the equation y(x) = sin(wax) sin(w,Bx) withw a = 3 andw,B = 5. We then generate (x, d) pairs by picking x uniformly from the interval [0, 7r /2] and obtaining the corresponding target d by adding normally distributed noise n(x) = N[0,u 2 (x)] totheunderlyingy(x), whereu 2(x) = 0.02+0.25 x [1-sin(w,Bx)j2. 493 Learnillg Local Error Bars for Nonlinear Regression Table 1: Results for Example #1. ENMS denotes the mean squared error divided by the overall variance of the target; "Mean cost" represents the cost function (Eq. (1)) averaged over all patterns . Row 4 lists these values for the ideal model (true y(x) and a 2 (x)) given the data generated. Row 5 gives the correlation coefficient between the network's predictions for the standard error (i.e., the square root of the &2 -unit's activation) and the actually occurring L1 residual errors, Id(Xi) - y(x;) I. Row 6 gives the correlation between the true a(x) and these residual errors . Rows 7-9 give the percentage of residuals smaller than one and two standard deviations for the obtained and ideal models as well as for an exact Gaussian. 1 Training 1 Phase I 2 3 4 Phase " Phase III n( x ) (exact additive noise) 5 6 ~ ~~ ~ \: residual errors) 7 8 9 % of errors % of errors I a( x ,reridual errors) < .,.~ xl; 2"'~ xl < a-(x); 2a-(x) (aact Gaussian) ENM ."< 0.576 0.576 0.552 0.545 (N -- 103 ) Er.. IU .,,< 0.593 0.593 0.570 0.563 0.853 0.542 0.440 0.430 I P 0.564 0.602 I sId 64.8 66.6 68.3 1 Evaluation Mean cost 2 sId 95.4 96.0 95.4 (N -- 105 ) 1 Mean cost 0.882 0.566 0.462 0.441 J p 0.548 0.584 I sId 67.0 68.4 68.3 2 sId 94.6 95.4 95.4 We generate 1000 patterns for training and an additional 105 patterns for post-training evaluation. Training follows exactly the three phases described above with the following details: 2 Phase I uses a network with one hidden layer of 10 tanh units and TJ = 10- 2 ? For Phase II we add an auxiliary layer of 10 tanh hidden units connected to the a.2-unit (see Figure 1) and use the same TJ. Finally, in Phase III the composite network is trained with TJ 10- 4. = At the end of Phase I (Figure 2a), the only available estimate of (1'2 (x ) is the global root-mean-squared error on the available data, and the model misspecification is roughly uniform over x-a typical solution were we training with only the traditional squared-error cost function. The corresponding error measures are listed in Table 1. At the end of Phase II, however, we have obtained an initial estimate of (1'2 (x ) (since the weights to the i)-unit are frozen during this phase, no modification of i) is made). Finally, at the end of Phase III, we have better estimates of both y{ x) and (1'2 (x). First we note that the correlations between the predicted errors and actual errors listed in Table 1 underscore the near-optimal prediction of local errors. We also see that these errors correspond, as expected, to the assumed Gaussian error model. Second, we note that not only has the value of the cost function dropped from Phase II to Phase III, but the generalization error has also dropped, indicating an improved estimate of y( x ). By comparing Phases I and III we see that the quality of i)(x) has improved significantly in the low-noise regions (roughly x < 0.6) at a minor sacrifice of accuracy in the high-noise region. Example #2: We now apply our method to a set of observed data, the 1000-point laser 2Purther details: all inputs are scaled to zero mean and unit variance. All initial weights feeding into hidden units are drawn from a uniform distribution between -1 j i and 1 j i where i is the number of incoming connections. All initial weights feeding into y or &2 are drawn from a uniform distribution between -sji and sji where s is the standard deviation of the (overall) target distribution. No momentum is used, and all weight updates are averaged over the forward passes of 20 patterns. 494 David A. Nix. Andreas S. Weigend (bl Phaao 1 Phaoolll Phalol _II _"I x x x 1~ 0.5 ~ Phaloll : 0 . -0.5 .? -1 o . ?.\ x 1 x x 11(1-11 Figure 2: (a) Example #1: Results after each phase of training. The top row gives the true y(x) (solid line) and network estimate y(x) (dotted line); the bottom row gives the true oo2(x) (solid line) and network estimate o-2(x) (dotted line). (b) Example #2: state-space embedding of laser data (evaluation set) using linear grey-scaling of 0.50 (lightest) < o-(Xt) < 6.92 (darkest). See text for details. intensity series from the Santa Fe competition. 3 Since our method is based on the network's observed errors, the predicted error a- 2 (x) actually represents the sum of the underlying system noise, characterized by 00 2 (x), and the model misspecification. Here, since we know the system noise is roughly uniform 8-bit sampling resolution quantization error, we can apply our method to evaluate the local quality of the manifold approximation. 4 The prediction task is easier if we have more points that lie on the manifold, thus better constraining its shape. In the competition, Sauer (1994) upsampled the 1000 available data points with an FFf method by a factor of 32. This does not change the effective sampling rate, but it "fills in" more points, more precisely defining the manifold. We use the same upsampling trick (without filtered embedding), and obtain 31200 full (x, d) patterns for learning. We apply the three-phase approach described above for the simple network of Figure 1 with 25 inputs (corresponding to 25 past values), 12 hidden units feeding the y-unit, and a libera130 hidden units feeding the a- 2 -unit (since we are uncertain as to the complexity of (j2(x) for this dataset). We use 11 = 10- 7 for Phase I and 11 = 10- 10 for Phases II and III. Since we know the quantization error is ?O. 5, error estimates less than this are meaningless. Therefore, we enforce a minimum value of (j2(x) = 0.25 (the quantization error squared) on the squared errors in Phases II and III. 3The data set and several predictions and characterizations are described in the volume edited by Weigend & Gershenfeld (1994). The data is available by anonymous ftp at ftp.cs.colorado.edu in /pub/Time-Series/SantaFe as A.dat. See also http://www . cs. colorado. edu/Time-Series/TSWelcome. html for further analyses of this and other time series data sets. 4When we make a single-step prediction where the manifold approximation is poor, we have little confidence making iterated predictions based on that predicted value. However, if we know we are in a low-error region, we can have increased confidence in iterated predictions that involve our current prediction. 495 Learning Local Error Bars for Nonlinear Regression Table 2: Results for Example #2 (See Table 1 caption for definitions) . I Training (N -- 975) I Evaluation (N - 23 950) row I . EMMC: I 3 Phase I Phase II Phase III 4 P( a( x ) . residual errors) p 0.557 % of errors < a( x); 2a( x) (exacl Gaussian) 1 sid 69A 68.3 2 0.00125 0.00125 0.00132 Mean <XlSt 1.941 1.939 1.725 E1"Juc: 0.0156 0.0156 0.0139 Mean cost 7.213 5.628 5.564 P 0.366 2std 94.9 95.4 1 sid 63.1 68.3 2 std 88.0 95.4 Results are given in Table 2 for patterns generated from the available 1000 points and 24,000 additional points used for evaluation. Even though we have used a Gaussian error model, we know the distribution of errors is not Gaussian. This is reflected in rows 5 and 6, where the training data is modeled as having a Gaussian em but the evaluation values become considerably distorted from an exact Gaussian. Again, however, not only do we obtain significant predictability of the errors, but the method also reduces the squared-error measure obtained in Phase I. We can use the estimated error to characterize the quality of the manifold approximation on 24,000 post-training evaluation points, as illustrated in Figure 2b. The manifold approximation is poorer (darker) for higher predicted values of Xt and for values nearer the edge of the manifold. Note the dark-grey (high-error) vertical streak leaving the origin and dark points to its left which represent patterns involving sudden changes in oscillation intensity. 3 Discussion Since we are in effect approximating two functions simultaneously, we can apply many of the existing variations for improving function approximation designed for networks learning only y(x). For example, when using limited amounts of data, especially if it is noisy, the particular split of data into training and cross-validation sets we use introduces significant variation in the resulting y(x) due to overfitting, as demonstrated on financial data by Weigend & LeBaron (1994). If we want to estimate local error bars, not only must we fear overfitting y(x), but we must also be concerned with overfitting &.2 (x). If the standard method of stopping at the minimum of an appropriate cross-validation set does not suffice for a given problem, it is straightforward to employ the usual anti-overfitting weaponry (smooth &2 as a function of x, pruning, weight-elimination, etc.). Furthermore, we can bootstrap over our available dataset and create multiple composite networks, averaging their predictions for both y(x) and &2(X) . Additionally, to incorporate prior information in a Bayesian framework as a form of regularization, Wolpert (personal communication, 1994) suggests finding the maximum a posteriori (instead of maximum-likelihood) conditional mean and variance using the same interpretation of the network outputs. In summary, we start with the maximum-likelihood principle and arrive at error estimates that vary with location in the input space. These local error estimates incorporate both underlying system noise and model misspecification. We have provided a computergenerated example to demonstrate the ease with which accurate error bars can be learned. We have also provided an example with real-world data in which the underlying system noise is small, uniform quantization error to demonstrate how the method can be used 496 David A. Nix, Andreas S. Weigend to characterize the local quality of the regression model. A significant feature of this method is its weighted-regression effect, which complicates learning by introducing local minima but can be potentially beneficial in constructing a more robust model with improved generalization abilities. In the framework presented, for any problem we must assume a specific parameterized cm then add one auxiliary output unit for each higher moment of the cm we wish to estimate locally. Here we have demonstrated the Gaussian case with a location parameter (conditional mean) and a scale parameter (local error bar) for a scalar output variable. The extension to multiple output variables is clear and allows a full covariance matrix to be used for weighted regression, including the cross-correlation between multiple targets. Acknowledgments This work is supported by the National Science Foundation under Grant No. RIA ECS-9309786 and by a Graduate Fellowship from the Office of Naval Research. We would like to thank Chris Bishop, Wray BUntine, Don Hush, Steve Nowlan, Barak Pearlmutter, Dave Rumelhart, and Dave Wolpert for helpful discussions. References C. Bishop. (1994) "Mixture Density Networks." Neural Computing Research Group Report NCRG/4288, Department of Computer Science, Aston University, Birmingham, UK. w.L. Buntine and A.S. Weigend. (1991) "Bayesian Backpropagation." Complex Systems, 5: 603--643 . A.M. Fraser and A. Dimitriadis. (1994) "Forecasting Probability Densities Using Hidden Markov Models with Mixed States." In Time Series Prediction: Forecasting the Future and Understanding the Past, A.S. Weigend and N.A. Gershenfeld. eds .. Addison-Wesley, pp. 265- 282. P.T. Kazlas and A.S. Weigend. (1995) "Direct Multi-Step TIme Series Prediction Using TD(>.)." In Advances in Neural Infonnation Processing Systems 7 (NIPS*94, this volume). San Francisco, CA: Morgan Kaufmann. D.E. Rumelhart, R. Durbin. R. Golden. and Y. Chauvin. (1995) "Backpropagation: The Basic Theory." In Backpropagation: Theory, Architectures and Applications, Y. Chauvin and D.E. Rumelhart, eds., Lawrence Erlbaum, pp. 1- 34. T. Sauer. (1994) 'T1l11.e Series Prediction by Using Delay Coordinate Embedding." In Time Series Prediction: Forecasting the Future and Understanding the Past, A.S. Weigend and N.A. Gershenfeld. eds., Addison-Wesley, pp. 175-193. A.N. Srivastava and A.S. Weigend. (1994) "Computing the Probability Density in Connectionist Regression." In Proceedings of the IEEE International Conference on Neural Networks (IEEEICNN'94), Orlando, FL, p. 3786--3789. IEEE-Press. A.S. Weigend and N.A. Gershenfeld, eds. (1994) Time Series Prediction: Forecasting the Future and Understanding the Past. Addison-Wesley. A.S. Weigend and B. LeBaron. (1994) "Evaluating Neural Network Predictors by Bootstrapping." In Proceedings of the International Conference on Neural Infonnation Processing (ICONIP'94), Seoul, }(orea,pp.1207-1212. A.S. Weigend and D.A. Nix. (1994) "Predictions with Confidence Intervals (Local Error Bars)." In Proceedings of the International Conference on Neural Information Processing (ICONIP'94), Seoul, }(orea, p. 847- 852.
896 |@word grey:2 covariance:1 solid:2 moment:1 initial:6 series:11 pub:2 seriously:1 past:4 existing:1 current:1 comparing:1 nowlan:1 activation:3 yet:1 must:5 additive:1 subsequent:1 shape:2 drop:1 designed:1 update:6 v:1 half:2 ria:1 underestimating:1 filtered:1 sudden:1 provides:2 characterization:1 location:2 direct:1 become:1 fitting:1 sacrifice:1 expected:2 indeed:1 roughly:3 frequently:1 mechanic:1 multi:1 discounted:1 td:1 little:2 actual:1 increasing:1 provided:2 estimating:1 underlying:5 suffice:1 what:1 cm:7 interpreted:3 finding:2 differentiation:1 bootstrapping:1 golden:1 xd:2 exactly:1 scaled:1 uk:1 normally:3 unit:43 grant:1 positive:1 before:1 dropped:2 local:16 troublesome:1 id:1 conversely:1 suggests:1 co:2 ease:1 limited:2 graduate:1 averaged:2 acknowledgment:1 backpropagation:5 bootstrap:1 significantly:1 composite:2 confidence:4 upsampled:1 risk:1 applying:2 influence:1 www:2 equivalent:1 demonstrated:2 maximizing:1 straightforward:1 resolution:1 simplicity:1 rule:3 fill:1 pull:1 financial:1 embedding:3 anyway:1 traditionally:1 variation:2 coordinate:1 target:14 colorado:8 exact:3 caption:1 ixi:1 us:1 origin:1 trick:1 rumelhart:5 approximated:1 std:2 binning:1 observed:3 invented:1 bottom:1 region:8 connected:5 edited:1 complexity:1 personal:1 trained:2 depend:2 santafe:1 basis:1 completely:1 multimodal:1 easily:1 laser:4 train:2 effective:2 monte:1 artificial:1 otherwise:1 ability:1 itself:3 noisy:1 frozen:1 j2:8 flexibility:1 competition:3 p:1 produce:1 ftp:4 derive:1 minor:1 nlx:1 eq:7 auxiliary:5 c:6 predicted:5 come:1 involves:1 fff:1 elimination:1 require:1 feeding:4 orlando:1 generalization:3 anonymous:1 extension:1 around:1 considered:2 exp:2 great:1 lawrence:1 mapping:1 predict:1 vary:2 early:1 a2:1 birmingham:1 tanh:3 infonnation:2 create:1 weighted:6 minimization:1 gaussian:14 avoid:1 cr:2 hj:1 varying:1 office:1 derived:2 naval:1 likelihood:6 hk:1 contrast:2 underscore:1 cr2:2 posteriori:1 helpful:1 stopping:1 typically:1 entire:1 initially:2 hidden:20 going:1 iu:1 overall:2 html:2 ill:1 constrained:1 summed:1 equal:1 construct:1 having:1 sampling:2 represents:2 future:3 connectionist:3 report:1 employ:2 randomly:1 simultaneously:1 national:1 familiar:1 phase:35 highly:2 evaluation:7 introduces:1 mixture:4 extreme:2 tj:3 chain:1 accurate:1 poorer:1 edge:1 experience:1 sauer:2 logarithm:1 desired:1 re:1 uncertain:1 complicates:1 increased:1 modeling:1 cost:13 introducing:1 deviation:2 uniform:5 predictor:1 delay:1 erlbaum:1 too:1 buntine:3 characterize:2 straightforwardly:1 considerably:1 density:5 international:3 off:1 picking:1 connectivity:3 squared:13 again:2 choose:2 cognitive:2 derivative:1 leading:1 bx:2 li:1 coefficient:1 explicitly:1 root:2 reached:1 start:2 bayes:1 minimize:1 formed:1 square:1 accuracy:1 variance:12 kaufmann:1 correspond:1 sid:6 bayesian:2 iterated:2 emphasizes:1 wray:1 carlo:1 dave:2 ed:4 definition:1 pp:4 naturally:1 di:5 monitored:1 stop:1 dataset:2 color:1 fractional:1 improves:1 actually:2 feed:1 steve:1 higher:3 wesley:3 reflected:1 specify:1 improved:3 though:1 furthermore:1 just:1 correlation:4 receives:1 nonlinear:7 quality:4 effect:3 concept:1 true:4 regularization:1 illustrated:1 sin:3 during:1 width:1 inferior:1 generalized:1 iconip:2 complete:3 demonstrate:4 pearlmutter:1 l1:1 hungry:1 iin:1 naf:1 functional:1 volume:3 ncrg:1 discussed:1 interpretation:1 significant:6 freeze:1 access:1 etc:1 add:3 own:2 sji:2 morgan:1 minimum:6 additional:3 impose:1 maximize:1 monotonically:1 ii:9 full:4 unimodal:2 multiple:3 reduces:1 smooth:1 characterized:1 cross:5 divided:1 post:2 e1:1 fraser:2 prediction:15 involving:1 regression:14 basic:1 represent:3 tailored:1 want:3 fellowship:1 interval:2 leaving:1 meaningless:1 pass:1 thing:1 near:1 ideal:2 constraining:2 split:3 iii:8 decent:1 concerned:1 independence:1 xj:1 architecture:6 suboptimal:1 andreas:7 reduce:2 forecasting:4 santa:2 listed:2 involve:1 clear:1 amount:1 discount:1 dark:2 locally:2 http:2 generate:2 percentage:1 dotted:2 lightest:1 estimated:1 per:1 write:1 group:1 drawn:3 gershenfeld:4 sum:2 weigend:19 parameterized:4 distorted:1 place:2 arrive:1 home:1 oscillation:1 scaling:1 bit:1 layer:7 bound:1 fl:1 durbin:1 precisely:1 erroneously:1 attempting:1 ern:1 department:3 combination:1 poor:1 beneficial:2 smaller:1 em:4 modification:1 making:1 outlier:1 boulder:2 taken:1 equation:4 resource:1 know:4 addison:3 end:4 available:10 gaussians:1 apply:5 away:1 appropriate:2 enforce:1 darkest:1 assumes:1 denotes:1 top:1 wax:1 especially:1 approximating:2 dat:1 bl:1 overemphasized:1 already:1 quantity:1 enm:1 parametric:2 usual:1 traditional:2 cid:2 gradient:5 separate:4 thank:1 upsampling:1 chris:1 manifold:7 chauvin:2 assuming:3 modeled:1 dimitriadis:2 equivalently:1 fe:2 potentially:1 negative:2 ordinarily:1 andw:1 vertical:1 markov:2 nix:8 descent:4 anti:1 heterogeneity:1 defining:1 communication:1 misspecification:4 intensity:2 david:5 pair:1 connection:2 learnillg:1 learned:1 hush:1 nearer:1 nip:1 bar:13 c1d:2 pattern:17 including:1 overlap:1 natural:1 attach:1 predicting:1 residual:5 aston:1 lk:1 text:1 prior:1 understanding:3 ina:1 mixed:2 validation:4 foundation:1 principle:1 share:1 row:8 summary:1 supported:1 bias:1 allow:1 barak:1 institute:2 distributed:3 world:1 evaluating:1 forward:2 interchange:1 made:1 san:1 avoided:2 ec:1 approximate:2 emphasize:1 pruning:1 keep:1 global:1 overfitting:6 incoming:1 assumed:4 francisco:1 xi:20 don:1 table:6 additionally:1 learn:3 transfer:1 robust:1 ca:1 streak:1 obtaining:3 improving:1 lebaron:2 complex:1 constructing:1 motivation:1 noise:19 allowed:1 darker:1 predictability:1 momentum:1 wish:2 explicit:1 exponential:3 xl:2 lie:1 ix:1 bishop:6 specific:3 xt:2 er:1 offset:2 list:1 quantization:5 adding:1 conditioned:1 justifies:1 occurring:1 easier:1 wolpert:2 simply:3 scalar:1 fear:1 corresponds:2 conditional:18 viewed:1 shared:1 change:2 typical:1 reducing:1 uniformly:1 averaging:1 yand:1 total:1 pas:1 indicating:1 maximumlikelihood:1 seoul:2 incorporate:2 evaluate:1 srivastava:2
7,131
897
PREDICTIVE CODING WITH NEURAL NETS: APPLICATION TO TEXT COMPRESSION J iirgen Schmidhuber Stefan Heil Fakultat fiir Informatik Technische Universitat Miinchen 80290 Miinchen, Germany Abstract To compress text files, a neural predictor network P is used to approximate the conditional probability distribution of possible "next characters", given n previous characters. P's outputs are fed into standard coding algorithms that generate short codes for characters with high predicted probability and long codes for highly unpredictable characters. Tested on short German newspaper articles, our method outperforms widely used Lempel-Ziv algorithms (used in UNIX functions such as "compress" and "gzip"). 1048 1 liirgen Schmidhuber, Stefan Heil INTRODUCTION The method presented in this paper is an instance of a strategy known as "predictive coding" or "model-based coding". To compress text files, a neural predictor network P approximates the conditional probability distribution of possible "next characters", given n previous characters. P's outputs are fed into algorithms that generate short codes for characters with low information content (characters with high predicted probability) and long codes for characters conveying a lot of information (highly unpredictable characters) [5]. Two such standard coding algorithms are employed: Huffman Coding (see e.g. [1]) and Arithmetic Coding (see e.g. [7]). With the off-line variant of the approach, P's training phase is based on a set F of training files. After training, the weights are frozen. Copies of P are installed at all machines functioning as message receivers or senders. From then on, P is used to encode and decode unknown files without being changed any more. The weights become part of the code of the compression algorithm. Note that the storage occupied by the network weights does not have to be taken into account to measure the performance on unknown files - just like the code for a conventional data compression algorithm does not have to be taken into account. The more sophisticated on-line variant of our approach will be addressed later. 2 A PREDICTOR OF CONDITIONAL PROBABILITIES Assume that the alphabet contains k possible characters Zl, Z2, ?.? , Z1c. The (local) representation of Zi is a binary k-dimensional vector r( Zi) with exactly one non-zero component (at the i-th position). P has nk input units and k output units. n is called the "time-window size". We insert n default characters Zo at the beginning of each file. The representation of the default character, r(zo), is the k-dimensional zero-vector. The m-th character of file f (starting from the first default character) is called efn. For all f E F and all possible m > n, P receives as an input r(e;"_n) 0 r(e;"_n+l) 0 ... 0 r(c!n_l), (1) where 0 is the concatenation operator, for vectors. P produces as an output Pin, a k-dimensional output vector. Using back-propagation [6][2][3][4], P is trained to mmlIDlze ~L L II r(c!n) - pin 112 . (2) jEFm>n Expression (2) is minimal if pin always equals E(r(efn) I e;"_n,?? ?,c!n-l), (3) the conditional expectation of r( efn), given r( e;"_n) ore e;"_n+1)o . . . or( c!n-l). Due to the local character representation, this is equivalent to (Pin); being equal to the 1049 Predictive Coding with Neural Nets conditional probability (4) for all / and for all appropriate m> n, where (P,{Jj denotes the j-th component of the vector P/n. In general, the (P/n)i will not quite match the corresponding conditional probabilities. For normalization purposes, we define _ PmI (.) 1 - (P/n)i f? L:j=I(Pm)j j: (5) No normalization is used during training, however. 3 HOW TO USE THE PREDICTOR FOR COMPRESSION We use a standard procedure for predictive coding. With the help of a copy of P, an unknown file / can be compressed as follows: Again, n default characters are inserted at the beginning. For each character cfn (m> n), the predictor emits its output P/n based on the n previous characters. There will be a k such that cfn = Zj:. The estimate of P(cfn = Zj: I Crn-l) is given by P/n(k). The code of cfn, code( cfn), is generated by feeding P/n (k) into the Huffman Coding algorithm (see below), or, alternatively, into the Arithmetic Coding algorithm (see below). code(cfn) is written into the compressed file. The basic ideas of both coding algorithms are described next. c!n-n, ... , 3.1 HUFFMAN CODING With a given probability distribution on a set of possible characters, Huffman Coding (e.g. [1]) encodes characters by bitstrings as follows. Characters are terminal nodes of a binary tree to be built in an incremental fashion. The probability of a terminal node is defined as the probability of the corresponding character. The probability of a non-terminal node is defined as the sum of the probabilities of its sons. Starting from the terminal nodes, a binary tree is built as follows: Repeat as long as possible: Among those nodes that are not children 0/ any non-terminal nodes created earlier, pick two with lowest associated probabilities. Make them the two sons 0/ a newly generated non-terminal node. The branch to the "left" son of each non-terminal node is labeled by a O. The branch to its "right" son is labeled by a 1. The code of a character c, code(c), is the bitstring obtained by following the path from the root to the corresponding node. Obviously, if c #- d, then code(c) cannot be the prefix of code(d). This makes the code uniquely decipherable. 1050 Jurgen Schmidhuber, Stefan Heil Characters with high associated probability are encoded by short bitstrings. Characters with low associated probability are encoded by long bitstrings. Huffman Coding guarantees minimal expected code length, provided all character probabilities are integer powers of ~. 3.2 ARlTHMETIC CODING In general, Arithmetic Coding works slightly better than Huffman Coding. For sufficiently long messages, Arithmetic Coding achieves expected code lenghts arbitrarily close to the information-theoretic lower bound. This is true even if the character probabilities are not powers of ~ (see e.g. [7]) . The basic idea of Arithmetic Coding is: a message is encoded by an interval of real numbers from the unit interval [0,1[. The output of Arithmetic Coding is a binary representation of the boundaries of the corresponding interval. This binary representation is incrementally generated during message processing. Starting with the unit interval, for each observed character the interval is made smaller, essentially in proportion to the probability of the character. A message with low information content (and high corresponding probability) is encoded by a comparatively large interval whose precise boundaries can be specified with comparatively few bits. A message with a lot of information content (and low corresponding probability) is encoded by a comparatively small interval whose boundaries require comparatively many bits to be specified. Although the basic idea is elegant and simple, additional technical considerations are necessary to make Arithmetic Coding practicable. See [7] for details. Neither Huffman Coding nor Arithmetic Coding requires that the probability distribution on the characters remains fixed. This allows for using "time-varying" conditional probability distributions as generated by the neural predictor. 3.3 HOW TO "UNCOMPRESS" DATA The information in the compressed file is sufficient to reconstruct the original file without loss of information. This is done with the "uncompress" algorithm, which works as follows: Again, for each character efn (m > n), the predictor (sequentially) emits its output based on the n previous characters, where the e{ with n < I < m were gained sequentially by feeding the approximations (k) of the probabilities P(e{ ZIc I e{-n,???, e{-l) into the inverse Huffman Coding procedure (see e.g. [1]), or, alternatively (depending on which coding procedure was used), into the inverse Arithmetic Coding procedure (e.g. [7]). Both variants allow for correct decoding of c{ from eode(c{) . With both variants, to correctly decode some character, we first need to decode all previous characters. Both variants are guaranteed to restore the original file from the compressed file. pin = p/ WHY NOT USE A LOOK-UP TABLE INSTEAD OF A NETWORK? Because a look-up table would be extremely inefficient. A look-up table requires k n +1 entries for all the conditional probabilities corresponding to all possible com- Predictive Coding with Neural Nets 1051 binations of n previous characters and possible next characters. In addition, a special procedure is required for dealing with previously unseen combinations of input characters. In contrast, the size of a neural net typically grows in proportion to n 2 (assuming the number of hidden units grows in proportion to the number of input units), and its inherent "generalization capability" is going to take care of previously unseen combinations of input characters (hopefully by coming up with good predicted probabilities). 4 SIMULATIONS We implemented both alternative variants of the encoding and decoding procedure described above. Our current computing environment prohibits extensive experimental evaluations of the method. The predictor updates turn out to be quite time consuming, which makes special neural net hardware recommendable. The limited software simulations presented in this section, however, will show that the "neural" compression technique can achieve "excellent" compression ratios. Here the term "excellent" is defined by a statement from [1]: "In general, good algorithms can be expected to achieve an average compression ratio of 1.5, while excellent algorithms based upon sophisticated processing techniques will achieve an average compression ratio exceeding 2.0." Here the average compression ratio is the average ratio between the lengths of original and compressed files. The method was applied to German newspaper articles. The results were compared to those obtained with standard encoding techniques provided by the operating system UNIX, namely "pack", "compress", and "gzip" . The corresponding decoding algorithms are "unpack", "uncompress", and "gunzip", respectively. ''pack'' is based on Huffman-Coding (e.g. [1]), while "compress" and "gzip" are based on techniques developed by Lempel and Ziv (e.g. [9]). As the file size goes to infinity, Lempel-Ziv becomes asymptotically optimal in a certain information theoretic sense [8]. This does not necessarily mean, however, that Lempel-Ziv is optimal for finite file sizes. The training set for the predictor was given by a set of 40 articles from the newspaper Miinchner M erkur, each containing between 10000 and 20000 characters. The alphabet consisted of k 80 possible characters, including upper case and lower case letters, digits, interpunction symbols, and special German letters like "0", "ii", "a.". P had 430 hidden units. A "true" unit with constant activation 1.0 was connected to all hidden and output units. The learning rate was 0.2. The training phase consisted of 25 sweeps through the training set. = The test set consisted of newspaper articles excluded from the training set, each containing between 10000 and 20000 characters. Table 1 lists the average compression ratios. The "neural" method outperformed the strongest conventional competitor, the UNIX "gzip" function based on a Lempel-Ziv algorithm. 1052 Jurgen Schmidhuber, Stefan Heil Method Huffman Coding (UNIX: pack) Lempel-Ziv Coding (UNIX: compress) Improved Lempel-Ziv ( UNIX: gzip -9) Neural predictor + Huffman Coding, n 5 Neural predictor + Arithmetic Coding, n 5 = = Compression Ratio 1.74 1.99 2.29 2.70 2.72 I Table 1: Compression ratios of various compression algorithms for short German 20000 Bytes) from the unknown test set. text files ? Method Huffman Coding (UNIX: pack) Lempel-Ziv Coding (UNIX: compress) Improved Lempel-Zlv ( UNIX: gzip -9) Neural predictor + Huffman Coding, n 5 Neural predictor + Arithmetic Coding, n 5 = = Compression Ratio 1.67 1.71 2.03 2.25 2.20 I Table 2: Compression ratios for articles from a different newspaper. The neural predictor was not retrained. How does a neural net trained on articles from Miinchner Merkurperform on articles from other sources? Without retraining the neural predictor, we applied all competing methods to 10 articles from another German newspaper (the Frankenpost). The results are given in table 2. The Frankenpost articles were harder to compress for all algorithms. But relative performance remained comparable. Note that the time-window was quite small (n = 5). In general, larger time windows will make more information available to the predictor. In turn, this will improve the prediction quality and increase the compression ratio. Therefore we expect to obtain even better results for n > 5 and for recurrent predictor networks. 5 DISCUSSION / OUTLOOK Our results show that neural networks are promising tools for loss-free data compression. It was demonstrated that even off-line methods based on small time windows can lead to excellent compression ratios - at least with small text files, they can outperform conventional standard algorithms. We have hardly begun, however, to exhaust the potential of the basic approach. 5.1 ON-LINE METHODS A disadvantage of the off-line technique above is that it is off-line: The predictor does not adapt to the specific text file it sees. This limitation is not essential, however. It is straight-forward to construct an on-line variant of the approach. Predictive Coding with Neural Nets 1053 With the on-line variant, the predictor continues to learn during compression. The on-line variant proceeds like this: Both the sender and the receiver start with exactly the same initial predictor. Whenever the sender sees a new character, it encodes it using its current predictor. The code is sent to the receiver who decodes it. Both the sender and the receiver use exactly the same learning protocol to modify their weights. This implies that the modified weights need not be sent from the sender to the receiver and do not have to be taken into account to compute the average compression ratio. Of course, the on-line method promises much higher compression ratios than the off-line method. 5.2 LIMITATIONS The main disadvantage of both on-line and off-line variants is their computational complexity. The current off-line implementation is clearly slower than conventional standard techniques, by about three orders of magnitude (but no attempt was made to optimize the code with respect to speed). And the complexity of the on-line method is even worse (the exact slow-down factor depends on the precise nature of the learning protocol, of course). For this reason, especially the promising on-line variants can be recommended only if special neural net hardware is available. Note, however, that there are many commercial data compression applications which rely on specialized electronic chips. 5.3 ONGOING RESEARCH There are a few obvious directions for ongoing experimental research: (1) Use larger time windows - they seem to be promising even for off-line methods (see the last paragraph of section 4). (2) Thoroughly test the potential of on-line methods. Both (1) and (2) should greatly benefit from fast hardware. (3) Compare performance of predictive coding based on neural predictors to the performance of predictive coding based on different kinds of predictors. 6 ACKNOWLEDGEMENTS Thanks to David MacKay for directing our attention towards Arithmetic Coding. Thanks to Margit Kinder, Martin Eldracher, and Gerhard Weiss for useful comments. 1054 Jiirgen Schmidhuber, Stefan Heil References [1] G. Held. Data Compression. Wiley and Sons LTD, New York, 1991. [2] Y. LeCun. Une procedure d'apprentissage pour reseau a. seuil asymetrique. Proceedings of Cognitiva 85, Paris, pages 599-604, 1985. [3] D. B. Parker. Learning-logic. Technical Report TR-47, Center for Compo Research in Economics and Management Sci., MIT, 1985. [4] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Learning internal representations by error propagation. In Parallel Distributed Processing, volume 1, pages 318-362. MIT Press, 1986. [5] J. H. Schmidhuber and S. Heil. Sequential neural text compression. IEEE Transactions on Neural Networks, 1994. Accepted for publication. [6] P. J. Werbos. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. PhD thesis, Harvard University, 1974. [7] I. H. Witten, R . M. Neal , and J. G. Cleary. Arithmetic coding for data compression. Communications of the ACM, 30(6):520-540, 1987. [8] A. Wyner and J . Ziv. Fixed data base version of the Lempel-Ziv data compression algorithm. IEEE Transactions In/ormation Theory, 37:878-880, 1991. [9] J. Ziv and A. Lempel. A universal algorithm for sequential data compression. IEEE Transactions on Information Theory, IT-23(5):337-343, 1977.
897 |@word version:1 compression:27 proportion:3 retraining:1 simulation:2 pick:1 tr:1 outlook:1 cleary:1 harder:1 initial:1 contains:1 prefix:1 outperforms:1 current:3 z2:1 com:1 activation:1 written:1 update:1 une:1 beginning:2 short:5 compo:1 node:9 miinchen:2 become:1 behavioral:1 paragraph:1 expected:3 pour:1 nor:1 terminal:7 unpredictable:2 window:5 becomes:1 provided:2 lowest:1 kind:1 prohibits:1 developed:1 guarantee:1 exactly:3 zl:1 unit:9 local:2 modify:1 installed:1 encoding:2 path:1 limited:1 lecun:1 digit:1 procedure:7 universal:1 cannot:1 close:1 operator:1 storage:1 optimize:1 equivalent:1 conventional:4 demonstrated:1 center:1 go:1 attention:1 starting:3 economics:1 williams:1 seuil:1 commercial:1 gerhard:1 decode:3 exact:1 harvard:1 rumelhart:1 continues:1 werbos:1 labeled:2 observed:1 inserted:1 connected:1 ormation:1 environment:1 complexity:2 trained:2 predictive:8 upon:1 chip:1 various:1 alphabet:2 zo:2 fast:1 quite:3 encoded:5 widely:1 whose:2 larger:2 reconstruct:1 compressed:5 unseen:2 obviously:1 frozen:1 net:8 coming:1 achieve:3 produce:1 incremental:1 help:1 depending:1 recurrent:1 jurgen:2 implemented:1 predicted:3 implies:1 direction:1 correct:1 require:1 feeding:2 generalization:1 insert:1 sufficiently:1 achieves:1 purpose:1 outperformed:1 bitstring:1 tool:2 stefan:5 mit:2 clearly:1 always:1 modified:1 occupied:1 zic:1 varying:1 publication:1 encode:1 greatly:1 contrast:1 sense:1 typically:1 hidden:3 going:1 germany:1 among:1 ziv:11 special:4 mackay:1 equal:2 construct:1 look:3 cfn:6 report:1 inherent:1 few:2 phase:2 attempt:1 message:6 highly:2 evaluation:1 held:1 necessary:1 tree:2 minimal:2 instance:1 earlier:1 disadvantage:2 technische:1 entry:1 predictor:23 universitat:1 thoroughly:1 thanks:2 off:8 decoding:3 again:2 thesis:1 management:1 containing:2 worse:1 inefficient:1 account:3 potential:2 coding:42 exhaust:1 depends:1 later:1 root:1 lot:2 start:1 capability:1 parallel:1 recommendable:1 who:1 conveying:1 decodes:1 informatik:1 straight:1 strongest:1 whenever:1 competitor:1 obvious:1 jiirgen:1 associated:3 emits:2 newly:1 begun:1 sophisticated:2 back:1 higher:1 improved:2 wei:1 done:1 just:1 receives:1 hopefully:1 propagation:2 incrementally:1 quality:1 grows:2 consisted:3 true:2 functioning:1 excluded:1 neal:1 during:3 uniquely:1 gzip:6 theoretic:2 consideration:1 bitstrings:3 specialized:1 witten:1 volume:1 approximates:1 pmi:1 pm:1 had:1 operating:1 base:1 schmidhuber:6 certain:1 binary:5 arbitrarily:1 additional:1 care:1 employed:1 recommended:1 arithmetic:13 ii:2 branch:2 technical:2 match:1 adapt:1 unpack:1 long:5 prediction:2 variant:11 basic:4 regression:1 essentially:1 expectation:1 normalization:2 huffman:13 addition:1 ore:1 addressed:1 interval:7 source:1 cognitiva:1 file:19 comment:1 elegant:1 sent:2 seem:1 integer:1 zi:2 competing:1 idea:3 expression:1 ltd:1 york:1 jj:1 hardly:1 useful:1 hardware:3 generate:2 outperform:1 zj:2 correctly:1 promise:1 neither:1 asymptotically:1 sum:1 inverse:2 unix:9 letter:2 electronic:1 comparable:1 bit:2 bound:1 guaranteed:1 infinity:1 software:1 encodes:2 speed:1 extremely:1 martin:1 combination:2 smaller:1 slightly:1 son:5 character:43 practicable:1 taken:3 remains:1 previously:2 pin:5 german:5 turn:2 fed:2 available:2 lempel:11 appropriate:1 alternative:1 slower:1 original:3 compress:8 denotes:1 n_l:1 lenghts:1 especially:1 comparatively:4 sweep:1 strategy:1 sci:1 concatenation:1 reason:1 assuming:1 code:18 length:2 ratio:14 statement:1 implementation:1 unknown:4 upper:1 finite:1 hinton:1 communication:1 precise:2 directing:1 retrained:1 david:1 namely:1 required:1 specified:2 extensive:1 paris:1 fakultat:1 efn:4 beyond:1 proceeds:1 below:2 built:2 including:1 power:2 rely:1 restore:1 improve:1 wyner:1 created:1 heil:6 text:7 byte:1 acknowledgement:1 relative:1 loss:2 expect:1 limitation:2 sufficient:1 apprentissage:1 article:9 course:2 changed:1 repeat:1 last:1 copy:2 free:1 allow:1 benefit:1 distributed:1 boundary:3 default:4 forward:1 made:2 newspaper:6 transaction:3 approximate:1 fiir:1 dealing:1 logic:1 sequentially:2 receiver:5 consuming:1 alternatively:2 crn:1 why:1 table:7 promising:3 pack:4 learn:1 nature:1 excellent:4 z1c:1 necessarily:1 protocol:2 main:1 child:1 fashion:1 parker:1 slow:1 wiley:1 position:1 exceeding:1 down:1 remained:1 specific:1 symbol:1 list:1 essential:1 sequential:2 gained:1 phd:1 magnitude:1 nk:1 sender:5 acm:1 conditional:8 towards:1 content:3 called:2 accepted:1 experimental:2 internal:1 ongoing:2 tested:1
7,132
898
Direction Selectivity In Primary Visual Cortex Using Massive Intracortical Connections Christof Koch CNS Program 216-76 Caltech Pasadena, CA 91125 Humbert Suarez CNS Program 216-76 Caltech Pasadena, CA 91125 Rodney Douglas MRC Anatomical Neuropharmacology Unit University of Oxford Oxford UK Abstract Almost all models of orientation and direction selectivity in visual cortex are based on feedforward connection schemes, where geniculate input provides all excitation to both pyramidal and inhibitory neurons. The latter neurons then suppress the response of the former for non-optimal stimuli. However, anatomical studies show that up to 90 % of the excitatory synaptic input onto any cortical cell is provided by other cortical cells. The massive excitatory feedback nature of cortical circuits is embedded in the canonical microcircuit of Douglas &. Martin (1991). We here investigate analytically and through biologically realistic simulations the functioning of a detailed model of this circuitry, operating in a hysteretic mode. In the model, weak geniculate input is dramatically amplified by intracortical excitation, while inhibition has a dual role: (i) to prevent the early geniculate-induced excitation in the null direction and (ii) to restrain excitation and ensure that the neurons fire only when the stimulus is in their receptive-field. Among the Humbert Suarez, Christo! Koch, Rodney Douglas 4 insights gained are the possibility that hysteresis underlies visual cortical function, paralleling proposals for short-term memory, and strong limitations on linearity tests that use gratings. Properties of visual cortical neurons are compared in detail to this model and to a classical model of direction selectivity that does not include excitatory corti co-cortical connections. The model explain a number of puzzling features of direction-selective simple cells, including the small somatic input conductance changes that have been measured experimentally during stimulation in the null direction. The model also allows us to understand why the velocity-response curve of area 17 neurons is different from that of their LG N afferents, and the origin of expansive and compressive nonlinearities in the contrast-response curve of striate cortical neurons. 1 INTRODUCTION Direction selectivity is the property of neurons that fire more strongly for one direction of motion of a bar (the preferred direction) than the other (nUll direction). It is one of the fundamental properties of neurons in visual cort~x and is intimately related to the processing of motion by the visual system. LGN neurons that provide input to visual cortex respond approximately symmetrically to motion in different directions; so cortical neurons must generate that direction specificity. Models of direction selectivity in primary visual. cortex generally overlook two important constraints. 1. at least 80% of excitatory synapses on cortical pyramidal cells originate from other pyramidal cells, and less than 10 % are thalamic afferents (Peters & Payne, 1993). 2. Intracellular in vivo recordings in cat simple cells by Douglas, et al. (1988) failed to detect significant changes in somatic input conductance during stimulation in the null direction, indicating that there is very little synaptic input to direction-selective neurons in that condition, including no massive "shunting" inhibition. One very attractive solution incorporating these two constraints was proposed by Douglas & Martin (1991) in the form of their canonical microcircuit: for motion of a visual stimulus in the preferred direction, weak geniculate excitation excites cortical pyramidal cells to respond moderately. This relatively small amount of cortical excitation is amplified via excitatory cortico-cortical connections. For motion in the null direction, the weak geniculate excitation is vetoed by weak inhibition (mediated via an interneuron) and the cortical loop is never activated. In order to test quantitatively this circuit against the large body of experimental data, it is imperative to model its operation through mathematical analysis and detailed simulations. 2 MODEL DESCRIPTION AND ANALYSIS The model consists of a retino-geniculate and a cortical stage (Fig. 1). The former includes a center-surround receptive field and bandpass temporal filtering (Victor, 1987). We simulate a I-D array of ON LGN neurons, with 208 LGN neurons at each of 6 spatial positions in this array. The output ofthe LGN-action potentials-feeds Direction Selectivity in Primary Visual Cortex 5 into the cortical module consisting of 640 pyramidal (excitatory) and 160 inhibitory neurons. Each neuron is modeled using 3-4 compartments whose parameters reflect, in a simplified way, the biophysics and morphology of cortical neurons; the somatic compartment produces action potentials in response to current injection. There are excitatory connections among all pyramidal neurons, and inhibitory connections from the inhibitory population to itself and to the pyramidal population. The receptive field of the geniculate input to the inhibitory cortical neurons is displaced in space from the geniculate input to the pyramidal neurons, so that in the null direction inhibition overlaps with geniculate excitation in the pyramidal neurons, resulting in direction selectivity in the pyramidal neurons. The time courses of the post-synaptic potentials (PSP's) are consistent with physiologically recorded PSP's. The EPSP's in our model arise exclusively from non-NMDA synapses, and the IPSP's originate from both GABAA and GABAB synapses. There are two operating modes of this cortical amplifier circuit, depending on parameter values. In the first mode, the pyramidal neurons' response increases proportionally to the stimulus strength over a substantial range of input values, before saturating. In the second mode, the response increases much faster over a narrow range of stimulus strengths, then saturates. Analytically, one can define a steadystate transfer function for the network of pyramidal neurons; then, in the first mode, the transfer function has a slope that is less than 1, so that the network's firing rate at equilibrium increases proportionally to the input strength and the network dynamics are rather slow. In the second mode the initial slope is larger than 1, so that the network can discharge briskly at equilibrium even without any input, show hysteresis, and has rather fast dynamics. The network does not fire without any input, because of the neuron's threshold. In this paper, we will show responses in that second, hysteretic mode of operation. We will compare the cortical amplifier model's response properties to a pure feedforward model, that has no excitatory connections between pyramidal neurons. In order to maintain strong responses in that model, the LGN was connected more strongly to the pyramidal neurons, and in order to maintain direction selectivity, the weights of the connections of the inhibitory neurons to the pyramids were increased as well. 3 AMPLIFICATION AND CONDUCTANCE CHANGE IN THE NULL DIRECTION The input conductance of pyramidal neurons, a measure of total synaptic input, changes by only 50 % in the cortical amplifier model versus 400 % in the feedforward model, and so is more consistent with physiology (Fig. 2). Indeed, most of the current causing firing in the preferred direction originates from other pyramidal neurons, so the connection weight from the LGN is small. Consequently, the inhibitory weight is also small, since it needs only be large enough to balance out the LG N current in the null direction. Since there is little firing in other pyramidal neurons in the null direction, there is also little total synaptic input to the fiducial cell. The cortical amplifier circuit provides substantial amplification of the LGN input. In the preferred direction, excitatory intracortical connections amplify the LGN input, providing a feedback current that is about 2.2 times larger than the 6 Humbert Suarez, Christo! Koch, Rodney Douglas 640 160 smooth cells AREA 17 208 neurons at each position LGN RETINA 111I11111111111111 ==-1111111111111111111 125 pixels Figure 1: Wiring diagram of the direction selectivity model. Input to LGN neurons comes from a one-dimensional array of retinal pixels. There are 208 LGN neurons at each of six spatial positions. The LG N neurons connect slightly differently with the two populations of cortical neurons (pyramidal and inhibitory) so that as a group the LGN inputs to the pyramids are displaced spatially by 5' from those to the inhibitory neurons. The open triangle symbols denote excitatory connections, the filled triangles inhibitory GABAA-mediated synapses, and the filled circles inhibitory GABAB-mediated synapses. The capital sigma symbol indicates convergence of inputs from many LGN neurons onto cortical neurons. Direction Selectivity in Primary Visual Cortex 7 0.5 b 04 ~ .s E ~ :; 1 0.3 r25 tOO ~:~I ~:':-~_.'~~::'>:~ :: ~"~:'~:':'.~:..~ .,)0.1:' 75 0.2 0 - 2sec 0 .1 0 0 0.1 0.2 0.3 0.5 0.4 Cumal 0 ? 2iec Tesl 0 .6 4()0 c 150 d 300 125 j ..... c 200 ~: 100 - :.....,...,..-..,.?? 100 o\V_,MJ _ ...: . .. . ; . :"-:~ .... ---..,;" '. ?o\...-ohl 75 501 - - - - - - - - - - ' o Time (seconds) Figure 2: (a) Total synaptic current in a simulated pyramidal cell from the LGN (with symbols) and from other pyramidal cells (continuous without symbols), during stimulation by a bar moving in the preferred direction . For the same stimulus moving in the null direction, somatic input conductance as a function of time, (b) for a direction-selective pyramidal neuron (data from Douglas et ai., 1988); (c) feedforward model; (d) cortical amplifier model; LGN current at 60 % contrast (Fig. 2). 4 CONTRAST-RESPONSE CURVES Contrast-response curves plot the peak firing rate to a grating moving in the preferred direction as function of its contrast, or stimulus amplitude (Albrecht & Hamilton, 1982). The cortical amplifier's contrast-response curve is very different from the LGN inputs' and is similar to those that have been described experimentally in cortex (Albrecht & Geisler 1991), having a steep power-function portion followed by abrupt saturation (Fig. 3). The network firing saturates at the fixed point of the transfer function mentioned above (see section 2) and the steep portion results from the fast rise to that fixed point when the stimulus has exceeded the neurons' threshold . In contrast, the feedforward model's contrast-response curve is similar to the LGN inputs' and does not match physiology. The response is very small in the null direction, resulting in very good direction selectivity at all contrasts (i.e., the average direction index is above 0.9). 5 VELOCITY-RESPONSE CURVES 8 Humbert Suarez, Christo! Koch, Rodney Douglas KlO a u 100 b S! '" <II .:a. L .. ~ 10 .. :< ~ 2' :,.; u. 1 1 100 100 10 Contrast('X.) 100 100 U- c S! . !.. 10 Contrast ('X.) <II ~ 10 ~ go :~ u. 1 1 U- ~ 10 Contrast ('X.) 100 d S! . <II 1 10 ~ 2' J; IL 1 1 Figure 3: Peak firing rate versus contrast during stimulation by a moving grating. (a) Visual cortical neuron. Data from Albrecht &, Hamilton, 1982. (b) LGN model. ( c) Feedforward model. (d) Cortical amplifier model. Velocity-response curves plot the peak response to a bar moving in the preferred direction as a function of its velocity. Again, the cortical amplifier's velocityresponse curve is very different from the LGN inputs' and is similar to physiology (Orban, 1984; Fig. 4). The LGN model is firing strongly at high velocities but the model pyramidal neurons are totally silent, due to a combination of GABAA inhibition, neuron threshold, and membrane low-pass filtering. At low velocities, the LGN model does not fire much, while the cortical neurons respond strongly. Indeed, the network firing has enough time to reach the fixed point of the transfer function and will reach it as long as the input is suprathreshold. In contrast, the feedforward model's velocity-response curve is again similar to the LGN input's and does not match physiology. Also shown in Fig. 4 is the response in the null direction for the cortical amplifier model. There is very good direction selectivity at all velocities, consistent with physiological data (Orban, 1984). The persistence of direction selectivity down to low velocities depends critically on the time constant of GABAB and the presence of a very small displacement between the LGN inputs to the pyramidal and inhibitory neurons. 6 OTHER PROPERTIES Recently, direction-selective cortical neurons have been tested for linearity using an intracellular grating superposition test and found to be quite linear (Jagadeesh et al., 1993). Despite that amplification in the present model is so nonlinear, the model is also linear according to that superposition test. An analysis of the test in the context of the model shows that such a test has limited usefulness and suggests improvements. Given that the network transfer function has a fixed point at high firing without any Direction Selectivity in Primary Visual Cortex U 9 400 ~ ~ lit .- ~ ~'OO .~ . LGN X a !! 200 .. .E .. iL D 1n0 GO 0 0.1 10 100 1000 Stimulus velocity (e/sec) 100 120 'S c 51100 CiJ Q) '! --... CIl Q) Cii CI c :~ u.. 'S 80 60 40 20 0 0.1 d 51 80 CiJ Q) '! ~ GN-::J 60 Cii ... 40 ell c: FF :~ 20 u.. 1 P Q) 10 Velocity (deg/sec) 100 0 0.1 NP 1 10 100 Velocity (deg/sec) Figure 4: Peak firing rate versus velocity during stimulation by a bar. (a) LGN neuron in cat (Frishman et aJ., 1983). (b) area 17 visual cortical neuron in cat (Orban, 1984). (c) LGN model neuron (curve labelled "LGN") and feedforward model (FF). (d) Cortical amplifier model in the preferred (curve labelled "P") and null directions (NP). 10 Humbert Suarez, Christo! Koch, Rodney Douglas input (see section 2), hysteresis may occur, whereby the network's discharge persists after the initial stimulus is withdrawn. Because of hysteresis, it is imperative to reset the network by presenting a negative, or inhibitory, input. A parallel can be drawn to several recent proposals for the mechanisms underlying short-term memory. 7 CONCLUSIONS From early on, neurophysiologists have proposed that the LG N provides most of the input to visual cortical neurons and shapes their receptive properties (Hubel and Wiesel, 1962). However, direction selectivity and several other stimulus selectivities are not present in LGN neurons, and other important discrepancies have appeared between the receptive field properties of cortical neurons and those of their LG N afferents. Anatomically, synaptic inputs from the LGN account for less than 10 % of synapses to pyramidal neurons in visual cortex; the remaining 90 % could clearly provide a substrate for these receptive field transformations. Although intracortical inhibition has often been invoked to explain various cortical properties, excitation is usually not mentioned, despite that most corti co-cortical synapses are excitatory. In this paper, we show that intracortical excitation can better account for several key properties of cortical neurons than a purely feedforward model, including the magnitude of the conductance change in the null direction, contrast-response curves, and velocity-response curves. Furthermore, surprinsingly, other key cell properties that are appear to point to feedforward models, such as linearity measured by superposition tests, are also properties of a model based on intracortical excitation. Acknowledgements This research was supported by the Office of Naval Research, the National Science Foundation, the National Eye Institute, and the McDonnell Foundation. References Albrecht, D.G., and Geisler, W.S . (1991) Visual Neuroscience 7, 531-546. Albrecht, D.B., and Hamilton, D.B. (1982) J. Neurophysiol. 48,217-237. Douglas, R.J., and Martin, K.A.C. (1991) J. Physiol. 440,735-769. Douglas, R.J ., Martin, K.A.C., and Whitteridge, D. (1988) Nature 332,642-644. Frishman, L.J., Schweitzer-Tong, D.E., and Goldstein, E.B. (1983) J. Neurophysiol. 50, 1393-1414. Hubel, D.H., and Wiesel, T.N. (1962) J. Physiol. 165,559-568. Jagadeesh, B., Wheat, H.S. and Ferster, D. (1993) Science 262, 1901-1904 Orban, G.A. (1984) Neuronal operations in the visual cortex. Springer, Berlin. Peters, A. and Payne, B. R. (1993) Cerebral Cortex 3, 69-78. Victor, J.D. (1987) J.Physiol. 386, 219-246.
898 |@word wiesel:2 open:1 simulation:2 initial:2 exclusively:1 cort:1 current:6 must:1 physiol:3 realistic:1 shape:1 plot:2 n0:1 v_:1 short:2 provides:3 ohl:1 mathematical:1 schweitzer:1 consists:1 indeed:2 morphology:1 little:3 totally:1 provided:1 linearity:3 underlying:1 circuit:4 null:14 compressive:1 transformation:1 temporal:1 uk:1 unit:1 originates:1 christof:1 hamilton:3 appear:1 before:1 persists:1 despite:2 oxford:2 firing:10 approximately:1 suggests:1 co:2 limited:1 range:2 displacement:1 area:3 physiology:4 persistence:1 specificity:1 onto:2 amplify:1 context:1 center:1 go:2 abrupt:1 pure:1 insight:1 array:3 population:3 discharge:2 massive:3 paralleling:1 substrate:1 origin:1 velocity:14 role:1 suarez:5 module:1 frishman:2 wheat:1 connected:1 jagadeesh:2 substantial:2 mentioned:2 moderately:1 dynamic:2 purely:1 gabaa:3 triangle:2 neurophysiol:2 differently:1 cat:3 various:1 fast:2 whose:1 quite:1 larger:2 itself:1 epsp:1 reset:1 causing:1 payne:2 loop:1 amplified:2 amplification:3 description:1 convergence:1 produce:1 depending:1 oo:1 measured:2 excites:1 strong:2 grating:4 come:1 direction:44 suprathreshold:1 koch:5 withdrawn:1 equilibrium:2 circuitry:1 early:2 geniculate:9 superposition:3 clearly:1 rather:2 office:1 naval:1 improvement:1 indicates:1 expansive:1 contrast:15 detect:1 pasadena:2 selective:4 lgn:29 pixel:2 dual:1 orientation:1 among:2 spatial:2 ell:1 field:5 never:1 having:1 lit:1 discrepancy:1 np:2 stimulus:11 quantitatively:1 retina:1 national:2 consisting:1 cns:2 fire:4 maintain:2 amplifier:10 conductance:6 investigate:1 possibility:1 activated:1 filled:2 circle:1 increased:1 gn:1 imperative:2 r25:1 usefulness:1 too:1 connect:1 fundamental:1 peak:4 geisler:2 again:2 reflect:1 recorded:1 albrecht:5 account:2 potential:3 nonlinearities:1 intracortical:6 retinal:1 sec:4 includes:1 hysteresis:4 afferent:3 depends:1 portion:2 thalamic:1 parallel:1 rodney:5 slope:2 vivo:1 il:2 compartment:2 ofthe:1 weak:4 critically:1 overlook:1 mrc:1 explain:2 synapsis:7 reach:2 synaptic:7 against:1 tesl:1 gabab:3 nmda:1 amplitude:1 goldstein:1 feed:1 exceeded:1 response:21 microcircuit:2 strongly:4 furthermore:1 stage:1 nonlinear:1 mode:7 aj:1 functioning:1 former:2 analytically:2 spatially:1 attractive:1 wiring:1 during:5 whereby:1 excitation:11 presenting:1 motion:5 steadystate:1 invoked:1 recently:1 stimulation:5 cerebral:1 neuropharmacology:1 significant:1 surround:1 ai:1 whitteridge:1 christo:4 moving:5 cortex:11 operating:2 inhibition:6 recent:1 selectivity:16 caltech:2 victor:2 cii:2 ii:4 smooth:1 faster:1 match:2 long:1 post:1 shunting:1 biophysics:1 underlies:1 pyramid:2 cell:12 proposal:2 diagram:1 pyramidal:24 induced:1 recording:1 symmetrically:1 presence:1 feedforward:10 enough:2 silent:1 six:1 peter:2 action:2 dramatically:1 generally:1 detailed:2 retino:1 proportionally:2 amount:1 generate:1 canonical:2 inhibitory:13 neuroscience:1 anatomical:2 group:1 key:2 restrain:1 hysteretic:2 threshold:3 drawn:1 capital:1 prevent:1 douglas:11 respond:3 almost:1 followed:1 strength:3 occur:1 constraint:2 simulate:1 orban:4 injection:1 martin:4 relatively:1 according:1 combination:1 mcdonnell:1 membrane:1 psp:2 slightly:1 intimately:1 biologically:1 anatomically:1 mechanism:1 operation:3 corti:2 remaining:1 ensure:1 include:1 classical:1 receptive:6 primary:5 striate:1 fiducial:1 simulated:1 berlin:1 originate:2 modeled:1 index:1 providing:1 balance:1 lg:5 steep:2 cij:2 sigma:1 negative:1 rise:1 suppress:1 neuron:54 displaced:2 saturates:2 somatic:4 connection:11 narrow:1 bar:4 usually:1 appeared:1 program:2 saturation:1 including:3 memory:2 power:1 overlap:1 scheme:1 vetoed:1 eye:1 mediated:3 acknowledgement:1 embedded:1 limitation:1 filtering:2 versus:3 foundation:2 consistent:3 excitatory:11 course:1 supported:1 cortico:1 understand:1 institute:1 feedback:2 curve:14 cortical:38 simplified:1 preferred:8 deg:2 hubel:2 ipsp:1 physiologically:1 continuous:1 why:1 nature:2 transfer:5 mj:1 ca:2 iec:1 intracellular:2 arise:1 body:1 neuronal:1 fig:6 ff:2 slow:1 cil:1 tong:1 position:3 bandpass:1 down:1 symbol:4 physiological:1 incorporating:1 gained:1 ci:1 magnitude:1 interneuron:1 visual:18 failed:1 saturating:1 springer:1 consequently:1 ferster:1 labelled:2 change:5 experimentally:2 neurophysiologists:1 total:3 pas:1 experimental:1 indicating:1 puzzling:1 latter:1 tested:1
7,133
899
SIMPLIFYING NEURAL NETS BY DISCOVERING FLAT MINIMA Sepp Hochreiter" Jiirgen Schmidhuber t Fakultat fiir Informatik, H2 Technische Universitat Miinchen 80290 Miinchen, Germany Abstract We present a new algorithm for finding low complexity networks with high generalization capability. The algorithm searches for large connected regions of so-called ''fiat'' minima of the error function. In the weight-space environment of a "flat" minimum, the error remains approximately constant. Using an MDL-based argument, flat minima can be shown to correspond to low expected overfitting. Although our algorithm requires the computation of second order derivatives, it has backprop's order of complexity. Experiments with feedforward and recurrent nets are described. In an application to stock market prediction, the method outperforms conventional backprop, weight decay, and "optimal brain surgeon" . 1 INTRODUCTION Previous algorithms for finding low complexity networks with high generalization capability are based on significant prior assumptions. They can be broadly classified as follows: (1) Assumptions about the prior weight distribution. Hinton and van Camp [3] and Williams [17] assume that pushing the posterior distribution (after learning) close to the prior leads to "good" generalization. Weight decay can be derived e.g. from Gaussian priors. Nowlan and Hinton [10] assume that networks with many similar weights generated by Gaussian mixtures are "better" a priori. MacKay's priors [6] are implicit in additional penalty terms, which embody the "hochreit@informatik. tu-muenchen .de t schmidhu@informatik.tu-muenchen.de 530 Sepp Hochreiter. Jurgen Schmidhuber assumptions made. (2) Prior assumptions about how theoretical results on early stopping and network complexity carryover to practical applications. Examples are methods based on validation sets (see [8]), Vapnik's "structural risk minimization" [1] [14], and the methods of Holden [5] and Wang et al. [15]. Our approach requires less prior assumptions than most other approaches (see appendix A.l). Basic idea of flat minima search. Our algorithm finds a large region in weight space with the property that each weight vector from that region has similar small error. Such regions are called "flat minima". To get an intuitive feeling for why ''flat'' minima are interesting, consider this (see also Wolpert [18]): a "sharp" minimum corresponds to weights which have to be specified with high precision. A ''flat'' minimum corresponds to weights many of which can be given with low precision. In the terminology of the theory of minimum description length (MDL), fewer bits of information are required to pick a ''flat'' minimum (corresponding to a "simple" or low complexity-network). The MDL principle suggests that low network complexity corresponds to high generalization performance (see e.g. [4, 13]). Unlike Hinton and van Camp's method [3] (see appendix A.3), our approach does not depend on explicitly choosing a "good" prior. Our algorithm finds "flat" minima by searching for weights that minimize both training error and weight precision. This requires the computation of the Hessian. However, by using Pearlmutter's and M~ner's efficient second order method [11, 7], we obtain the same order of complexity as with conventional backprop. A utomatically, the method effectively reduces numbers of units, weigths, and input lines, as well as the sensitivity of outputs with respect to remaining weights and units. Excellent experimental generalization results will be reported in section 4. 2 TASK / ARCHITECTURE / BOXES Generalization task. The task is to approximate an unknown relation [) c X x Z between a set of inputs X C RN and a set of outputs Z C RK. [) is taken to be a function. A relation D is obtained from [) by adding noise to the outputs. All training information is given by a finite relation Do C D. Do is called the training set. The pth element of Do is denoted by an input/target pair (xp, dp ). Architecture. For simplicity, we will focus on a standard feedforward net (but in the experiments, we will use recurrent nets as well). The net has N input units, K output units, W weights, and differentiable activation functions. It maps input vectors xp E RN to output vectors op E RK. The weight from unit j to i is denoted by Wij. The W -dimensional weight vector is denoted by w. Training error. Mean squared error Eq(w, Do) := l.riol E(xp,dp)ED o II dp - op 112 is used, where II . II denotes the Euclidian norm, and 1.1 denotes the cardinality of a set. To define regions in weight space with the property that each weight vector from that region has "similar small error", we introduce the tolerable error Etal, a positive constant. "Small" error is defined as being smaller than Etal. Eq(w, Do) > E tal implies "underfitting" . Boxes. Each weight W satisfying Eq(w, Do) ~ E tal defines an "acceptable minimum". We are interested in large regions of connected acceptable minima. 531 Simplifying Neural Nets by Discovering Flat Minima Such regions are called fiat minima. They are associated with low expected generalization error (see [4]). To simplify the algorithm for finding large connected regions (see below), we do not consider maximal connected regions but focus on so-called "boxes" within regions: for each acceptable minimum w, its box Mw in weight space is a W-dimensional hypercuboid with center w . For simplicity, each edge of the box is taken to be parallel to one weight axis. Half the length of the box edge in direction of the axis corresponding to weight Wij is denoted by ~wii ' which is the maximal (positive) value such that for all i, j, all positive K.ii :5 ~Wij can be added to or subtracted from the corresponding component of W simultaneously without violating Eq(. , Do) :5 Etol (~Wij gives the precision of Wij). Mw's box volume is defined by ~w := 2w ni,j ~Wij. 3 THE ALGORITHM The algorithm is designed to find a W defining a box Mw with maximal box volume ~w. This is equivalent to finding a box Mw with minimal B( w, Do) := -log(~w/2W) = Li,j -log ~Wi.j. Note the relationship to MDL (B is the number of bits required to describe the weights). In appendix A.2, we derive the following algorithm. It minimizes E(w, Do) = Eq(w, Do) + >'B(w, Do), where B= ~ (-WIOg{+ ~logL(:~~j)2 + WlogL ' ,3 " " (L: /~ ,)2) . (1) L"(8w;) ' ,3 0" Here is the activation of the kth output unit, { is a constant, and >. is a positive variable ensuring either Eq(w, Do) :5 Etol, or ensuring an expected decrease of E q (., Do) during learning (see [16] for adjusting >.). E(w, Do) is minimized by gradient descent. To minimize B(w, Do), we compute 8B(w, Do) 8 wuv 2 8B(w, Do) 8 0" = '"" L8 " 8Wij 8 wuv 8(~) I. .? "',' ,) ~ II lor a u, v . (2) vw., It can be shown (see [4]) that by using Pearlmutter's and M~ller's efficient second order method [11, 7], the gradient of B( w, Do) can be computed in O(W) time (see details in [4]) . Therefore, our algorithm has the same order of complexity as standard backprop. 4 EXPERIMENTAL RESULTS (see [4] for details) EXPERIMENT 1 - noisy classification. The first experiment is taken from Pearlmutter and Rosenfeld [12]. The task is to decide whether the x-coordinate of a point in 2-dimensional space exceeds zero (class 1) or does not (class 2). Noisy training examples are generated as follows: data points are obtained from a Gaussian with zero mean and stdev 1.0, bounded in the interval [-3.0,3.0]. The data points are misclassified with a probability of 0.05. Final input data is obtained by adding a zero mean Gaussian with stdev 0.15 to the data points. In a test with 2,000,000 data points, it was found that the procedure above leads to 9.27 per cent 532 Sepp Hochreiter, Jurgen Schmidhuber 1 2 3 4 5 Backprop MSE dto 0.220 1.35 0.223 1.16 0.222 1.37 0.213 1.18 0.222 1.24 New approach dto MSE 0.193 0.00 0.189 0.09 0.13 0.186 0.181 0.01 0.195 0.25 6 7 8 9 10 Backprop MSE dto 0.219 1.24 0.215 1.14 0.214 1.10 0.218 1.21 0.214 1.21 New approach dto MSE 0.187 0.04 0.187 0.07 0.01 0.185 0.190 0.09 0.188 0.07 Table 1: 10 comparisons of conventional backprop (BP) and our new method (FMS) . The second row (labeled "MSE") shows mean squared error on the test set. The third row ("dto") shows the difference between the fraction (in per cent) of misclassifications and the optimal fraction (9.27). The remaining rows provide the analoguous information for the new approach, which clearly outperforms backprop. misclassified data. No method will misclassify less than 9.27 per cent, due to the inherent noise in the data. The training set is based on 200 fixed data points. The test set is based on 120,000 data points. Results. 10 conventional backprop (BP) nets were tested against 10 equally initialized networks based on our new method ("flat minima search", FMS). After 1,000 epochs, the weights of our nets essentially stopped changing (automatic "early stopping"), while backprop kept changing weights to learn the outliers in the data set and overfit. In the end, our approach left a single hidden unit h with a maximal weight of 30.0 or -30.0 from the x-axis input. Unlike with backprop, the other hidden units were effectively pruned away (outputs near zero). So was the y-axis input (zero weight to h) . It can be shown that this corresponds to an "optimal" net with minimal numbers of units and weights. Table 1 illustrates the superior performance of our approach. EXPERIMENT 2 - recurrent nets. The method works for continually running fully recurrent nets as well. At every time step, a recurrent net with sigmoid activations in [0,1] sees an input vector from a stream of randomly chosen input vectors from the set {(0,0), (0,1),(1,0),(1,1)}. The task is to switch on the first output unit whenever an input (1,0) had occurred two time steps ago, and to switch on the second output unit without delay in response to any input (0,1). The task can be solved by a single hidden unit. Results. With conventional recurrent net algorithms, after training, both hidden units were used to store the input vector . Not so with our new approach. We trained 20 networks. All of them learned perfect solutions. Like with weight decay, most weights to the output decayed to zero. But unlike with weight decay, strong inhibitory connections (-30.0) switched off one of the hidden units, effectively pruning it away. EXPERIMENT 3 - stock market prediction. We predict the DAX (German stock market index) based on fundamental (experiments 3.1 and 3.2) and technical (experiment 3.3) indicators. We use strictly layered feedforward nets with sigmoid units active in [-1,1]' and the following performance measures: Confidence: output 0 > a - positive tendency, 0 < -a - negative tendency. Performance: Sum of confidently, incorrectly predicted DAX changes is subtracted Simplifying Neural Nets by Discovering Flat Minima 533 from sum of confidently, correctly predicted ones. The result is divided by the sum of absolute changes. EXPERIMENT 3.1: Fundamental inputs: (a) German interest rate ("Umlaufsrendite"), (b) industrial production divided by money supply, (c) business sentiments ("IFO Geschiiftsklimaindex"). 24 training examples, 68 test examples, quarterly prediction, confidence: 0: = 0.0/0.6/0.9, architecture: (3-8-1) . EXPERIMENT 3.2: Fundamental inputs: (a), (b) , (c) as in expo 3.1, (d) dividend rate, (e) foreign orders in manufacturing industry. 228 training examples, 100 test examples, monthly prediction, confidence: 0: = 0.0/0.6/0.8, architecture: (5-8-1). EXPERIMENT 3.3: Technical inputs: (a) 8 most recent DAX-changes, (b) DAX, (c) change of 24-week relative strength index ("RSI"), (d) difference of "5 week statistic", (e) "MACD" (difference of exponentially weighted 6 week and 24 week DAX). 320 training examples, 100 test examples, weekly predictions, confidence: 0: 0.0/0.2/0.4, architecture: (12-9-1). The following methods are tested: (1) Conventional backprop (BP), (2) optimal brain surgeon (OBS [2]), (3) weight decay (WD [16]), (4) flat minima search (FMS). = Results. Our method clearly outperforms the other methods . FMS is up to 63 per cent better than the best competitor (see [4] for details) . APPENDIX - THEORETICAL JUSTIFICATION A.t. OVERFITTING ERROR In analogy to [15] and [1], we decompose the generalization error into an "overfitting" error and an "underfitting" error. There is no significant underfitting error (corresponding to Vapnik's empirical risk) if Eq(w, Do) ~ E tol . Some thought is required, however, to define the "overfitting" error. We do this in a novel way. Since we do not know the relation D, we cannot know p(o: I D), the "optimal" posterior weight distribution we would obtain by training the net on D (- "sure thing hypothesis") . But, for theoretical purposes, suppose we did know p(o: I D) . Then we could use p(o: I D) to initialize weights before learning the training set Do . Using the Kullback-Leibler distance, we measure the information (due to noise) conveyed by Do, but not by D. In conjunction with the initialization above, this provides the conceptual setting for defining an overfitting error measure. But, the initialization does not really matter, because it does not heavily influence the posterior (see [4]). The overfittin~ error is the Kullback-Leibler distance of the posteriors: Eo(D , Do) = J p(o: I Do) log (p(o: I Do)/p(o: I D?) do:. Eo(D, Do) is the expectation of log (p(o: I Do)/p(o: I D)) (the expected difference of the minimal description of 0: with respect to D and Do, after learning Do). Now we measure the expected overfitting error relative to Mw (see section 2) by computing the expectation of log (p( 0: I Do) / p( 0: I D? in the range Mw: Ero(w) = f3 (1M", PM", (0: I Do)Eq(O:, D)do: - Eq(Do, M w ?) . (3) Here PM",(O: I Do) := p(o: I Do)/ IM", p(a: I Do)da: is the posterior of Do scaled to obtain a distribution within Mw, and Eq(Do, Mw) := IM", PM", (a I Do)Eq(a, Do)do: is the mean error in Mw with respect to Do. 534 Sepp Hochreiter. JiJrgen Schmidhuber Clearly, we would like to pick W such that Ero( w) is minimized. Towards this purpose, we need two additional prior assumptions, which are actually implicit in most previous approaches (which make additional stronger assumptions, see section 1): (1) "Closeness assumption": Every minimum of E q (., Do) is "close" to a maximum of p(aID) (see formal definition in [4]). Intuitively, "closeness" ensures that Do can indeed tell us something about D, such that training on Do may indeed reduce the error on D. (2) "Flatness assumption": The peaks of p(aID)'s maxima are not sharp. This MDL-like assumption holds if not all weights have to be known exactly to model D. It ensures that there are regions with low error on D. A.2. HOW TO FLATTEN THE NETWORK OUTPUT To find nets with flat outputs, two conditions will be defined to specify B(w, Do) (see section 3). The first condition ensures flatness. The second condition enforces "equal flatness" in all weight space ? directions. In both cases, linear approximations will be made (to be justified in [4]). We are looking for weights (causing tolerable error) that can be perturbed without causing significant output changes. Perturbing the weights w by 6w (with components 6Wij), we obtain ED(w,6w) := L,,(o"(w + 6w) - o"(w?)2, where o"(w) expresses o"'s dependence on w (in what follows, however, w often will be suppressed for convenience). Linear approximation (justified in [4]) gives us "Flatness Condition 1": ED(w, 6w) ~ 00" 00" L(L -;;;-:-:6Wij)2 :$ L(L 1-;;;-:-:1I 6wij1)2 :$ ( , .. .. uW'J .. .. uW'J .. '.J .. (4) '.J where ( > 0 defines tolerable output changes within a box and is small enough to allow for linear approximation (it does not appear in B(w, Do)'s gradient, see section 3). Many Mw satisfy flatness condition 1. To select a particular, very flat M w, the following "Flatness Condition 2" uses up degrees of freedom left by (4): . . ( 2 " ( 00")2 'VZ,),U,V: 6Wij) L-~ " 00" 2 = (6w uv ) 2L-(-~-)' (5) " uW'J " uWuv Flatness Condition 2 enforces equal "directed errors" EDij(W,6wij) L,,(O"(Wij + 6Wij) - O"(Wij)? ~ L,,(:;:;6wij?, where Ok(Wij) has the obvious meaning. It can be shown (see [4]) that with given box volume, we need flatness condition 2 to minimize the expected description length of the box center. Flatness condition 2 influences the algorithm as follows: (1) The algorithm prefers to increase the 6Wij'S of weights which currently are not important to generate the target output. (2) The algorithm enforces equal sensitivity of all output units with respect to the weights. Hence, the algorithm tends to group hidden units according to their relevance for groups of output units. Flatness condition 2 is essential: flatness condition 1 by itself corresponds to nothing more but first order derivative reduction (ordinary sensitivity reduction, e.g. [9]). Linear approximation is justified by the choice of f in equation (4). = We first solve equation (5) for 16w,; I = 16W., I ( 2:, (a':::.) 2/ 2:, U.::J 2) Simplifying Neural Nets by Discovering Flat Minima 535 (fixing u, v for all i, j) . Then we insert 16wij I into equation (4) (replacing the second "$" in (4) by ";;;:"). This gives us an equation for the 16wijl (which depend on w, but this is notationally suppressed): 16wi?1 J = Vii ook ( "(_)2 LowIJ.. k (6) The 16wijl approximate the ~Wij from section 2. Thus, B(w,Do) (see section 3) can be approximated by B(w, Do) :;;;: Ei,j -log 16wijl. This immediately leads to the algorithm given by equation (1) . How can this approximation be justified? The learning process itself enforces its validity (see justification in [4]). Initially, the conditions above are valid only in a very small environment of an "initial" acceptable minimum. But during search for new acceptable minima with more associated box volume, the corresponding environments are enlarged, which implies that the absolute values of the entries in the Hessian decrease . It can be shown (see [4]) that the algorithm tends to suppress the following values: (1) unit activations, (2) first order activation derivatives, (3) the sum of all contributions of an arbitary unit activation to the net output. Since weights, inputs, activation functions, and their first and second order derivatives are bounded, it can be shown (see [4]) that the entries in the Hessian decrease where the corresponding 16wij I increase. A.3. RELATION TO HINTON AND VAN CAMP Hinton and van Camp [3] minimize the sum of two terms: the first is conventional enor plus variance, the other is the distance f p( a I Do) log (p( a I Do) I p( a? da between posterior pea I Do) and prior pea). The problem is to choose a "good" prior. In contrast to their approach, our approach does not require a "good" prior given in advance. Furthermore, Hinton and van Camp have to compute variances of weights and units, which (in general) cannot be done using linear approximation. Intuitively speaking, their weight variances are related to our ~Wij. Our approach, however, does justify linear approximation. References [1] I. Guyon, V. Vapnik, B. Boser, L. Bottou, and S. A. Solla. Structural risk minimization for character recognition. In J. E. Moody, S. J. Hanson, and R. P. Lippman, editors, Advances in Neural Information Processing Systems 4, pages 471-479. San Mateo, CA: Morgan Kaufmann, 1992. [2] B. Hassibi and D. G. Stork. Second order derivatives for network pruning: Optimal bra.i.n surgeon. In J. D. Cowan S. J . Hanson and C. L. Giles, editors, Advances in Neural Information Processing Systems 5, pages 164-171 . San Mateo, CA: Morgan Kaufmann , 1993. [3] G. E. Hinton and D. van Camp. Keeping neural networks simple. In Proceedings of the International Conference on Artificial Neural Networks, Amsterdam, pa.ges 11-18. Springer, 1993. 536 Sepp Hoc/zreiter, Jiirgen Schmidhuber [4] S. Hochreiter and J. Schmidhuber. Flat Inllllma search for discovering simple nets. Technical Report FKI-200-94, Fakultiit fiir Informatik, Technische Universitiit Munchen, 1994. [5] S. B. Holden. On the Theory of Generalization and Self-Structuring in Linearly Weighted Connectionist Networks. PhD thesis, Cambridge University, Engineering Department, 1994. [6] D. J. C . MacKay. A practical Bayesian framework for backprop networks. Neural Computation, 4:448-472, 1992. [7] M. F. M~ller. Exact calculation of the product of the Hessian matrix offeed-forward network error functions and a vector in O(N) time. Technical Report PB-432, Computer Science Department, Aarhus University, Denmark, 1993. [8] J. E. Moody and J . Utans. Architecture selection strategies for neural networks: Application to corporate bond rating prediction. In A. N. Refenes, editor, Neural Networks in the Capital Markets. John Wiley & Sons, 1994. [9] A. F. Murray and P. J. Edwards. Synaptic weight noise during MLP learning enhances fault-tolerance, generalisation and learning trajectory. In J. D. Cowan S. J. Hanson and C. L. Giles, editors, Advances in Neural Information Processing Systems 5, pages 491-498. San Mateo, CA: Morgan Kaufmann, 1993. [10] S. J. Nowlan and G. E. Hinton. Simplifying neural networks by soft weight sharing. Neural Computation, 4:173-193, 1992. [11] B. A. Pearlmutter. Fast exact multiplication by the Hessian. Neural Computation, 1994. [12] B. A. Pearlmutter and R. Rosenfeld. Chaitin-Kolmogorov complexity and generalization in neural networks. In R. P. Lippmann, J. E. Moody, and D. S. Touretzky, editors, Advances in Neural Information Processing Systems 3, pages 925-931. San Mateo, CA: Morgan Kaufmann, 1991. [13] J. H. Schmid huber. Discovering problem solutions with low Kolmogorov complexity and high generalization capability. Technical Report FKI-194-94, Fakultiit fUr Informatik, Technische U niversitiit Munchen, 1994. [14] V. Vapnik. Principles of risk minimization for learning theory. In J. E . Moody, S. J. Hanson, and R. P. Lippman, editors, Advances in Neural Information Processing Systems 4, pages 831-838. San Mateo, CA: Morgan Kaufmann, 1992. [15] C . Wang, S. S. Venkatesh, and J. S. Judd. Optimal stopping and effective machine complexity in learning. In J . D. Cowan, G. Tesauro, and J. Alspector, editors, Advances in Neural Information Processing Systems 6, pages 303-310. Morgan Kaufmann, San Mateo, CA, 1994. [16] A. S. Weigend, D. E. Rumelhart, and B. A. Huberman. Generalization by weightelimination with application to forecasting. In R. P. Lippmann, J. E. Moody, and D. S. Touretzky, editors, Advances in Neural Information Processing Systems 3, pages 875-882. San Mateo, CA: Morgan Kaufmann, 1991. [17] P. M. Williams. Bayesian regularisation and pruning using a Laplace prior. Technical report, School of Cognitive and Computing Sciences, University of Sussex, Falmer, Brighton, 1994. [18] D. H. Wolpert. Bayesian backpropagation over i-o functions rather than weights. In J. D. Cowan, G. Tesauro, and J. Alspector, editors, Advances in Neural Information Processing Systems 6, pages 200-207. San Mateo, CA: Morgan Kaufmann, 1994.
899 |@word norm:1 stronger:1 simplifying:5 pick:2 euclidian:1 analoguous:1 reduction:2 initial:1 outperforms:3 wd:1 nowlan:2 activation:7 john:1 hochreit:1 designed:1 half:1 discovering:6 fewer:1 provides:1 miinchen:2 lor:1 supply:1 underfitting:3 introduce:1 huber:1 expected:6 indeed:2 alspector:2 market:4 embody:1 brain:2 cardinality:1 bounded:2 dax:5 what:1 minimizes:1 enor:1 finding:4 every:2 weekly:1 exactly:1 refenes:1 scaled:1 unit:21 appear:1 continually:1 positive:5 ner:1 before:1 engineering:1 wuv:2 tends:2 approximately:1 plus:1 initialization:2 mateo:8 suggests:1 range:1 directed:1 practical:2 enforces:4 backpropagation:1 lippman:2 procedure:1 empirical:1 thought:1 confidence:4 flatten:1 get:1 cannot:2 close:2 layered:1 convenience:1 selection:1 risk:4 influence:2 equivalent:1 conventional:7 map:1 center:2 sepp:5 williams:2 simplicity:2 immediately:1 lowij:1 searching:1 coordinate:1 justification:2 laplace:1 target:2 suppose:1 heavily:1 exact:2 us:1 hypothesis:1 pa:1 element:1 rumelhart:1 satisfying:1 approximated:1 recognition:1 labeled:1 wang:2 solved:1 region:12 ensures:3 connected:4 solla:1 decrease:3 environment:3 complexity:11 trained:1 depend:2 surgeon:3 stock:3 kolmogorov:2 stdev:2 fast:1 describe:1 effective:1 artificial:1 tell:1 choosing:1 solve:1 statistic:1 rosenfeld:2 noisy:2 itself:2 final:1 hoc:1 differentiable:1 net:20 ero:2 maximal:4 product:1 tu:2 causing:2 chaitin:1 intuitive:1 description:3 perfect:1 derive:1 recurrent:6 fixing:1 school:1 op:2 jurgen:2 eq:11 edward:1 strong:1 predicted:2 implies:2 direction:2 pea:2 backprop:13 require:1 generalization:12 really:1 decompose:1 im:2 strictly:1 insert:1 hold:1 predict:1 week:4 early:2 purpose:2 bond:1 currently:1 vz:1 weighted:2 minimization:3 clearly:3 gaussian:4 rather:1 conjunction:1 structuring:1 derived:1 focus:2 fur:1 industrial:1 contrast:1 camp:6 stopping:3 foreign:1 holden:2 initially:1 hidden:6 relation:5 wij:21 misclassified:2 interested:1 germany:1 classification:1 denoted:4 priori:1 mackay:2 initialize:1 equal:3 f3:1 minimized:2 carryover:1 report:4 simplify:1 inherent:1 connectionist:1 randomly:1 simultaneously:1 freedom:1 schmidhu:1 misclassify:1 interest:1 mlp:1 mdl:5 mixture:1 edge:2 dividend:1 initialized:1 theoretical:3 minimal:3 stopped:1 industry:1 soft:1 giles:2 ordinary:1 technische:3 entry:2 delay:1 universitat:1 reported:1 perturbed:1 decayed:1 fundamental:3 sensitivity:3 peak:1 international:1 off:1 moody:5 squared:2 thesis:1 choose:1 cognitive:1 derivative:5 li:1 de:2 ifo:1 matter:1 satisfy:1 explicitly:1 stream:1 capability:3 parallel:1 contribution:1 minimize:4 ni:1 variance:3 kaufmann:8 correspond:1 bayesian:3 fki:2 informatik:5 trajectory:1 ago:1 classified:1 touretzky:2 whenever:1 ed:3 synaptic:1 definition:1 sharing:1 against:1 competitor:1 jiirgen:2 obvious:1 associated:2 adjusting:1 fiat:2 actually:1 ok:1 violating:1 response:1 specify:1 done:1 box:14 furthermore:1 implicit:2 overfit:1 logl:1 replacing:1 ei:1 defines:2 validity:1 hence:1 leibler:2 during:3 dto:5 self:1 sussex:1 brighton:1 pearlmutter:5 meaning:1 novel:1 superior:1 sigmoid:2 perturbing:1 stork:1 exponentially:1 volume:4 occurred:1 significant:3 monthly:1 cambridge:1 automatic:1 uv:1 pm:3 etal:2 had:1 money:1 something:1 posterior:6 recent:1 tesauro:2 schmidhuber:6 store:1 fault:1 morgan:8 minimum:24 additional:3 eo:2 bra:1 ller:2 ii:5 flatness:11 corporate:1 reduces:1 exceeds:1 technical:6 calculation:1 divided:2 equally:1 ensuring:2 prediction:6 basic:1 muenchen:2 essentially:1 expectation:2 hochreiter:5 justified:4 interval:1 unlike:3 sure:1 thing:1 cowan:4 structural:2 mw:10 vw:1 near:1 feedforward:3 enough:1 switch:2 misclassifications:1 architecture:6 fm:4 reduce:1 idea:1 whether:1 forecasting:1 penalty:1 sentiment:1 rsi:1 hessian:5 speaking:1 prefers:1 tol:1 weigths:1 generate:1 inhibitory:1 utans:1 per:4 correctly:1 broadly:1 express:1 group:2 terminology:1 pb:1 capital:1 changing:2 weightelimination:1 falmer:1 kept:1 uw:3 fraction:2 sum:5 weigend:1 guyon:1 decide:1 ob:1 appendix:4 acceptable:5 bit:2 strength:1 bp:3 expo:1 flat:17 tal:2 argument:1 pruned:1 notationally:1 department:2 according:1 smaller:1 son:1 suppressed:2 character:1 wi:2 outlier:1 intuitively:2 taken:3 equation:5 remains:1 german:2 know:3 ge:1 end:1 wii:1 quarterly:1 munchen:2 away:2 tolerable:3 subtracted:2 cent:4 denotes:2 remaining:2 running:1 pushing:1 murray:1 added:1 strategy:1 dependence:1 enhances:1 gradient:3 dp:3 kth:1 distance:3 denmark:1 length:3 index:2 relationship:1 negative:1 suppress:1 unknown:1 finite:1 descent:1 incorrectly:1 defining:2 hinton:8 looking:1 rn:2 sharp:2 rating:1 venkatesh:1 pair:1 required:3 specified:1 connection:1 hanson:4 fakultat:1 learned:1 boser:1 below:1 confidently:2 business:1 indicator:1 axis:4 schmid:1 prior:13 epoch:1 aarhus:1 multiplication:1 relative:2 regularisation:1 fully:1 interesting:1 analogy:1 validation:1 h2:1 switched:1 degree:1 conveyed:1 xp:3 principle:2 editor:9 production:1 row:3 l8:1 keeping:1 formal:1 allow:1 absolute:2 van:6 tolerance:1 judd:1 valid:1 forward:1 made:2 san:8 feeling:1 pth:1 approximate:2 pruning:3 wijl:3 lippmann:2 kullback:2 fiir:2 overfitting:6 active:1 conceptual:1 search:6 why:1 table:2 learn:1 ca:8 mse:5 excellent:1 bottou:1 da:2 did:1 linearly:1 noise:4 nothing:1 enlarged:1 aid:2 wiley:1 precision:4 hassibi:1 third:1 rk:2 decay:5 closeness:2 essential:1 vapnik:4 adding:2 effectively:3 phd:1 illustrates:1 vii:1 wolpert:2 amsterdam:1 springer:1 corresponds:5 manufacturing:1 towards:1 universitiit:1 change:6 generalisation:1 huberman:1 justify:1 wij1:1 called:5 experimental:2 tendency:2 arbitary:1 select:1 fakultiit:2 relevance:1 tested:2
7,134
9
22 LEARNING ON A GENERAL NETWORK Amir F. Atiya Department of Electrical Engineering California Institute of Technology Ca 91125 Abstract This paper generalizes the backpropagation method to a general network containing feedback t;onnections. The network model considered consists of interconnected groups of neurons, where each group could be fully interconnected (it could have feedback connections, with possibly asymmetric weights), but no loops between the groups are allowed. A stochastic descent algorithm is applied, under a certain inequality constraint on each intra-group weight matrix which ensures for the network to possess a unique equilibrium state for every input. Introduction It has been shown in the last few years that large networks of interconnected "neuron" -like elemp.nts are quite suitable for performing a variety of computational and pattern recognition tasks. One of the well-known neural network models is the backpropagation model [1]-[4]. It is an elegant way for teaching a layered feedforward network by a set of given input/output examples. Neural network models having feedback connections, on the other hand, have also been devised (for example the Hopfield network [5]), and are shown to be quite successful in performing some computational tasks. It is important, though, to have a method for learning by examples for a feedback network, since this is a general way of design, and thus one can avoid using an ad hoc design method for each different computational task. The existence of feedback is expected to improve the computational abilities of a given network. This is because in feedback networks the state iterates until a stable state is reached. Thus processing is perforrr:.ed on several steps or recursions. This, in general allows more processing abilities than the "single step" feedforward case (note also the fact that a feedforward network is a special case of a feedback network). Therefore, in this work we consider the problem of developing a general learning algorithm for feedback networks. In developing a learning algorithm for feedback networks, one has to pay attention to the following (see Fig. 1 for an example of a configuration of a feedback network). The state of the network evolves in time until it goes to equilibrium, or possibly other types of behavior such as periodic or chaotic motion could occur. However, we are interested in having a steady and and fixed output for every input applied to the network. Therefore, we have the following two important requirements for the network. Beginning in any initial condition, the state should ultimately go to equilibrium. The other requirement is that we have to have a unique ? American Institute of Physics 1988 23 equilibrium state. It is in fact that equilibrium state that determines the final output. The objective of the learning algorithm is to adjust the parameters (weights) of the network in small steps, so as to move the unique equilibrium state in a way that will result finally in an output as close as possible to the required one (for each given input). The existence of more than op.e equilibrium state for a given input causes the following problems. In some iterations one might be updating the weights so as to move one of the equilibrium states in a sought direction, while in other iterations (especially with different input examples) a different equilibrium state is moved. Another important point is that when implementing the network (after the completion oflearning), for a fixed input there can be more than one possible output. Independently, other work appeared recently on training a feedback network [6],[7],[8]. Learning algorithms were developed, but solving the problem of ensuring a unique equilibrium was not considered. This problem is addressed in this paper and an appropriate network and a learning algorithm are proposed. neuron 1 outputs inputs Fig . 1 A recurrent network The Feedback Network Consider a group of n neurons which could be fully inter-connected (see Fig. 1 for an example). The weight matrix W can be asymmetric (as opposed to the Hopfield network). The inputs are also weighted before entering into the network (let V be the weight matrix). Let x and y be the input and output vectors respectively. In our model y is governed by the following set of differential equations, proposed by Hopfield [5]: du Tdj = Wf(u) - u + Vx, y = f(u) (1) 24 where f(u) = (J(ud, ... , f(un)f, T denotes the transpose operator, f is a bounded and differentiable function, and.,. is a positive constant. For a given input, we would like the network after a short transient period to give a steady and fixed output, no matter what the initial network state was. This means that beginning any initial condition, the state is to be attracted towards a unique equilibrium. This leads to looking for a condition on the matrix W. Theorem: A network (not necessarily symmetric) satisfying L L w'fi i < l/max(J')2, i exhibits no other behavior except going to a unique equilibrium for a given input. Proof : Let udt) and U2(t) be two solutions of (1). Let where " II is the two-norm. Differentiating J with respect to time, one obtains Using (1) , the expression becomes dJ(t) = --lluI(t) 2 -d- u2(t))11 2 t 1" 2 + -(uI(t) - U2(t)) T W [f( uI(t) ) - f ( uz(t) )] . .,. Using Schwarz's Inequality, we obtain Again, by Schwarz's Inequality, i = 1, ... ,n where Wi denotes the ith row of W. Using the mean value theorem, we get Ilf(udt)) - f(U2(t))II ~ (maxl!'I)IIUl(t) - uz(t)ll. Using (2),(3), and the expression for J(t), we get d~~t) ~ where -aJ(t) (4) (3) (2) 25 By hypothesis of the Theorem, a is strictly positive. Multiplying both sides of (4) by exp( at), the inequality results, from which we obtain J(t) ~ J(O)e- at . From that and from the fact that J is non-negative, it follows that J(t) goes to zero as t -+ <Xl. Therefore, any two solutions corresponding to any two initial conditions ultimately approach each other. To show that this asymptotic solution is in fact an equilibrium, one simply takes U2(t) = Ul(t + T), where T is a constant, and applies the above argument (that J(t) -+ 0 as t -+ <Xl), and hence Ul(t + T) -+ udt) as t -+ <Xl for any T, and this completes the proof. For example, if the function I is of the following widely used sigmoid-shaped form, 1 I(u) = l+e- u ' then the sum of the square of the weights should be less than 16. Note that for any function I, scaling does not have an effect on the overall results. We have to work in our updating scheme subject to the constraint given in the Theorem. In many cases where a large network is necessary, this constraint might be too restrictive. Therefore we propose a general network, which is explained in the next Section. The General Network We propose the following network (for an example refer to Fig. 2). The neurons are partitioned into several groups. Within each group there are no restrictions on the connections and therefore the group could be fully interconnected (i.e. it could have feedback connections) . The groups are connected to each other, but in a way that there are no loops. The inputs to the whole network can be connected to the inputs of any of the groups (each input can have several connections to several groups). The outputs of the whole network are taken to be the outputs (or part of the outputs) of a certain group, say group I. The constraint given in the Theorem is applied on each intra-group weight matrix separately. Let (qa, s"), a = 1, .. . , N be the input/output vector pairs of the function to be implemented. We would like to minimize the sum of the square error, given by a=l where M e" = I)y{ - si}2, i=l and yf is the output vector of group f upon giving input qa, and M is the dimension of vector s". The learning process is performed by feeding the input examples qU sequentially to the network, each time updating the weights in an attempt to minimize the error. 26 inputs J---V outputs Fig. 2 An example of a general network (each group represents a recurrent network) Now, consider a single group l. Let Wi be the intra-group weight matrix of group l, vrl be the matrix of weights between the outputs of group,. and the inputs of group l, and yl be the output vector of group I. Let the respective elements be w~i' V[~., and y~. Furthermore, let be the number of neurons of group l. Assume that the time constant l' is sufficiently small so as to allow the network to settle quickly to the equilibrium state, which is given by the solution of the equation n, yl = f(W'yl + L vrlyr) . (5) r?A I where A, is the set of the indices of the groups whose outputs a.re connected to the inputs of group ,. We would like each iteration to update the weight matrices Wi and vrl so as to move the equilibrium in a direction to decrease the error. We need therefore to know the change in the error produced by a small change in the weight matrices. Let .:;';, , and aa~~, denote the matrices whose (i, j)th element are :~'.' and ::~ respectively. Let ~ be the column vector '1 '1 :r whose ith element is ~. We obtain the following relations: uy. 8e a = 8W' 8e a 8V tl [A' _ (W')T] -1 8ea ( 8yl Y ')T , a ( r)T = [A' _ (W')T] -1 8e 8yl y , where A' is the diagonal matrix whose ith diagonal element is l/f'(Lk w!kY~ + LrLktJ[kyk) for a derivation refer to Appendix). The vector ~ associated with groUp l can be obtained in terms of the vectors ~, fEB" where B, is the set of the indices of the groups whose inputs are connected to the outputs of group ,. We get (refer to Appendix) 8e a 8yl = '" (V'i)T[Ai _ (Wi{r 1 8e".. ~ 8y3 JlBI (6) The matrix A' ~ (W')T for any group l can never be singular, so we will not face any problem in the updating process. To prove that, let z be a vector satisfying [A' - (W'f]z = o. 27 We can write zdmaxlf' I ~ LW~.Zk' i = I, ... , nl k where Zi is the ,"th element of z. Using Schwarz's Inequality, we obtain i = I, ... ,nl Squaring both sides and adding the inequalities for i = I, ... , nl, we get L/; ~ max(J')2(Lz~) LL(w~i)2. i k Since the condition (7) k LL(W!k)2 < I/max(J')2), k is enforced, it follows that (7) cannot be satisfied unless z is the zero vector. Thus, the matrix A' - (W')T cannot be singular. For each iteration we begin by updating the weights of group f (the group contammg the final outputs). For that group ~ equals simply 2(y{ - SI, ... , yf.t - SM, 0, ... , O)T). Then we move backwards to the groups connected to that group and obtain their corresponding vectors using (6), update the weights, and proceed in the same manner until we complete updating all the groups. Updating the weights is performed using the following stochastic descent algorithm for each group, !!J: 8e a t:. V = -a3 8V + a4 ea R , where R is a noise matrix whose elements are characterized by independent zero-mean unityvariance Gaussian densities, and the a's are parameters. The purpose of adding noise is to allow escaping local minima if one gets stuck in any of them. Note that the control parameter is taken to be ea. Hence the variance of the added noise tends to decrease the more we approach the ideal zero-error solution. This makes sense because for a large error, i.e. for an unsatisfactory solution, it pays more to add noise to the weight matrices in order to escape local minima. On the other hand, if the error is small, then we are possibly near the global minimum or to an acceptable solution, and hence we do not want too much noise in order not to be thrown out of that basin. Note that once we reach the ideal zero-error solution the added noise as well as the gradient of ea become zero for all a and hence the increments of the weight matrices become zero. If after a certain iteration W happens to violate the constraint Liiwlj ~ constant < I/max(J')2, then its elements are scaled so as to project it back onto the surface of the hypershere. Implementation Example A pattern recognition example is considered. Fig. 3 shows a set of two-dimensional training patterns from three classes. It is required to design a neural network recognizer with 28 three output neurons. Each of the neurons should be on if a sample of the corresponding class is presented, and off otherwise, i.e. we would like to design a "winner-take-all" network. A singlelayer three neuron feedback network is implemented. We obtained 3.3% error. Performing the same experiment on a feedforward single-layer network with three neurons, we obtained 20% error. For satisfactory results, a feedforward network should be two-layer. With one neuron in the first layer and three in the second layer, we got 36.7% error. Finally, with two neurons in the first layer and three in the second layer, we got a match with the feedback case, with 3.3% error . z z z z z z z z z z z z z z z z zil 1 33 3 3 3 1 3 3 33 3 3 3 ~ 3 3 3 3 3 3 Fig. 3 A pattern recognition example Conclusion A way to extend the backpropagation method to feedback networks has been proposed . A condition on the weight matrix is obtained, to insure having only one fixed point, so as to prevent having more than one possible output for a fixed input. A general structure for networks is presented, in which the network consists of a number of feedback groups connected to each other in a feedforward manner. A stochastic descent rule is used to update the weights. The lJ!ethod is applied to a pattern recognition example. With a single-layer feedback network it obtained good results. On the other hand, the feedforward backpropagation method achieved good resuls only for the case of more than one layer, hence also with a larger number of neurons than the feedback case. 29 Acknow ledgement The author would like to gratefully acknowledge Dr . Y. Abu-Mostafa for the useful discussions. This work is supported by Air Force Office of Scientific Research under Grant AFOSR-86-0296. Appendix Differentiating (5), one obtains aI - Yj a I = w kp aI k,p = 1, ... ,n, I Ym p jk , f '(')(,,", Zj L..,Wjm-a I +Y'6) w kp m where if j = k otherwise, and We can write a~' = (A' _ Wi) -lbkz> (A - 1) aw kp where b kp is the nt-dimensional vector whose b~l> = {y~ ? 0 ith component is given by ifi = k otherwise. By the chain rule, aea _ ""' ae a ay; -a I -L..,-aI - a I' w kp j Yj w kp which, upon substituting from (A - 1), can be put in the form y!,gk~' where gk is the column of (A' - Wt)-l. Finally, we obtain the required expression, which is ae" = [At _ (WI)T] aw' -1 ae" ( ,)T ayl y . Regarding a()~~I' it is obtained by differentiating (5) with respect to vr~,. We get similarly where C kl' is the nt-dimensional vector whose ith component is given by if i = k otherwise. kth 30 A derivation very similar to the case of :~l results in the following required expression: Be a = BVrl 8 8 [A' _ (w,)T] -1 Be a ( r)T. By' y 8yJ j Now, finally consider ~. Let ~, jf.B, be the matrix whose (k,p)th element is ~. The elements of ~ can be obtained by differentiating the equation for the fixed point for group . uy J, as follows, Hence, :~~. = (Ai - Wi) -IV'i. (A - 2) Using the chain rule, one can write Be a By' ?T = ~(ByJ) ~ Byl JEEr Be a By;' We substitute from (A - 2) into the previous equation to complete the derivation by obtaining References 111 P. Werbos, "Beyond regression: New tools for prediction and analysis in behavioral sciences", Harvard University dissertation, 1974. [21 D. Parker, "Learning logic", MIT Tech Report TR-47, Center for Computational Research in Economics and Management Science, 1985. [31 Y. Le Cun, "A learning scheme for asymmetric threshold network", Proceedings of Cognitiva, Paris, June 1985. [41 D. Rumelhart, G.Hinton, and R. Williams, "Learning internal representations by error propagation", in D. Rumelhart, J. McLelland and the PDP research group (Eds.), Parallel distributed processing: Explorations in the microstructure of cognition, Vol. 1, MIT Press, Cambridge, MA, 1986. 151 J. Hopfield, "Neurons with graded response have collective computational properties like those of two-state neurons", Proc. N atl. Acad. Sci. USA, May 1984. [61 L. Ahneida, " A learning rule for asynchronous perceptrons with feedback in a combinatorial environment", Proc. of the First Int. Annual Conf. on Neural Networks, San Diego, June 1987. [71 R. Rohwer, and B. Forrest, "Training time-dependence in neural networks", Proc. of the First Int. Annual Conf. on Neural Networks, San Diego, June 1987. [81 F. Pineda, "Generalization of back-propagation to recurrent neural networks", Phys. Rev. Lett., vol. 59, no. 19, 9 Nov. 1987.
9 |@word norm:1 tr:1 initial:4 configuration:1 nt:3 si:2 attracted:1 update:3 amir:1 kyk:1 beginning:2 ith:5 short:1 dissertation:1 iterates:1 differential:1 become:2 consists:2 prove:1 behavioral:1 manner:2 inter:1 expected:1 behavior:2 uz:2 becomes:1 begin:1 project:1 bounded:1 insure:1 what:1 developed:1 every:2 y3:1 scaled:1 control:1 grant:1 before:1 positive:2 engineering:1 local:2 tends:1 acad:1 might:2 uy:2 unique:6 yj:3 backpropagation:4 chaotic:1 got:2 get:6 cannot:2 close:1 layered:1 operator:1 onto:1 vrl:2 put:1 ilf:1 restriction:1 center:1 go:3 attention:1 economics:1 independently:1 williams:1 rule:4 atl:1 increment:1 diego:2 hypothesis:1 harvard:1 element:9 rumelhart:2 recognition:4 satisfying:2 updating:7 jk:1 asymmetric:3 werbos:1 electrical:1 ensures:1 connected:7 decrease:2 environment:1 ui:2 ultimately:2 solving:1 upon:2 hopfield:4 derivation:3 kp:6 quite:2 whose:9 widely:1 larger:1 say:1 otherwise:4 ability:2 final:2 hoc:1 pineda:1 differentiable:1 propose:2 interconnected:4 loop:2 moved:1 ky:1 requirement:2 recurrent:3 completion:1 op:1 implemented:2 direction:2 stochastic:3 exploration:1 vx:1 transient:1 settle:1 implementing:1 byj:1 feeding:1 microstructure:1 generalization:1 strictly:1 sufficiently:1 considered:3 exp:1 equilibrium:15 cognition:1 mostafa:1 substituting:1 sought:1 purpose:1 recognizer:1 proc:3 combinatorial:1 schwarz:3 tool:1 weighted:1 mit:2 gaussian:1 avoid:1 office:1 resuls:1 june:3 unsatisfactory:1 tech:1 wf:1 sense:1 squaring:1 lj:1 relation:1 going:1 interested:1 overall:1 special:1 equal:1 once:1 never:1 having:4 shaped:1 represents:1 report:1 escape:1 few:1 attempt:1 thrown:1 intra:3 adjust:1 nl:3 chain:2 necessary:1 respective:1 unless:1 iv:1 re:1 column:2 oflearning:1 successful:1 too:2 aw:2 periodic:1 density:1 physic:1 yl:6 off:1 ym:1 quickly:1 again:1 satisfied:1 management:1 containing:1 opposed:1 possibly:3 dr:1 conf:2 american:1 udt:3 int:2 matter:1 ad:1 performed:2 reached:1 parallel:1 minimize:2 square:2 air:1 variance:1 produced:1 multiplying:1 reach:1 phys:1 ed:2 rohwer:1 proof:2 associated:1 ea:4 back:2 response:1 though:1 furthermore:1 until:3 hand:3 propagation:2 yf:2 aj:1 scientific:1 usa:1 effect:1 hence:6 entering:1 symmetric:1 satisfactory:1 ll:3 steady:2 ay:1 complete:2 motion:1 recently:1 fi:1 sigmoid:1 winner:1 mclelland:1 extend:1 refer:3 zil:1 cambridge:1 ai:5 similarly:1 teaching:1 gratefully:1 dj:1 stable:1 surface:1 feb:1 add:1 certain:3 inequality:6 minimum:3 ud:1 period:1 ii:2 violate:1 match:1 characterized:1 devised:1 ensuring:1 prediction:1 regression:1 ae:3 iteration:5 achieved:1 want:1 separately:1 addressed:1 completes:1 singular:2 posse:1 cognitiva:1 subject:1 elegant:1 near:1 backwards:1 feedforward:7 ideal:2 variety:1 zi:1 ifi:1 escaping:1 regarding:1 expression:4 ul:2 aea:1 proceed:1 cause:1 useful:1 byl:1 atiya:1 zj:1 write:3 vol:2 abu:1 group:39 threshold:1 prevent:1 year:1 sum:2 enforced:1 forrest:1 appendix:3 scaling:1 acceptable:1 layer:8 pay:2 annual:2 occur:1 constraint:5 argument:1 performing:3 department:1 developing:2 wi:7 partitioned:1 qu:1 rev:1 evolves:1 cun:1 happens:1 explained:1 taken:2 equation:4 know:1 generalizes:1 appropriate:1 existence:2 substitute:1 denotes:2 a4:1 giving:1 restrictive:1 especially:1 graded:1 objective:1 move:4 added:2 dependence:1 diagonal:2 exhibit:1 gradient:1 kth:1 sci:1 index:2 gk:2 acknow:1 negative:1 design:4 implementation:1 ethod:1 collective:1 neuron:15 sm:1 acknowledge:1 descent:3 hinton:1 looking:1 pdp:1 pair:1 required:4 kl:1 paris:1 connection:5 california:1 qa:2 beyond:1 pattern:5 appeared:1 max:4 suitable:1 force:1 recursion:1 scheme:2 improve:1 technology:1 lk:1 singlelayer:1 asymptotic:1 afosr:1 fully:3 basin:1 row:1 supported:1 last:1 transpose:1 asynchronous:1 side:2 allow:2 institute:2 face:1 differentiating:4 distributed:1 feedback:20 maxl:1 dimension:1 lett:1 stuck:1 author:1 san:2 lz:1 nov:1 obtains:2 logic:1 global:1 sequentially:1 un:1 zk:1 ca:1 obtaining:1 du:1 necessarily:1 whole:2 noise:6 allowed:1 fig:7 tl:1 parker:1 vr:1 xl:3 governed:1 lw:1 theorem:5 a3:1 adding:2 simply:2 u2:5 applies:1 aa:1 determines:1 ma:1 towards:1 jf:1 change:2 except:1 wjm:1 wt:1 perceptrons:1 internal:1
7,135
90
584 PHASOR NEURAL NETVORKS Andr~ J. Noest, N.I.B.R., NL-ll0S AZ Amsterdam, The Netherlands. ABSTRACT A novel network type is introduced which uses unit-length 2-vectors for local variables. As an example of its applications, associative memory nets are defined and their performance analyzed. Real systems corresponding to such 'phasor' models can be e.g. (neuro)biological networks of limit-cycle oscillators or optical resonators that have a hologram in their feedback path. INTRODUCTION Most neural network models use either binary local variables or scalars combined with sigmoidal nonlinearities. Rather awkward coding schemes have to be invoked if one wants to maintain linear relations between the local signals being processed in e.g. associative memory networks, since the nonlinearities necessary for any nontrivial computation act directly on the range of values assumed by the local variables. In addition, there is the problem of representing signals that take values from a space with a different topology, e.g. that of the circle, sphere, torus, etc. Practical examples of such a signal are the orientations of edges or the directions of local optic flow in images, or ~he phase of a set of (sound or EM) waves as they arrive on an array of detectors. Apart from the fact that 'circular' signals occur in technical as well as biological systems, there are indications that some parts of the brain (e.g. olfactory bulb, cf. Dr.B.Baird's contribution to these proceedings) can use limit-cycle oscillators formed by local feedback circuits as functional building blocks, even for signals without circular symmetry. Vith respect to technical implementations, I had speculated before the conference whether it could be useful to code information in the phase of the beams of optical neurocomputers, avoiding slow optical switching elements and using only (saturating) optical amplification and a ? American Institute of Physics 1988 585 hologram encoding the (complex) 'synaptic' weight factors. At the conference, I learnt that Prof. Dana Anderson had independently developed an optical device (cf. these proceedings) that basically works this way, at least in the slow-evolution limit of the dynamic hologram. Hopefully, some of the theory that I present here can be applied to his experiment. In turn, such implementations call for interesting extensions of the present models. BASIC ELEMENTS OF GENERAL PHASOR NETVORKS Here I study the perhaps simplest non-scalar network by using unitlength 2-vectors (phasors) as continuous local variables. The signals processed by the network are represented in the relative phaseangles. Thus, the nonlinearities (unit-length 'clipping') act orthogonally to the range of the variables coding the information. The behavior of the network is invariant under any rigid rotation of the complete set of phasors, representing an arbitrary choice of a global reference I phase. Statistical physicists will recognize the phasor model as a generalization of 02-spin models to include vector-valued couplings. All 2-vectors are treated algebraically as complex numbers, writing x for Ixl for the length, Ixl for the phase-angle, and the complex conjugate of a 2-vector x. A phasor network then consists of N?l phasors s. , with Is.l=l, 1 interacting via couplings c .. , with 1J C .. 11 1 = O. The c 1J .. are allowed to be complex-valued quantities. For optical implementations this is clearly a natural choice, but it may seem less so for biological systems. However, if the coupling between two limitcycle oscillators with frequency f is mediated via a path having propagationdelay d, then that coupling in fact acquires a phaseshift of f.d.2~ radians. Thus, complex couplings can represent such systems more faithfully than the usual models which neglect propagationdelays altogether. Only 2-point couplings are treated here, but multi-point couplings c.1)'k' etc., can be treated similarly. The dynamics of each phasor depends only on its local field h.= 1 !z:4~ c 1J .. s. J J + n. 1 where z is the number of inputs 586 c .. ~O per cell and n. is a local noise term (complex and Gaussian). 1J 1 Various dynamics are possible, and yield largely similar results: Continuous-time, parallel evolution: ("type A") d (/s./) = Ih. l.sin(/h.1 - Is./) (IT 1 1 1 1 s.(t+dt)= h.1 Ih. I , either serially in Discrete-time updating: 1 1 1 random i-sequence ("type B"), or in parallel for all i ("type C"). The natural time scale for type-B dynamics is obtained by scaling the discrete time-interval eft as ,.., liN ; type-C dynamics has cl't=l. LYAPUNOV FUNCTION (alias "ENERGY", or "HAMILTONIAN" ) If one limits the attention temporarily to purely deterministic (n.=O) models, then the question suggests itself whether a class of 1 couplings exists for which one can easily find a Lyapunov function i.e. a function of the network variables that is monotonic under the dynamics. A well-known example 1 is the 'energy' of the binary and scalar Hopfield models with symmetric interactions. It turns out that a very similar function exists for phasor networks with type-A or B dynamics and a Hermitian matrix of couplings. -H = L ? 1 Hermiticity (c .. 1J 5.1 h. 1 = =c .. ) makes J 1 (lIz) L 5.1 c 1J .. s. J ? . 1,J H real-valued and non-increasing in time. This can be shown as follows, e.g. for the serial dynamics (type B). Suppose, without loss of generality, that phasor i=l is updated. Then -z H Ls. c ' l sl 1>1 + sl' 2: c ' 1 5. i>l = + 1 z 51 h1 + 1 1 I.I: i ,j>l -s. 1 c .. s. 1J J + constant. 1 Vith Hermitian couplings, H becomes real-valued, and one also has l:c 1 ? i>l I:c' 5. i>1 1 l 1 Thus, - H - constant 1 s. 1 = z h1 . 51 h1 + sl h1 = 2 Re(sl h 1 ) Clearly, H is minimized with respect to sl by sl(t+1) = hll Ih11 ? Type-A dynamics has the same Lyapunovian, but type C is more complex. = The existence of Hermitian interactions and the corresponding energy function simplifies greatly the understanding and design of phasor networks, although non-Hermitian networks can still have a Lyapunov- 587 function, and even networks for which such a function is not readily found can be useful, as will be illustrated later. AN APPLICATION: ASSOCIATIVE MEMORY. A large class of collective computations, such as optimisations and content-addressable memory, can be realised with networks having an energy function. The basic idea is to define the relevant penalty function over the solution-space in the form of the generic 'energy' of the net, and simply let the network relax to minima of this energy. As a simple example, consider an associative memory built within the framework of Hermitian phasor networks. In order to store a set of patterns in the network, i.e. to make a set of special states (at least approximatively) into attractive fixed points of the dynamics, one needs to choose an appropriate set of couplings. One particularly simple way of doing this is via the phasor-analog of "Hebb's rule" (note the Hermiticity) p s(.k). s-(.k), h c .. = were s.(k).IS phasor 1. .In I earne d pattern k . IJ rk 1 J 1 The rule is understood to apply only to the input-sets 'i of each i. Such couplings should be realisable as holograms in optical networks, but they may seem unrealistic in the context of biological networks of oscillators since the phase-shift (e.g. corresponding to a delay) of a connection may not be changeable at will. However, the required coupling can still be implemented naturally if e.g. a few paths with different fixed delays exist between pairs of cells. The synaps in each path then simply becomes the projection of the complex coupling on the direction given by the phase of its path, i.e. it is just a classical Hebb-synapse that computes the correlation of its pre- and post-synaptic (imposed) signals, which now are phase-shifted versions of the phasors s~~)The required complex c .. are then realised as the 1 IJ vector sum over at least two signals arriving via distinct paths with corresponding phase-shift and real-valued synaps. Two paths suffice if they have orthogonal phase-shifts, but random phases will do as well if there are a reasonable number of paths. Ve need to have a concise way of expressing how 'near' any state of the net is to one or more of the stored patterns. A natural way 588 of doing this is via a set of p order parameters called "overlaps" N -(k) 1 s 1.. s.1 ? N 11: 1 I ; 1 Note the constraint on the p overlaps P I < k -< p - ? 2 Mk ~ 1 if all the patterns k are orthogonal, or merely random in the limit N-.QO. This will be assumed from now on. Also, one sees at once that the whole behaviour of the network does not depend on any rigid rotation of all phasors over some angle since H, Mk , c .. and the dynamics are invariant under 1J multiplication of all s. by a fixed phasor : s~ = S.s. with ISI=1. I I I Let us find the performance at low loading: N,p,z .. oo, with p/z.. O and zero local noise. Also assume an initial overlap m)O with only one pattern, say with k=1. Then the local field is hi 1 s~1~ hP~ Z 1 1 and h.* 1 f s~k) ~s .. k s~k~ 1 J J j' i 1 = -z I: sP~s. jl'i = J J (1) = m1 . si ? S ~ fs~k). L: s~k~s. z k=2 1 j(~i J J h(1) i + + O(1//Z) h71 , with O( ./( p-l) Iz') where S~f(i);ISI=1, . Thus, perfect recall (M 1=1) occurs in one 'pass' at loadings p/z ... O. EXACTLY SOLVABLE CASE: SPARSE and ASYMMETRIC couplings Although it would be interesting to develop the full thermodynamics of Hermitian phasor networks with p and z of order N (analogous to the analysis of the finite-T Hopfield model by the teams of Amit 2 and van Hemmen 3 ), I will analyse here instead a model with sparse, asymmetric connectivity, which has the great advantages of being exactly solvable with relative ease, and of being arguably more realistic biologically and more easily scalable technologically. In neurobiological networks a cell has up to z;10 4 asymmetric connections, whereas N;101~ This probably has the same reason as applies to most VLSI chips, namely to alleviate wiring problems. For my present purposes, the theoretical advantage of getting some exact results is of primary interest 4 Suppose each cell has z incoming connections from randomly selected other cells. The state of each cell at time t depends on at most zt . cells at time t=O. Thus, If z t ?N 112 and N large, then the respective 589 4 trees of 'ancestors' of any pair cells have no cells in common. In x particular, if z_ (logN) , for any finite x, then there are no common ancestors for any finite time t in the limit N-.OO. For fundamental information-theoretic reasons, one can hope to be able to store p patterns with p at most of order z for any sort of 2-point couplings. Important questions to be settled are: Yhat are the accuracy and speed of the recall process, and how large are the basins of the attractors representing recalled patterns? Take again initial conditions (t=O) with, say, m(t)= Hl > H>l = O. Allowing again local random Gaussian (complex) noise n., the local ? ld s become, In . now f amI'1'Iar notatIon, . h .= h(l) 1 n .? f Ie . + h*. + 1 1 1 1 As in the previous section, the h~l)term consists of the 'signal' 1 m(t).s. (modulo the rigid rotation S) and a random term of variance * 1 at most liz. For p _ z, the h. term becomes important. Being sums of * 1 z(p-1) phasors oriented randomly relative to the signal, the h. are 1 independent Gaussian zero-mean 2-vectors with variance (p-1)/z , as p,z and N.. oo . Finally, let the local noises n.1 have variance r2. Then the distribution of the s.(t+l) phasors can be found in terms of 1 * .? 2 the signal met) and the total variance a=(p/z)+r of the random h.+n 1 1 After somewhat tedious algebraic manipulations (to be reported in detail elsewhere) one obtains the dynamic behaviour of met) m(t+1) = F(m(t),a) for discrete parallel (type-C) dynamics, and d met) = F(m(t),a) - met) for type-A or type-B dynamics , Tt where the function m F(m,a) = +" 2 Idx.(1+cos2x).expl-(m.sinx) la].(l+erfl(m.cosx)/~) -1'C The attractive fixed points H* (a)= F(H * ,a) represent the retrieval accuracy when the loading-pIus-noise factor equals a. See figure 1. For a?l one obtains the expansion 1-H* (a) = a/4 + 3a 2 132 + O(a 3 ). The recall solutions vanish continuously as H*_(a -a) 112 at a =tc/4. c c One also obtains (at any t) the distribution of the phase scatter of the phasors around the ideal values occurring in the stored pattern. 590 P(/u./) 1 = (1/2n).exp(-m 2 /a).(1+I1t.L.exp(L 2 ).(1+erf(L? -(k) , and L = (m/la).cos(/u./) 1 where u.= s. s. 111 , (modulo S). Useful approximations for the high, respectively low M regimes are: M ?ra: PUu./) 1 (MIl'a1l).exp[-(M./u./)2 /a ] 1 ; I/u./1 ?"XI2 1 M ? f i : PUu./) = (1I21t).(1+L ?./;l) 1 Figure ~ RETRIEVAL-ERROR and BASIN OF ATTRACTION versus LOADING + NOISE. Q Q Q en Q Q .,; Q ".,; Q I: UI ..,) Q -C0 a. 1:) CD x - c- Q Q . Q Q Q '"Q 0 Q 0 Q 0 c: "'0.00 0.10 0.20 O. 30 0 ? 40 a 0 ? 50 = p/z 0 ? 60 + r-r O. 70 0 ? 80 0 ? 90 1. 00 591 DISCUSSION It has been shown that the usual binary or scalar neural networks can be generalized to phasor networks, and that the general structure of the theoretical analysis for their use as associative memories can be extended accordingly. This suggests that many of the other useful applications of neural nets (back-prop, etcJ can also be generalized to a phasor setting. This may be of interest both from the point of view of solving problems naturally posed in such a setting, as well as from that of enabling a wider range of physical implementations, such as networks of limit-cycle oscillators, phase-encoded optics, or maybe even Josephson-junctions. The performance of phasor networks turns out to be roughly similar to that of the scalar systems; the maximum capacity p/z=~/4 for phasor nets is slightly larger than its value 2/n for binary nets, but there is a seemingly faster growth of the recall error 1-M at small a (linear for phasors, against exp(-1/(2a? for binary nets). However, the latter measures cannot be compared directly since they stem from quite different order parameters. If one reduces recalled phasor patterns to binary information, performance is again similar. Finally, the present methods and results suggest several roads to further generalizations, some of which may be relevant with respect to natural or technical implementations. The first class of these involves local variables ranging over the k-sphere with k>l. The other generalizations involve breaking the O(n) (here n=2) symmetry of the system, either by forcing the variables to discrete positions on the circle (k-sphere), and/or by taking the interactions between two variables to be a more general function of the angular distance between them. Such models are now under development. REFERENCES 1. J.J.Hopfield, Proc.Nat.Acad.Sci.USA 79, 2554 (1982) and idem, Proc.Nat.Acad.Sci.USA 81, 3088 (1984). 2. D.J.Amit, H.Gutfreund and H.Sompolinski, Ann.Phys. 173, 30 (1987). 3. D.Grensing, R.Kuhn and J.L. van Hemmen, J.Phys.A 20, 2935 (1987). 4. B.Derrida, E.Gardner and A.Zippelius, Europhys.Lett. 4, 167 (1987)
90 |@word version:1 loading:4 c0:1 tedious:1 concise:1 ld:1 initial:2 si:1 scatter:1 readily:1 realistic:1 selected:1 device:1 accordingly:1 hamiltonian:1 sigmoidal:1 become:1 consists:2 hermitian:6 olfactory:1 ra:1 roughly:1 isi:2 behavior:1 multi:1 brain:1 increasing:1 becomes:3 notation:1 suffice:1 circuit:1 developed:1 gutfreund:1 zippelius:1 act:2 growth:1 exactly:2 unit:2 arguably:1 before:1 understood:1 local:15 limit:7 physicist:1 switching:1 acad:2 encoding:1 path:8 suggests:2 co:1 ease:1 range:3 practical:1 block:1 addressable:1 projection:1 pre:1 road:1 suggest:1 cannot:1 context:1 writing:1 deterministic:1 imposed:1 expl:1 attention:1 independently:1 l:1 rule:2 attraction:1 array:1 his:1 analogous:1 updated:1 suppose:2 modulo:2 exact:1 us:1 element:2 particularly:1 updating:1 asymmetric:3 cycle:3 ui:1 dynamic:14 depend:1 solving:1 purely:1 easily:2 hopfield:3 chip:1 represented:1 various:1 phasor:19 distinct:1 iar:1 europhys:1 quite:1 encoded:1 posed:1 valued:5 larger:1 say:2 neurocomputers:1 relax:1 erf:1 analyse:1 itself:1 associative:5 seemingly:1 sequence:1 indication:1 advantage:2 net:7 interaction:3 relevant:2 amplification:1 az:1 getting:1 perfect:1 wider:1 coupling:16 oo:3 develop:1 derrida:1 ij:2 implemented:1 involves:1 met:4 lyapunov:3 direction:2 kuhn:1 a1l:1 behaviour:2 generalization:3 alleviate:1 biological:4 extension:1 around:1 liz:2 exp:4 great:1 vith:2 purpose:1 proc:2 faithfully:1 hope:1 clearly:2 gaussian:3 rather:1 mil:1 greatly:1 rigid:3 idem:1 relation:1 vlsi:1 ancestor:2 orientation:1 logn:1 development:1 special:1 field:2 once:1 equal:1 having:2 minimized:1 few:1 randomly:2 oriented:1 recognize:1 ve:1 phase:12 attractor:1 maintain:1 sinx:1 interest:2 circular:2 analyzed:1 nl:1 edge:1 necessary:1 respective:1 orthogonal:2 tree:1 circle:2 re:1 theoretical:2 mk:2 technologically:1 clipping:1 h71:1 delay:2 stored:2 reported:1 learnt:1 my:1 combined:1 fundamental:1 ie:1 physic:1 continuously:1 connectivity:1 again:3 settled:1 choose:1 dr:1 american:1 nonlinearities:3 realisable:1 coding:2 baird:1 depends:2 later:1 h1:4 view:1 doing:2 realised:2 wave:1 sort:1 parallel:3 contribution:1 formed:1 spin:1 accuracy:2 variance:4 largely:1 yield:1 basically:1 detector:1 phys:2 synaptic:2 against:1 energy:6 frequency:1 naturally:2 radian:1 recall:4 back:1 dt:1 awkward:1 synapse:1 anderson:1 generality:1 just:1 angular:1 correlation:1 qo:1 hopefully:1 perhaps:1 usa:2 building:1 evolution:2 symmetric:1 illustrated:1 attractive:2 wiring:1 sin:1 acquires:1 generalized:2 complete:1 theoretic:1 tt:1 image:1 ranging:1 invoked:1 novel:1 limitcycle:1 common:2 rotation:3 functional:1 physical:1 jl:1 analog:1 he:1 eft:1 m1:1 expressing:1 similarly:1 hp:1 had:2 etc:2 apart:1 forcing:1 manipulation:1 store:2 binary:6 minimum:1 somewhat:1 hologram:4 algebraically:1 signal:11 full:1 sound:1 reduces:1 stem:1 technical:3 faster:1 sphere:3 lin:1 retrieval:2 serial:1 post:1 neuro:1 basic:2 scalable:1 optimisation:1 represent:2 cell:9 beam:1 whereas:1 want:1 addition:1 interval:1 probably:1 flow:1 seem:2 call:1 near:1 ideal:1 topology:1 simplifies:1 idea:1 shift:3 whether:2 penalty:1 f:1 algebraic:1 useful:4 involve:1 netherlands:1 maybe:1 processed:2 simplest:1 sl:6 exist:1 andr:1 shifted:1 per:1 discrete:4 iz:1 sompolinski:1 merely:1 josephson:1 sum:2 angle:2 arrive:1 reasonable:1 scaling:1 hi:1 nontrivial:1 occur:1 optic:2 constraint:1 speed:1 optical:7 conjugate:1 slightly:1 em:1 biologically:1 hl:1 invariant:2 turn:3 xi2:1 junction:1 apply:1 generic:1 appropriate:1 altogether:1 existence:1 changeable:1 cf:2 include:1 neglect:1 prof:1 amit:2 classical:1 question:2 quantity:1 occurs:1 primary:1 usual:2 distance:1 sci:2 capacity:1 idx:1 reason:2 length:3 code:1 implementation:5 design:1 collective:1 zt:1 allowing:1 noest:1 finite:3 enabling:1 extended:1 team:1 interacting:1 arbitrary:1 introduced:1 pair:2 required:2 namely:1 connection:3 recalled:2 able:1 pattern:9 regime:1 built:1 memory:6 unrealistic:1 overlap:3 treated:3 natural:4 serially:1 solvable:2 representing:3 scheme:1 hll:1 thermodynamics:1 orthogonally:1 alias:1 gardner:1 mediated:1 understanding:1 multiplication:1 relative:3 loss:1 interesting:2 dana:1 ixl:2 versus:1 bulb:1 basin:2 cd:1 elsewhere:1 arriving:1 institute:1 taking:1 sparse:2 van:2 feedback:2 lett:1 computes:1 obtains:3 neurobiological:1 global:1 incoming:1 assumed:2 continuous:2 symmetry:2 expansion:1 complex:10 cl:1 z_:1 sp:1 whole:1 noise:6 allowed:1 resonator:1 en:1 hemmen:2 hebb:2 slow:2 position:1 torus:1 vanish:1 breaking:1 rk:1 r2:1 exists:2 ih:2 nat:2 occurring:1 tc:1 simply:2 amsterdam:1 saturating:1 temporarily:1 scalar:5 approximatively:1 speculated:1 monotonic:1 applies:1 prop:1 ann:1 oscillator:5 content:1 ami:1 called:1 total:1 pas:1 la:2 i1t:1 latter:1 avoiding:1
7,136
900
Active Learning for Function Approximation Kah Kay Sung (sung@ai.mit.edu) Massachusetts Institute of Technology Artificial Intelligence Laboratory 545 Technology Square Cambridge, MA 02139 Partha Niyogi (pn@ai.mit.edu) Massachusetts Institute of Technology Artificial Intelligence Laboratory 545 Technology Square Cambridge, MA 02139 Abstract We develop a principled strategy to sample a function optimally for function approximation tasks within a Bayesian framework. Using ideas from optimal experiment design, we introduce an objective function (incorporating both bias and variance) to measure the degree of approximation, and the potential utility of the data points towards optimizing this objective. We show how the general strategy can be used to derive precise algorithms to select data for two cases: learning unit step functions and polynomial functions. In particular, we investigate whether such active algorithms can learn the target with fewer examples. We obtain theoretical and empirical results to suggest that this is the case. 1 INTRODUCTION AND MOTIVATION Learning from examples is a common supervised learning paradigm that hypothe- sizes a target concept given a stream of training examples that describes the concept. In function approximation, example-based learning can be formulated as synthesizing an approximation function for data sampled from an unknown target function (Poggio and Girosi, 1990). Active learning describes a class of example-based learning paradigms that seeks out new training examples from specific regions of the input space, instead of passively accepting examples from some data generating source. By judiciously selecting ex- 594 Kah Kay Sung, Parlha Niyogi amples instead of allowing for possible random sampling, active learning techniques can conceivably have faster learning rates and better approximation results than passive learning methods. This paper presents a Bayesian formulation for active learning within the function approximation framework. Specifically, here is the problem we want to address: Let Dn {(Xi, Yi)li 1, ... , n} be a set of n data points sampled from an unknown target function g, possibly in the presence of noise. Given an approximation function concept class, :F, where each f E :F has prior probability P;:-[J], one can use regularization techniques to approximate 9 from Dn (in the Bayes optimal sense) by means of a function 9 E:F. We want a strategy to determine at what input location one should sample the next data point, (XN+l, YN+d, in order to obtain the "best" possible Bayes optimal approximation of the unknown target function 9 with our concept class :F. = = The data sampling problem consists of two parts: 1) Defining what we mean by the "best" possible Bayes optimal approximation of an unknown target function. In this paper, we propose an optimality criterion for evaluating the "goodness" of a solution with respect to an unknown target function. 2) Formalizing precisely the task of determining where in input space to sample the next data point. We express the above mentioned optimality criterion as a cost function to be minimized, and the task of choosing the next sample as one of minimizing the cost function with respect to the input space location of the next sample point. Earlier work (Cohn, 1991; MacKay, 1992) have tried to use similar optimal experiment design (Fedorov, 1972) techniques to collect data that would provide maximum information about the target function. Our work differs from theirs in several respects. First, we use a different, and perhaps more general, optimality criterion for evaluating solutions to an unknown target function, based on a measure of function uncertainty that incorporates both bias and variance components of the total output generalization error. In contrast, MacKay and Cohn use only variance components in model parameter space. Second, we address the important sample complexity question, i.e., does the active strategy require fewer examples to learn the target to the same degree of uncertainty? Our results are stated in PAC-style (Valiant, 1984). After completion of this work, we learnt that Sollich (1994) had also recently developed a similar formulation to ours. His analysis is conducted in a statistical physics framework. The rest of the paper is organized as follows: Section 2, develops our active sampling paradigm. In Sections 3 and 4, we consider two classes offunctions for which active strategies are obtained, and investigate their performance both theoretically and empirically. 2 THE MATHEMATICAL FRAMEWORK In order to optimally select examples for a learning task, one should first have a clear notion of what an "ideal" learning goal is for the task. We can then measure an example's utility in terms of how well the example helps the learner achieve the Active Learning for Function Approximation 595 goal, and devise an active sampling strategy that selects examples with maximum potential utility. In this section, we propose one such learning goal - to find an approximation function g E :F that "best" estimates the unknown target function g. We then derive an example utility cost function for the goal and finally present a general procedure for selecting examples. 2.1 EVALUATING A SOLUTION TO AN UNKNOWN TARGET THE EXPECTED INTEGRATED SQUARED DIFFERENCE Let 9 be the target function that we want to estimate by means of an approximation function 9 E :F. If the target function 9 were known, then one natural measure of how well (or badly) g approximates 9 would be the Integrated Squared Difference (ISD) of the two functions: 8(g,g) = 1x", (g(x) - g(x))2dx. (1) Xlo In most function approximation tasks, the target 9 is unknown, so we clearly cannot express the quality of a learning result, g, in terms of g. We can, however, obtain an expected integrated squared difference (EISD) between the unknown target, g, and its estimate, g, by treating the unknown target 9 as a random variable from the approximation function concept class:F. Taking into account the n data points, D n , seen so far, we have the following a-posteriori likelihood for g: P(gIDn) ex PF[g]P(Dnlg). The expected integrated squared difference (EISD) between an unknown target, g, and its estimate, g, given D n , is thus: EF[8(g,g)IDn] = 2.2 f P(gIDn)8(g,g)dg 19EF =f PF[g]P(D n lg)8(y,g)dg. (2) 19EF SELECTING THE NEXT SAMPLE LOCATION We can now express our learning goal as minimizing the expected integrated squared difference (EISD) between the unknown target 9 and its estimate g. A reasonable sampling strategy would be to choose the next example from the input location that minimizes the EISD between g and the new estimate gn+l. How does one predict the new EISD that results from sampling the next data point at location Xn+l ? Suppose we also know the target output value (possibly noisy), Yn+l, at Xn+l. The EISD between 9 and its new estimate gn+l would then be EF[8(gn+l, g)IDn U (x n+l,Yn+d]' where gn+l can be recovered from Dn U (X n+l,Yn+l) via regularization. In reality, we do not know Yn+l, but we can derive for it the following conditional probability distribution: P(Yn+llxn+l,Dn) ex f P(Dn U(Xn+l,Yn+dlf)PF[J]df. (3) liEF This leads to the following expected value for the new EISD, if we sample our next data point at X n +l: U(Yn+lIDn , xn+1) = 1: P(Yn+lIXn+1' Dn)EF[8(gn+1' g)IDn U (Xn+b Yn+l)]dYn+1' ~l Clearly, the optimal input location to sample next is the location that minimizes cost function in Equation 4 (henceforth referred to as the total output uncertainty), l.e., (5) 596 Kah Kay Sung, Partha Niyogi 2.3 SUMMARY OF ACTIVE LEARNING PROCEDURE We summarize the key steps involved in finding the optimal next sample location: 1) Compute P(gIDn). This is the a-posteriori likelihood of the different functions 9 given Dn , the n data points seen so far. 2) Fix a new point Xn+l to sample. 3) Assume a value Yn+l for this Xn+l . One can compute P(glDn U (Xn+l' Yn+d) and hence the expected integrated squared difference between the target and its new estimate. This is given by EF[8(Un+l, g)IDn U (Xn+l' Yn+t>]. See also Equation 2. 4) At the given Xn+l, Yn+l has a probability distribution given by Equation 3. Averaging over all Yn+l 's, we obtain the total output uncertainty for Xn+l, given by U(Un+lIDn, xn+t> in Equation 4. 5) Sample at the input location that minimizes the total output uncertainty cost function. 3 EXAMPLE 1: UNIT STEP FUNCTIONS To demonstrate the usefulness of the above procedure, let us first consider the following simple class of indicator functions parameterized by a single parameter a which takes values in [0,1]. Thus = :F {I[a,l]IO ~ a ~ I} We obtain a prior peg = l[a,l]? by assuming that a has an a-priori uniform distribution on [0,1]. Assume that data, Dn = {(Xi;Yi); i 1, .. n} consistent with some unknown target function l[at,l] (which the learner is to approximate) has been obtained. We are interested in choosing a point x E [0,1] to sample which will provide us with maximal information. Following the general procedure outlined above we go through the following steps. = For ease of notation, let x R be the right most point belonging to Dn whose Y value maXi=l, ..n{xdYi = OJ. Similarly let XL millj=l, .. n{Xs!Yi I} and is 0, i.e., XR let XL - XR = W. = = = 1) We first need to get P(gIDn ). It is easy to show that P(g = l[a,l]IDn) =~ if a E [XR' XL]; 0 otherwise 2) Suppose we sample next at a particular distribution P(y = OIDn , x) = (XL - x) XL - XR = (XL - X E [0,1]' we would obtain Y with the x) if x E [XR' XL]; 1 if x ~ XR; 0 otherwise W = For a particular y, the new data set would be Dn+l Dn U (x, y) and the corresponding EISD can be easily obtained using the distribution P(gIDn+d. Averaging this over P(YIDn, x) as in step 4 of the general procedure, we obtain if x ~ x R or x otherwise ~ XL 597 Active Learning for Function Approximation Clearly the point which minimizes the expected total output uncertainty is the midpoint of XL and XR. . Xn+1 arg mm U(gIDn, x) (XL + xR)/2 = = XE[O,l] Thus applying the general procedure to this special case reduces to a binary search learning algorithm which queries the midpoint of XR and XL. An interesting question at this stage is whether such a strategy provably reduces the sample complexity; and if so, by how much. It is possible to. prove the following theorem which shows that for a certain pre-decided total output uncertainty value, the active learning algorithm takes fewer examples to learn the target to the same degree of total output uncertainty than a random drawing of examples according to a uniform distribution. Theorem 1 Suppose we want to collect examples so that we are guaranteed with high probability (i. e. probability> 1 - 6) that the total output uncertainty is less than f. Then a passive learner would require at least ~ In(I/6) examples while y(48f) the active strategy described earlier would require at most (1/2) In(1 / 12f) examples. 4 . EXAMPLE 2: THE CASE OF POLYNOMIALS In this section we turn our attention to a class of univariate polynomials (from [-5,5] to ~) of maximum degree K, i.e., K :F = {g(ao, ... , aK) = L: aixi} i:=O As before, the prior on :F is obtained here by assuming a prior on the parameters; in particular we assume that a = (ao, aI, ... , aK) has a multivariate normal distribution N(O, S). For simplicity, it is assumed that the parameters are independent, i.e., S is a diagonal matrix with Si,i = (7;' In this example we also incorporate noise (distributed normally according to N(O, (72)). As before, there is a target gt E G which the learner is to approximate on the basis of data. Suppose the learner is in possession of a data set Dn = {(Xi, Yi = gt(Xi) + 1]); i = 1 ... n} and is to receive another data point. The two options are 1) to sample the function at a point X according to a uniform distribution on the domain [-5,5] (passive learning) and 2) follow our principled active learning strategy to select the next point to be sampled. 4.1 ACTIVE STRATEGY Here we derive an exact expression for Xn+1 (the next query point) by applying the general procedure described earlier. Going through the steps as before, 1)It is possible to show that p(g(a)IDn ) = P(aIDn ) is again a multivariate normal distribution N(J1., 1: n ) where J1. L~l Yixi, Xi (1, Xi, xl, . .. , xf)T and 1 n T ) 1 1:S-l + _ " (x.x. n 2(72 L..J 1 1 = = = i:=l 2)Computation of the total output uncertainty U(gn+1IDn, x) requires several steps. Taking advantage of the Gaussian distribution on both the parameters a and the noise, we obtain (see Niyogi and Sung, 1995 for details): U(gIDn, x) = l1:n+1AI Kah Kay Sung. Partha Niyogi 598 20 20 i\ Noise=0.1 15 to/ I~J \ A "\ 10 15 0\ I'\\J\l r\ \ \.j\ 10 '\ "'- 5 0 Noise=1.0 \ J 50 10". 'v.J\ 5 0 50 Figure 1: Comparing active and passive learning average error rates at different sample noise levels. The two graphs above plot log error rates against number of samples. See text for detailed explanation. The solid and dashed curves are the active and passive learning error rates respectively. where A is a matrix of numbers whose i,j element is J~5 t(i+i- 2)dt . En+l has the same form as En and depends on the previous data, the priors, noise and Xn+l. When minimized over Xn+l, we get xn +! as the maximum utility location where the active learner should next sample the unknown target function. 4.2 SIMULATIONS We have performed several simulations to compare the performance of the active strategy developed in the previous section to that of a passive learner (who receives examples according to a uniform random distribution on the domain [-5,5]). The following issues have been investigated. 1) Average error rate as a function of the number of examples: Is it indeed the case that the active strategy has superior error performance for the same number of examples? To investigate this we generated 1000 test target polynomial functions (of maximum degree 9) according ~o the following Gaussian prior on the parameters: for each ai,P(ai) = N(0,0.9'). For each target polynomial, we collected data according to the active strategy as well as the passive (random) strategy for varying number of data points. Figure 1 shows the average error rate (i .e., the integrated squared difference between the actual target function and its estimate, averaged over the 1000 different target polynomials) as a function of the number of data points. Notice that the active strategy has a lower error rate than the passive for the same number of examples and is particularly true for small number of data. The active strategy uses the same priors that generate the test target functions. We show results of the same simulation performed at two noise levels (noise standard deviation 0.1 and 1.0). In both cases the active strategy outperforms the passive learner indicating robustness in the face of noise. 2) Incorrect priors: How sensitive is the active learner to possible differences between its prior assumptions on the class :F and the true priors? We repeated the function learning task of the earlier case with the test targets generated in the same way as before. The active learner assumes a slightly different Gaussian prior and polynomial degree from the target (Std(ai) 0.7i and K 7 for the active learner versus Std(ad = O.Si and K = S for the target). Despite its inaccurate priors, the = = 599 Active Learning for Function Approximation 8.5 r--------r----r---~--...---___, OJ 8 1tS ~7.5 o Y' '- Different Gauss Prior & Deg. \ \ "- ~ ~ w 7 C) o -6.5 -- '\ ""' ....... ........ --... ....... -- -- 6~--~~--~----~----~----~ o 10 20 30 40 50 Figure 2: Active lea.rning results with different Gaussian priors for coefficients, and a lower a priori polynomial degree K. See text for detailed explanation. The solid and dashed curves a.re the active a.nd passive learning error rates respectively. active learner outperforms the passive case. 3) The distribution of points: How does the active learner choose to sample the domain for maximally reducing uncertainty? There are a few sampling trends which are noteworthy here. First, the learner does not simply sample the domain on a uniform grid. Instead it chooses to cluster its samples typically around K + 1 locations for concept classes with maximum degree K as borne out by simulations where K varies from 5 to 9. One possible explanation for this is it takes only K + 1 points to determine the target in the absence of noise. Second, as the noise increases, although the number of clusters remains fixed, they tend to be distributed away from the origin. It seems that for higher noise levels, there is less pressure to fit the data closely; consequently the prior assumption of lower order polynomials dominates. For such lower order polynomials, it is profitable to sample away from the origin as it reduces the variance of the resulting fit. (Note the case of linear regression) . Remarks 1) Notice that because the class of polynomials is linear in its model parameters, a, the new sample location (xn+d does not depend on the y values actually observed but only on the x values sampled . Thus if the learner is to collect n data points, it can pre-compute the n points at which to sample from the start. In this sense the active algorithm is not really adaptive. This behavior has also been observed by MacKay (1992) and Sollich (1994) . 2) Needless to say, the general framework from optimal design can be used for any function class within a Bayesian framework. We are currently investigating the possibility of developing active strategies for Radial Basis Function networks. While it is possible to compute exact expressions for Xn+l for such RBF networks with fixed centers, for the case of moving centers, one has to resort to numerical minimization. For lack of space we do not include those results in this paper. 600 Kah Kay Sung, Partha Niyogi 10 (DO 0 00 (D CJX:? 0 o 2 8 o 00 (I) 0 0 00 6 1 0 0 o~~~~----~~~~ -5 o 5 4 -5 0 0 00 0 0 0 5 Figure 3: Distribution of active learning sample points as a function of (i) noise strength and (ii) a-priori polynomial degree. The horizontal axis of both graphs represents the input space [-5,5]. Each circle indicates a sample location. The Left graph shows the distribution of sample locations (on x axis) for different noise level (indicated on y-axis). The Right graph shows the distribution of sample locations (on x-axis) for different assumptions on the maximum polynomial degree K (indicated on y-axis) . 5 CONCLUSIONS We have developed a Bayesian framework for active learning using ideas from optimal experiment design. Our focus has been to investigate the possibility of improved sample complexity using such active learning schemes. For a simple case of unit step functions, we are able to derive a binary search algorithm from a completely different standpoint. Such an algorithm then provably requires fewer examples for the same error rate. We then show how to derive specific algorithms for the case of polynomials and carry out extensive simulations to compare their performance against the benchmark of a passive learner with encouraging results. This is an application of the optimal design paradigm to function learning and seems to bear promise for the design of more efficient learning algorithms. References D. Cohn. (1991) A Local Approach to Optimal Queries. In D. Touretzky (ed.), Proc. of 1990 Connectionist Summer School, San Mateo, CA , 1991. Morgan Kaufmann Publishers. V. Fedorov. (1972) Theory of Optimal Experiments. Academic Press, New York, 1972. D. MacKay. (1992) Bayesian Methods for Adaptive Models. PhD thesis, CalTech, 1992. P. Niyogi and K. Sung. (1995) Active Learning for Function Approximation: Paradigms from Optimal Experiment Design. Tech Report AIM-1483, AI Lab. , MIT, In Preparation. M. Plutowski and H. White. (1991) Active Selection of Training Examples for Network Learning in Noiseless Environments. Tech Report CS91-180, Dept. of Computer Science and Engineering, University of California, San Diego, 1991. T. Poggio and F. Girosi. (1990) Regularization Algorithms for Learning that are Equivalent to Multilayer Networks. Science, 247:978-982, 1990. P. Sollich. (1994) Query Construction, Entropy, Generalization in Neural Network Models. Physical Review E, 49 :4637-4651, 1994. 1. Valiant. (1984) A Theory of Learnable. Proc. of the 1984 STOC, p436-445, 1984.
900 |@word polynomial:14 seems:2 nd:1 seek:1 tried:1 simulation:5 pressure:1 solid:2 carry:1 selecting:3 ours:1 outperforms:2 recovered:1 comparing:1 si:2 dx:1 numerical:1 j1:2 girosi:2 offunctions:1 treating:1 plot:1 intelligence:2 fewer:4 accepting:1 cjx:1 location:15 mathematical:1 dn:12 incorrect:1 consists:1 prove:1 introduce:1 theoretically:1 indeed:1 expected:7 behavior:1 actual:1 encouraging:1 pf:3 notation:1 formalizing:1 what:3 minimizes:4 developed:3 finding:1 possession:1 sung:8 unit:3 normally:1 yn:15 before:4 engineering:1 local:1 io:1 despite:1 ak:2 noteworthy:1 lief:1 mateo:1 collect:3 ease:1 averaged:1 kah:5 decided:1 differs:1 xr:9 procedure:7 rning:1 empirical:1 pre:2 radial:1 suggest:1 get:2 cannot:1 needle:1 selection:1 applying:2 equivalent:1 center:2 go:1 attention:1 simplicity:1 kay:5 his:1 notion:1 profitable:1 target:34 suppose:4 diego:1 construction:1 exact:2 aixi:1 us:1 origin:2 element:1 trend:1 particularly:1 std:2 observed:2 region:1 principled:2 mentioned:1 environment:1 complexity:3 depend:1 learner:16 basis:2 completely:1 easily:1 artificial:2 query:4 choosing:2 whose:2 say:1 drawing:1 otherwise:3 niyogi:7 noisy:1 advantage:1 propose:2 maximal:1 achieve:1 cluster:2 generating:1 help:1 derive:6 develop:1 completion:1 school:1 closely:1 require:3 fix:1 generalization:2 ao:2 really:1 mm:1 around:1 normal:2 predict:1 proc:2 currently:1 sensitive:1 minimization:1 mit:3 clearly:3 gaussian:4 aim:1 pn:1 varying:1 dlf:1 focus:1 likelihood:2 indicates:1 tech:2 contrast:1 sense:2 posteriori:2 inaccurate:1 integrated:7 typically:1 going:1 selects:1 interested:1 provably:2 arg:1 issue:1 priori:3 special:1 mackay:4 sampling:7 represents:1 minimized:2 connectionist:1 report:2 develops:1 few:1 dg:2 investigate:4 possibility:2 dyn:1 poggio:2 re:1 circle:1 theoretical:1 cs91:1 earlier:4 gn:6 goodness:1 cost:5 deviation:1 uniform:5 usefulness:1 conducted:1 optimally:2 varies:1 learnt:1 chooses:1 physic:1 squared:7 again:1 thesis:1 choose:2 possibly:2 henceforth:1 borne:1 resort:1 style:1 li:1 account:1 potential:2 amples:1 coefficient:1 depends:1 stream:1 ad:1 performed:2 lab:1 start:1 bayes:3 option:1 partha:4 square:2 variance:4 who:1 kaufmann:1 bayesian:5 touretzky:1 ed:1 against:2 involved:1 sampled:4 massachusetts:2 organized:1 actually:1 higher:1 dt:1 supervised:1 follow:1 maximally:1 improved:1 formulation:2 stage:1 receives:1 horizontal:1 cohn:3 lack:1 quality:1 indicated:2 perhaps:1 concept:6 true:2 regularization:3 hence:1 laboratory:2 white:1 criterion:3 demonstrate:1 l1:1 passive:12 ef:6 recently:1 common:1 superior:1 empirically:1 physical:1 approximates:1 theirs:1 cambridge:2 ai:8 outlined:1 grid:1 similarly:1 had:1 moving:1 gt:2 multivariate:2 optimizing:1 certain:1 binary:2 xe:1 yi:4 devise:1 caltech:1 plutowski:1 seen:2 morgan:1 determine:2 paradigm:5 dashed:2 ii:1 reduces:3 faster:1 xf:1 academic:1 regression:1 multilayer:1 noiseless:1 df:1 lea:1 receive:1 want:4 source:1 standpoint:1 publisher:1 rest:1 tend:1 incorporates:1 presence:1 ideal:1 easy:1 fit:2 idea:2 judiciously:1 whether:2 expression:2 utility:5 york:1 remark:1 clear:1 detailed:2 generate:1 peg:1 notice:2 promise:1 express:3 key:1 isd:1 graph:4 parameterized:1 uncertainty:11 reasonable:1 guaranteed:1 summer:1 badly:1 strength:1 precisely:1 optimality:3 passively:1 hypothe:1 developing:1 according:6 idn:7 belonging:1 describes:2 sollich:3 slightly:1 conceivably:1 equation:4 remains:1 turn:1 know:2 away:2 robustness:1 assumes:1 include:1 objective:2 question:2 strategy:19 diagonal:1 collected:1 assuming:2 minimizing:2 lg:1 stoc:1 stated:1 synthesizing:1 design:7 unknown:15 allowing:1 fedorov:2 benchmark:1 t:1 defining:1 precise:1 extensive:1 california:1 address:2 able:1 summarize:1 oj:2 explanation:3 natural:1 indicator:1 scheme:1 technology:4 axis:5 text:2 prior:15 review:1 determining:1 bear:1 interesting:1 versus:1 degree:10 consistent:1 summary:1 bias:2 institute:2 taking:2 face:1 midpoint:2 distributed:2 curve:2 xn:20 evaluating:3 adaptive:2 san:2 far:2 approximate:3 deg:1 active:40 investigating:1 assumed:1 xi:6 un:2 search:2 reality:1 learn:3 yixi:1 ca:1 investigated:1 domain:4 motivation:1 noise:15 repeated:1 referred:1 en:2 xl:12 theorem:2 specific:2 pac:1 maxi:1 learnable:1 x:1 dominates:1 incorporating:1 valiant:2 phd:1 entropy:1 simply:1 univariate:1 xlo:1 ma:2 conditional:1 goal:5 formulated:1 consequently:1 rbf:1 towards:1 absence:1 specifically:1 reducing:1 averaging:2 total:9 gauss:1 indicating:1 select:3 preparation:1 incorporate:1 dept:1 ex:3
7,137
901
Higher Order Statistical Decorrelation without Information Loss Gustavo Deco SiemensAG Central Research Otto-Hahn-Ring 6 81739 Munich GeIIDany Wilfried Brauer Technische UniversiUit MUnchen Institut fur InfoIIDatik Arcisstr. 21 80290 Munich GeIIDany Abstract A neural network learning paradigm based on information theory is proposed as a way to perform in an unsupervised fashion, redundancy reduction among the elements of the output layer without loss of information from the sensory input. The model developed performs nonlinear decorrelation up to higher orders of the cumulant tensors and results in probabilistic ally independent components of the output layer. This means that we don't need to assume Gaussian distribution neither at the input nor at the output. The theory presented is related to the unsupervised-learning theory of Barlow, which proposes redundancy reduction as the goal of cognition. When nonlinear units are used nonlinear principal component analysis is obtained. In this case nonlinear manifolds can be reduced to minimum dimension manifolds. If such units are used the network performs a generalized principal component analysis in the sense that non-Gaussian distributions can be linearly decorrelated and higher orders of the correlation tensors are also taken into account. The basic structure of the architecture involves a general transfOlmation that is volume conserving and therefore the entropy, yielding a map without loss of infoIIDation. Minimization of the mutual infoIIDation among the output neurons eliminates the redundancy between the outputs and results in statistical decorrelation of the extracted features. This is known as factorialleaming. 248 Gustavo Deco, Wilfried Brauer 1 INTRODUCTION One of the most important theories of feature extraction is the one proposed by Barlow (1989). Barlow describes the process of cognition as a preprocessing of the sensorial information performed by the nervous system in order to extract the statistically relevant and independent features of the inputs without loosing information. This means that the brain should statistically decorrelate the extracted information. As a learning strategy Barlow (1989) formulated the principle of redundancy reduction. This kind of learning is called factorial learning. Recently Atick and Redlich (1992) and Redlich (1993) concentrate on the original idea of Barlow yielding a very interesting formulation of early visual processing and factorial learning. Redlich (1993) reduces redundancy at the input by using a network structure which is a reversible cellular automaton and therefore guarantees the conservation of information in the transformation between input and output. Some nonlinear extensions of PCA for decorrelation of sensorial input signals were recently introduced. These follow very closely Barlow's original ideas of unsupervised learning. Redlich (1993) use similar information theoretic concepts and reversible cellular automata architectures in order to define how nonlinear decorrelation can be performed. The aim of our work is to formulate a neural network architecture and a novel learning paradigm that performs Barlow's unsupervised learning in the most general fashion. The basic idea is to define an architecture that assures perfect transmission without loss of information. Consequently the nonlinear transformation defined by the neural architecture is always bijective. The architecture performs a volume-conserving transformation (determinant of the Jacobian matrix is equal one). As a particular case we can derive the reversible cellular automata architecture proposed by Redlich (1993). The learning paradigm is defined so that the components of the output signal are statistically decorrelated. Due to the fact that the output distribution is not necessarily Gaussian, even if the input is Gaussian, we perform a cumulant expansion of the output distribution and find the rules that should be satisfied by the higher order correlation tensors in order to be decorrelated. 2 THEORETICAL FORMALISM Let us consider an input vector x of dimensionality d with components distributed according to the probability distribution P (X) , which is not factorial, i.e. the components of x are correlated. The goal of Barlow's unsupervised learning rule is to find a transformation ,. ->,. y = F (x) (2.1) such that the components of the output vector d -dimensional yare statistically decorrelated. This means that the probability distributions of the components Y j are independent and therefore, d pO) = IT P (y) . (2.2) j The objl?,.ctive of factorial learning is to find a neural network, which performs the transformation F ( ) such that the joint probability distribution P 0) of the output signals is factorized as in eq. (2.2). In order to implement factorial learning, the information contained in the input should be transferred to the output neurons without loss but, the probability distribution of the output neurons should be statistically decorrelated. Let us now define Higher Order Statistical Decorrelation without /n/onl1atioll Loss 249 these facts from the information theory perspective. The first aspect is to assure the entropy is conselVed, i.e. H (x) = H (y) (2.3) where the symbol !! (a) denotes the entropy of a and H (a/b) the conditional entropy of given b. One way to achieve this goal is to construct an architecture that independently of its synaptic parameters satisfies always eq. (2.3). Thus the architecture will conselVe information or entropy. The transmitted entropy satisfies a H (J) ~ H (x) + JP (x) In (det (~r? dx (2.4) -" where equality holds only if F is bijective, i.e. reversible. ConselVation of information and bijectivity is assured if the neural transformation conselVes the volume, which mathematically can be expressed by the fact that the Jacobian of the transformation should have determinant unity. In section 3 we fonnulate an architecture that always conselVes the entropy. Let us now concentrate on the main aspect of factorial learning, namely the decorrelation of the output components. Here the problem is to find a volume-conselVing transformation that satisfies eq. (2.2). The major problem is that the distribution of the output signal will not necessarily be Gaussian. Therefore it is impossible to use the technique of minimizing the mutual information between the components of the output as done by Redlich (1993). The only way to decorrelate non-Gaussian distributions is to expand the distribution in higher orders of the correlation matrix and impose the independence condition of eq. (2.2). In order to achieve this we propose to use a cumulant expansion of the output distribution. Let us define the Fourier transform of the output distribution, ~(K) = JdY ei(K?h P(J;~(Ki) = fdYi ei(Ki?Yi) P(y) (2.5) The cumulant expansion of a distribution is (Papoulis, 1991) (2.6) In the Fourier space the independence condition is given by (papoulis, 1991) (2.7) which is equivalent to ...>. In(~(K? = Incrl~(Ki? = Eln(~(Ki? I (2.8) I Putting eq. (2.8) and the cumulant expansions of eq. (2.6) together, we obtain that in the case of independence the following equality is satisfied (2.9) In both expansions we will only consider the first four cumulants. Mter an extra transformation ~ .. -,.- y' .. y - (y) (2.10) Gustavo Deco, Wilfried Brauer 250 to remove the bias <y). we can rewrite eq. (2.9) using the cumulants expression derived in the Papoulius (1991): 1 -21: KjKj {Cjj - Cj (2) '.J 1 + 24 i 5jj} -"6 E KiKjKd Cijl.: - Cj 5jjk } (3) (2.11) '.J.k E KjKjKkKd (Cjjkl-3CijCkl) - (4) (Cj (2) -3 (C; 2 ? 5jjkl } =0 i ,j, t , I Equation (2.11) should be satisfied for all values of K. The multidimensional correlation tensors C j . .. j and the one-dimensional higher order moments C?) are given by ~j ... j = Jdy' P(y') y'j ... y'j ; C?) = Jdy'j P(y'j) (y'i) (2.12) The 5. . denotes Kroenecker's delta. Due to the fact that eq. (2.11) should be satisfied for all K,'iiU coefficients in each summation of eq. (2.11) must be zero. This means that C jj = 0, if(i oF j) (2.13) C jjk C jjkl = 0, - 0, if(i oF j v i oF k) (2.l4) if( {ioFjvioFkvioFl} I\-,L) C Ujj - CijCjj = 0, if(i oF j) . (2.15) (2.16) In eq. (2.15) L is the logical expression L - {(i .. j 1\ k .. 11\ j oF k) v (i = k 1\ j = I 1\ i oF j) v (i - 11\ j .. k 1\ i oF j) }, (2.17) which excludes the cases considered in eq. (2.16). The conditions of independence given by eqs. (2. 13-2.l6) can be achieved by minimization of the cost function E - aE Cfj+ 13i<j~k E Cfjk+Yi<j~';'1 , Ctkl+ 5E (C;;jj - CjjCjj) 2 ;<j (2.18) i<j where a, 13, Y, 5 are the inverse of the number of elements in each summation respectively. In conclusion, minimizing the cost given by eq. (2.18) with a volume-conserving network, we achieve nonlinear decorrelation of non-Gaussian distributions. It is very easy to test wether a factorized probability distribution (eq. 2.2) satisfies the eqs. (2.13-2.16). As a particular case if only second order terms are used in the cumulant expansion, the learning rule reduces to eq. (2.13), which expresses nothing more than the diagonalization of the second order covariance matrix. In this case, by anti-transforming the cumulant expansion of the Fourier transform of the distribution,we obtain a Gaussian distribution. Diagonalization of the covariance matrix decorrelates statistically the components of the output only if we assume a Gaussian distribution of the outputs. In general the distribution of the output is not Gaussian and therefore higher orders of the cumulant expansion should be taken into account. yielding the learning rule conditions eqs. (2.13-2.l6) (up to fourth order, generalization to higher orders is straightforward). In the case of Gaussian distribution, minimization of the sum of the variances at each output leads to statistically decorrelation. This fact has a nice information theoretic background namely the minimization of the mutual information between the output components. Statistical independence as expressed in eq. (2.2) is equivalent to (Atick and Redlich. 1992) MH = EH(Yj) J -H(~) =0 (2.19) Higher Order Statistical Decorrelatioll without Illformation Loss 251 This means that in order to minimize the redundancy at the output we minimize the mutual information between the different components of the output vector. Due to the fact that the volume-conserving structure of the neural network conserves the entropy, the minimizaH (Yj) . tion of MH reduces to the minimization of E J 3 VOLUME-CONSERVING ARCHITECTURE AND LEARNING RULE In this section we define a neural network architecture that is volume-conserving and therefore can be used for the implementation of the learning rules described in the last section. Figure l.a shows the basic architecture of one layer. The dimensionality of input and output layer is the same and equal to d. A similar architecture was proposed by Redlich (1993b) using the theory of reversible cellular automata. ~ x ~ x ~ x ~ Y (c) (b) (a) Figure 1: Volume-conserving Neural Architectures. The analytical definition of the transformation defined by this architecture can be written as, )'i = xi +1, (x o' ... , Xj' Wi)' with j < i (3.1) where Wi represents a set of parameters of the function h' Note that independently of the functions fi the network is always volume-conserving. In particular h can be calculated by another neural network, by a sigmoid neuron, by polynomials (higher order neurons), etc. Due to the asymmetric dependence on the input variables and the direct connections with weights equal to 1 between corresponding components of input and output neurons, the Jacobian matrix of the transformation defined in eq. (3.1) is an upper triangular matrix with diagonal elements all equal to one, yielding a determinant equal to one. A network with inverted asymmetry also can be defined as )'i = Xi + gi (xj' ... , Xd' 8;>, with i <j (3.2) corresponding to a lower triangular Jacobian matrix 'fith diagonal elements all equal to one, being therefore volume-conserving. The vectors 8 j represent the parameters of functions gj. In order to yield a general nonlinear transformation from inputs to outputs (without asymmetric dependences) it is possible to build a multilayer architecture like the one shown in Fig. 1.b, which involves mixed versions of networks described by eq. (3.1) and eq. (3.2), respectively. Due to the fact that successive application of volume-conserving transformation is also volume-conserving, the multilayer architecture is also volume-con- 252 Gustavo Deco, Wilfried Brauer serving. In the two-layer case (Fig. I.c) the second layer can be interpreted as asymmetric lateral connections between the neurons of the first layer. However, in our case the feedfOlWard connections between input layer and first layer are also asymmetric. As demonstrated in the last section, we minimize a cost function E to decorrelate nonlinearly correlated non-Gaussian inputs. Let us analyze for simplicity a two-layer architecture (Fig. l.c) with the first layer given by eq. (3.l) and the second layer by eq. (3.2). Let us denote the output of the hidden layer by h and use it as input of eq. (3.2) with output y. The extension to multilayer architectures is straightfolWard. The learning rule can be easily expressed by gradient descent method: aE e.=e.-T\ae; .>. .>. I I (3.3) ..>. In order to calculate the derivative of the cost functions we need a _ _ Ci ... j 1 == N-~ {",,\"::"(y.-y .) ... (y .-y.) o i..J P 00 I I ] ] _ a _. ayi 1 a + (y.-y.) ... ""\"::"(y . -y, -a = -~ {,\::"(y .)}(3.4) I I 00 ] 0 N i..J 00 I p J e e where represents the parameters j and Wi. The sums in both equations extend over the N training patterns. The gradients of the different outputs are a _ a .j _ a a a W; Wi ---;:-Yi - - -g; .. ,~Yt - (ahgt) (a ..... !;)&i>k+ (a ..... !;) ae; ae; [OJ I (3.5) where &i> k is equal to I if i > k and 0 othelWise. In this paper we choose a polynomial fonn for the functions I and g. This model involves higher order neurons. In this case each function Ii or g; is a product of polynomial functions of the inputs. The update equations are given by (3.6) where R is the order of the polynomial used. In this case the two-layer architecture is a higher order network with a general volume-conserving structure. The derivatives involved in the learning rule are given by a _ -a-Yk 0) ? . ljr (3.7) 253 Higher Order Statistical Decorrelation without Information Loss 4 RESULTS AND SIMULATIONS We will present herein two different experiments using the architecture defined in this paper. The input space in all experiments is two-dimensional in order to show graphically the results and effects of the presented model. The experiments aim at learning noisy nonlinear polynomial and rational curves. Figure 2.a and 2.b plot the input and output space of the second experiment after training is finished, respectively. In this case the noisy logistic map was used to generate the input: (4.1) where'\) introduces I % Gaussian noise. In this case a one-layer polynomial network with R - 2 was used. The learning constant was 11 - 0.01 and 20000 iterations of training were performed. The result of Fig. 2.b is remarkable. The volume-conserving network decorrelated the output space extracting the strong nonlinear correlation that generated the curve in the input space. This means that after training only one coordinate is important to describe the curve. (b) (a) ?? ?? o? os ?? 04 02 ' ?1'=" .2 ---.,....---:.:'-:-.---:.:':-.---,.,':-.-----::':: ..-----7----,,' .. .g7.-~-~.:'-:-2---,O~.--0~.-~ ??~~-~ Figure 2: Input and Output space distribution after training with a one-layer polynomial volume-conseIVing network of order for the logistic map. (a) input space; (b) output space. The whole information was compressed into the first coordinate of the output. This is the generalization of data compression normally performed by using linear peA (also called Karhunen-Loewe transformation). The next experiment is similar, but in this case a twolayer network of order R .. 4 was used. The input space is given by the rational function 3 x2 = O.2x I + Xl 2 (1 +x I ) + '\l (4.2) where Xl and '\) are as in the last case. The results are shown in Fig. 4.a (input space) and Fig. 4.b (output space). Fig. 4.c shows the evolution of the four summands of eq (2.18) during learning. It is important to remark that at the beginning the tensors of second and third order are equally important. During learning all summands are simultaneously minimized, resulting in a statistically decorrelated output. The training was performed during 20000 iterations and the learning constant was 11 = 0.005 . 254 Gustavo Deco, Wilfried Brauer (b) (a) " (c) " etrIlIt' II. " .. .... " " ?? .- /' ~st2 II ? .l I. I ~. _,,, ...... H"'t . I. , . . . . . . .1? .. -"' - ' ... . . t. ,'" ..I ... ...t.. I .. 'I. .. . " '1. " " " cost4b .~ .. II. M.. ,- Figure 4: Input and Output space distribution after training with a two-layer polynomial volume-conselVing network of order for the noisy CUIVe of~. (4.2). (a) input space; (b) output space (c) Development of the four summands of the cost function (~ 2.18) during learning: (cost2) fiist summand (second order COIrelation tensor); (cost 3) second summand (tliird correlation order tensor); (cost 4a) third summand (fourth order correlation tensor); (cost4b) fourth summand (fourth order correlation tensor). 5 CONCLUSIONS We proposed a unsupervised neural paradigm, which is based on Infonnation Theory. The algorithm perfonns redundancy reduction among the elements of the output layer without loosing infonnation, as the data is sent through the network. The model developed perfonns a generalization of Barlow's unsupervised learning, which consists in nonlinear decorrelation up to higher orders of the cumulant tensors. After training the components of the output layer are statistically independent. Due to the use of higher order cumulant expansion arbitrary non-Gaussian distributions can be rigorously handled. When nonlinear units are used nonlinear principal component analysis is obtained. In this case nonlinear manifolds can be reduced to a minimum dimension manifolds. When linear units are used, the network performs a generalized principal component analysis in the sense that non-Gaussian distribution can be linearly decorrelated.This paper generalizes previous works on factorial learning in two ways: the architecture performs a general nonlinear transformation without loss of information and the decorrelation is perfonned without assuming Gaussian distributions. References: H. Barlow. (1989) Unsupervised Learning. Neural Computation, 1,295-311. A. Papoulis. (1991) Probability, Random Variables, and Stochastic Processes. 3. Edition, McGraw-Hill, New York. A. N. Redlich. (1993) Supervised Factorial Learning. Neural Computation, 5, 750-766.
901 |@word determinant:3 version:1 compression:1 polynomial:8 simulation:1 covariance:2 decorrelate:3 fonn:1 twolayer:1 papoulis:3 moment:1 reduction:4 dx:1 must:1 written:1 remove:1 plot:1 update:1 nervous:1 beginning:1 successive:1 direct:1 consists:1 nor:1 brain:1 factorized:2 kind:1 interpreted:1 developed:2 transformation:15 guarantee:1 multidimensional:1 xd:1 unit:4 normally:1 g7:1 statistically:9 yj:2 cfj:1 implement:1 impossible:1 equivalent:2 map:3 demonstrated:1 yt:1 straightforward:1 graphically:1 independently:2 automaton:4 formulate:1 iiu:1 simplicity:1 rule:8 coordinate:2 element:5 assure:1 conserve:1 asymmetric:4 calculate:1 yk:1 transforming:1 brauer:5 rigorously:1 rewrite:1 po:1 joint:1 mh:2 easily:1 describe:1 universiuit:1 compressed:1 otto:1 triangular:2 gi:1 transform:2 noisy:3 analytical:1 propose:1 product:1 relevant:1 conserving:13 achieve:3 transmission:1 asymmetry:1 perfect:1 ring:1 derive:1 jjk:2 eq:25 strong:1 involves:3 concentrate:2 closely:1 pea:1 stochastic:1 generalization:3 perfonns:2 summation:2 mathematically:1 extension:2 cost2:1 hold:1 considered:1 cognition:2 major:1 early:1 infonnation:2 minimization:5 gaussian:16 always:4 aim:2 derived:1 fur:1 sense:2 hidden:1 expand:1 sensorial:2 among:3 proposes:1 development:1 mutual:4 equal:7 construct:1 extraction:1 represents:2 unsupervised:8 minimized:1 summand:4 simultaneously:1 introduces:1 yielding:4 institut:1 theoretical:1 formalism:1 cumulants:2 cost:7 technische:1 jdy:3 probabilistic:1 together:1 deco:5 central:1 satisfied:4 choose:1 derivative:2 account:2 coefficient:1 fonnulate:1 performed:5 wether:1 tion:1 ayi:1 analyze:1 minimize:3 variance:1 yield:1 eln:1 decorrelated:8 synaptic:1 definition:1 involved:1 con:1 rational:2 logical:1 dimensionality:2 cj:3 higher:16 supervised:1 follow:1 formulation:1 done:1 atick:2 correlation:8 ally:1 ei:2 nonlinear:16 reversible:5 o:1 ujj:1 logistic:2 effect:1 concept:1 barlow:10 evolution:1 equality:2 during:4 generalized:2 bijective:2 hill:1 theoretic:2 performs:7 novel:1 recently:2 fi:1 sigmoid:1 jp:1 volume:18 extend:1 gj:1 etc:1 summands:3 perspective:1 yi:3 inverted:1 transmitted:1 minimum:2 impose:1 paradigm:4 signal:4 ii:4 reduces:3 cjj:1 equally:1 basic:3 ae:5 wilfried:5 multilayer:3 iteration:2 represent:1 achieved:1 background:1 extra:1 eliminates:1 sent:1 extracting:1 easy:1 independence:5 xj:2 architecture:23 idea:3 det:1 expression:2 pca:1 handled:1 york:1 jj:3 remark:1 factorial:8 reduced:2 generate:1 delta:1 serving:1 express:1 putting:1 four:3 redundancy:7 neither:1 excludes:1 sum:2 inverse:1 fourth:4 layer:19 ki:4 x2:1 aspect:2 fourier:3 transferred:1 munich:2 according:1 describes:1 unity:1 wi:4 taken:2 equation:3 assures:1 generalizes:1 yare:1 munchen:1 original:2 denotes:2 l6:2 hahn:1 build:1 tensor:10 strategy:1 dependence:2 diagonal:2 gradient:2 lateral:1 manifold:4 cellular:4 assuming:1 minimizing:2 implementation:1 perform:2 upper:1 neuron:8 st2:1 descent:1 anti:1 arbitrary:1 introduced:1 namely:2 nonlinearly:1 connection:3 herein:1 pattern:1 oj:1 perfonned:1 decorrelation:12 eh:1 finished:1 extract:1 nice:1 loss:9 mixed:1 interesting:1 remarkable:1 principle:1 last:3 bias:1 decorrelates:1 distributed:1 curve:3 dimension:2 calculated:1 sensory:1 preprocessing:1 mcgraw:1 mter:1 conservation:1 xi:2 don:1 expansion:9 necessarily:2 assured:1 main:1 linearly:2 whole:1 noise:1 edition:1 nothing:1 fig:7 redlich:9 fashion:2 xl:2 jacobian:4 third:2 symbol:1 gustavo:5 ci:1 diagonalization:2 karhunen:1 entropy:8 visual:1 expressed:3 contained:1 satisfies:4 extracted:2 conditional:1 goal:3 formulated:1 loosing:2 consequently:1 principal:4 called:2 bijectivity:1 l4:1 cumulant:10 ljr:1 correlated:2
7,138
902
A Neural Model of Delusions and Hallucinations in Schizophrenia Eytan Ruppin and James A. Reggia Department of Computer Science University of Maryland, College Park, MD 20742 ruppin@cs.umd .edu reggia@cs.umd.edu David Horn School of Physics and Astronomy, Tel Aviv University, Tel Aviv 69978, Israel horn@vm.tau.ac.il Abstract We implement and study a computational model of Stevens' [19921 theory of the pathogenesis of schizophrenia. This theory hypothesizes that the onset of schizophrenia is associated with reactive synaptic regeneration occurring in brain regions receiving degenerating temporal lobe projections. Concentrating on one such area, the frontal cortex, we model a frontal module as an associative memory neural network whose input synapses represent incoming temporal projections. We analyze how, in the face of weakened external input projections, compensatory strengthening of internal synaptic connections and increased noise levels can maintain memory capacities (which are generally preserved in schizophrenia) . However, These compensatory changes adversely lead to spontaneous, biased retrieval of stored memories, which corresponds to the occurrence of schizophrenic delusions and hallucinations without any apparent external trigger, and for their tendency to concentrate on just few central themes. Our results explain why these symptoms tend to wane as schizophrenia progresses, and why delayed therapeutical intervention leads to a much slower response. 150 1 Eytan Ruppin, James A. Reggia, David Hom Introduction There has been a growing interest in recent years in the use of neural models to investigate various brain pathologies and their cognitive and behavioral effects. Recent published examples of such studies include models of cortical plasticity following stroke, Alzheimer's disease and schizophrenia, and cognitive and behavioral explorations of aphasia, acquired dyslexia and affective disorders (reviewed in [1, 2]). Continuing this line of study, we present a computational account linking specific pathological synaptic changes that are postulated to occur in schizophrenia, and the emergence of schizophrenic delusions and hallucinations. The latter symptoms denote persistent, unrealistic, psychotic thoughts (delusions) or percepts (hallucinations) that may at times flood the patient in an overwhelming, stressful manner. The wealth of data gathered concerning the pathophysiology of schizophrenia supports the involvement of both the frontal and the temporal lobes. On the one hand, there are atrophic changes in the hippocampus and parahippocampal areas including neuronal loss and gliosis. On the other hand, neurochemical and morphometric studies testify to an expansion of various receptor binding sites and increased dendritic branching in the frontal cortex of schizophrenics. Stevens has recently presented a theory linking these temporal and frontal findings, claiming that the onset of schizophrenia is associated with reactive anomalous sprouting and synaptic reorganization taking place in the projection sites of degenerating temporal neurons, including (among various cortical and subcortical structures) the frontal lobes [3]. This paper presents a computational study of Stevens' theory. Within the framework of a memory model of hippocampal-frontal interaction, we show that the introduction of the 'microscopic' synaptic changes that underlie Stevens' hypothesis can help preserve memory function but results in specific 'pathological' changes in the 'macroscopic' behavior of the network. A small subset of the patterns stored in the network are now spontaneously retrieved at times, without being cued by any specific input pattern. This emergent behavior shares some of the important characteristics of schizophrenic delusions and hallucinations, which frequently appear in the absence of any apparent external trigger, and tend to concentrate on a limited set of recurrent themes [4]. Memory capacities are fairly preserved in schizophrenics, until late stages of the disease [5]. In Section 2 we present our model. The analytical and numerical results obtained are described in Section 3, followed by our conclusions in Section 4. 2 The Model As illustrated in Figure 1, we model a frontal module as an associative memory attractor neural network, receiving its input memory cues from decaying external input fibers (representing the degenerating temporal projections). The network's internal connections, which store the memorized patterns, undergo synaptic strengthening changes that model the reactive synaptic regeneration within the frontal module. The effect of other diffuse external projections is modeled as background noise. A frontal module represents a macro-columnar unit that has been suggested as a basic functional building block of the neocortex [6]. The assumption that memory retrieval from the frontal cortex is invoked by the firing of incoming A Neural Model of Delusions and Hallucinations in Schizophrenia 151 temporal projections is based on the notion that temporal structures have an important role in establishing long-term memory in the neocortex and in the retrieval of facts and events (e.g., [7]). ,,- --- " Other cortical : modules I T I I ....... : (Influence nxxJeled as,' .... .. ! noise) ," ......... .. \ ,I" .... .. Figure 1: A schematic illustration of the model. A frontal module is modeled as an attractor neural network whose neurons receive inputs via three kinds of connections: internal connections from other frontal neurons, external connections from temporal lobe neurons, and diffuse external connections from other cortical modules, modeled as noise. The attractor network we use is a biologically-motivated variant of Hopfield's ANN model, proposed by Tsodyks & Feigel'man [8]. Each neuron i is described by a binary variable Si = {1,0} denoting an active (firing) or passive (quiescent) state, respectively. M = aN distributed memory patterns IJ are stored in the network. The elements of each memory pattern are chosen to be 1 (0) with probability p (1- p) respectively, with p ~ 1. All N neurons in the network have a fixed uniform threshold O. e In its initial, undamaged state, the weights of the internal synaptic connections are M Wij =N CO" L....t(e IJ i - p)(e j - p) , (1) IJ=I where Co = 1. The post-synaptic potential (input field) hi of neuron i is the sum of internal contributions from other neurons and external projections Fi e hi(t) = L WijSj(t - 1) + Fie. (2) j The updating rule for neuron i at time t is given by S .(t) = 1 {I,0, with p~ob . G(hi(t) - 0) otherwIse (3) 152 Eytan Ruppin, James A. Reggia, David Hom where G is the sigmoid function G(x) = 1/(1+exp( -x/T)), and T denotes the noise level. The activation level of the stored memories is measured by their overlaps mil with the current state of the network, defined by 1 mll(t) = N (1 _ )N P P L)er - p)Si(t) . (4) i=l Stimulus-dependent retrieval is modeled by orienting the field Fe with one of the memorized patterns (the cued pattern, say e), such that F/ = e . e\ , (e > 0) . (5) Following the presentation of an external input cue, the network state evolves until it converges to a stable state. The network parameters are tuned such that in its initial, undamaged state it correctly retrieves the cued patterns (eo = 0.035 , Co = 1, T = 0.005). We also examine the network's behavior in the absence of any specific stimulus. The network may either continue to wander around in a state of random low baseline activity, or it may converge onto a stored memory state. We refer to the latter process as spontaneous retrieval. Our investigation of Stevens' work proceeds in two stages. First we examine and analyze the behavior of the network when it undergoes uniform synaptic changes that represent the pathological changes occurring in accordance with Stevens' theory. These include the weakening of external input projections (e !) and the increase in the internal projections (c i) and noise levels (T i). In the second stage, we add the assumption that the internal synaptic compensatory changes have an additional Hebbian activity-dependent component, and examine the effect of the rule (6) where Sk is 1 (0) only if neuron k has been consecutively firing (quiescent) for the last r iterations, and 'Y is a constant. 3 Results We now show some simulation and analytic results, examining the effects of the 'microscopic' pathological changes, taking place in accordance with Stevens' theory, on the 'macroscopic' behavior of the network . The analytical results presented have been derived by calculating the magnitude of randomly formed initial 'biases', and comparing their effect on the network 's dynamics versus the effect of externally presented input cues. This comparison is performed by formulating a corresponding overlap master equation , whose fixed point dynamics are investigated via phaseplane analysis, as described in [9]. First, we study whether the reactive synaptic changes (occurring in both internal and external, diffuse synapses) are really compensatory, i.e., to what extent can they help maintaining memory capacities in the face of degenerating external input synapses. As illustrated in Figure 2, we find that increased noise levels can (up to some degree) preserve memory retrieval in the face of decreased external input strength. Increased synaptic strengthening preserves 153 A Neural Model of Delusions and Hallucinations in Schizophrenia (a) (b) 1.0 1.0 ,/ i' , - . _ ,! O.O35 ------ ?? iO.025 O.B ---- ?? '0.015 - -? ? 10.014 G---o ?? ; 0.013 <)--0._ 10.012 - - - ?? ; 0.005 O.B 0.6 0.8 . t ! 6 6 0.( . ~_.__ ? _ 0.025 - --- ? ? 0.015 - - - ? ? 0.005 '- -- 0.2 0.2 0.0 0.000 0.( : - . _ 0.035 I - :... ---- 0.010 T 0.020 0?8. 0.020 T Figure 2: Stimulus-dependent retrieval performance, measured by the average final overlap m, as a function of the noise level T. Each curve displays this relation at a different magnitude of external input projections e. (a) Simulation results. (b) Analytic approximation. memory retrieval in a similar manner, and the combined effect of these synaptic compensatory measures is synergistic. Second, although the compensatory synaptic changes help maintain memory retrieval capacities, they necessarily have adverse effects, leading eventually to the emergence of spontaneous activation of non-cued memory patterns; the network converges to some of its memory patterns in a pathological, autonomous manner , in the absence of any external input stimuli. This emergence of pathological spontaneous retrieval , when either the noise level or the internal synaptic strength (or both) are increased beyond some point, is demonstrated in Figure 3. Third, when the compensatory regeneration of internal synapses has an additional Hebbian component (representing a period of increased activity-dependent plasticity due to the regenerative synaptic changes), a biased spontaneous retrieval distribution is obtained. That is, as time evolves (measured in time units of 'trials'), the distribution of patterns spontaneously retrieved by the network in a pathological manner tends to concentrate only on one or two of all the memory patterns stored in the network, as is shown in Figure 4a. This highly peaked distribution is maintained for a few hundred additional trials until memory retrieval sharply collapses to zero as a global mixed-state attractor is formed. Such a mixed attractor state does not have very high overlap with any memorized pattern, and thus does not represent any well-defined cognitive or perceptual item. It is an end state of the Hebbian, activity-dependent evolution of the network. Yet, even after activitydependent changes ensue , if spontaneous activity does not emerge the distribution of retrieved memories remains homogeneous (see Figure 4b). Eventually, a global 154 Eytan Ruppin. James A. Reggia. David Hom (b) (a) 1.0 , ,--- - . 1.0 , ........... . ~ .. .. .. ,1' " , , O.B , O.B , 0.6 0.6 , t ! f AnalytIc _Imollon - - - - Slmulallon 0.4 0.4 0.2 0.2 ,, 0.0 0.005 ' , 0.010 0.015 0.020 0.0 ':--~-'-:-'-::--~~--~----:' 2.0 2.5 3.0 3.5 4.0 T Figure 3: (a) Spontaneous retrieval, measured as the highest final overlap m achieved with any of the stored memory patterns, displayed as a function of the noise level T. c = 1. (b) Spontaneous retrieval as a function of internal synaptic compensation factor c. T 0.009. = mixed-state attractor is formed, and the network looses its retrieval capacities, but during this process no memory pattern gets to dominate the retrieval output. Our results remain qualitatively similar even when bounds are placed on the absolute magnitude of the synaptic weights. 4 Conclusions Our results suggest that the formation of biased spontaneous retrieval requires the concomitant occurrence of both degenerative changes in the external input fibers, and regenerative Hebbian changes in the intra-modular synaptic connections. They add support to the plausibility of Stevens' theory by showing that it may be realized within a neural model, and account for a few characteristics of schizophrenic symptoms: ? The emergence of spontaneous, non-homogeneous retrieval is a self-limiting phenomenon (as eventually a cognitively meaningless global attractor is formed) - this parallels the clinical finding that as schizophrenia progresses both delusions and hallucinations tend to wane, while negative symptoms are enhanced [10]. ? Once converged to, the network has a much larger tendency to remain in a biased memory state than in a non biased one - this is in accordance with the persistent characteristic of schizophrenic florid symptoms. ? As more spontaneous retrieval trials occur the frequency of spontaneous retrieval increases - indeed, while early treatment in young psychotic adults A Neural Model of Delusions and Hallucinations in Schizophrenia 155 (b) (a) 1.0 ~-~-~~-:----~-----. 0.100 0.8 0.080 G---f] ~ 0.6 t - v Aft... 800triala Q .,. 1 ~ ? ( ? Aft ... 400tr1a1a ~ 0.4 >' ., .,~ ~. ; , ~ 1 0.000 t 1 l .~ 0.2 ~AftOf200trlals Aft.r 200 tria_ {3 - -(:~Aft.r500trials ~ ~----"-~~--~-----. 0.040 0.020 Figure 4: (a) The distribution of memory patterns spontaneously retrieved . The xaxis enumerates the memories stored, and the y-axis denotes the retrieval frequency of each memory. 'Y = 0.0025. (b) The distribution of stimulus-dependent retrieval of memories. 'Y = 0.0025 . leads to early response within days, late, delayed intervention leads to a much slower response during one or more months [11]. The current model generates some testable predictions: ? On the neuroanatomical level, the model can be tested quantitatively by searching for a positive correlation between a recent history of florid psychotic symptoms and postmortem neuropathological findings of synaptic compensation. (For example, this kind of correlation, between indices of synaptic area and cognitive functioning was found in Alzheimer patients [12]). ? On the physiological level, the increased compensatory noise should manifest itself in increased spontaneous neural activity. While this prediction is obviously difficult to examine directly, EEG studies in schizophrenics show significant increase in slow-wave delta activity which may reflect increased spontaneous activity [13]. ? On the clinical level, due to the formation of a large and deep basin of attraction around the memory pattern which is at the focus of spontaneous retrieval, the proposed model predicts that its retrieval (and the elucidation of the corresponding delusions or hallucinations) may be frequently triggered by various environmental cues. A recent study points in this direction [14]. 156 Eytan Ruppin, James A. Reggia, David Hom Acknowledgements This research has been supported by a Rothschild Fellowship to Dr. Ruppin. References [1] J. Reggia, R. Berndt, and L. D' Autrechy. Connectionist models in neuropsychology. In Handbook of Neuropsychology, volume 9. 1994, in press. [2] E. Ruppin. Neural modeling of psychiatric disorders. Network: Computation in Neural Systems, 1995. Invited review paper, to appear. [3] J .R. Stevens. Abnormal reinnervation as a basis for schizophrenia: A hypothesis. Arch. Gen . Psychiatry, 49 :238-243, 1992. [4] S.K. Chaturvedi and V.D. Sinha. Recurrence of hallucinations in consecutive episodes of schizophrenia and affective disorder. Schizophrenia Research, 3: 103106, 1990. [5] M. Marsel Mesulam. Schizophrenia and the brain. New England Journal of Medicine, 322(12) :842-845, 1990. [6] P.S. Goldman and W.J .H . Nauta. Columnar distribution of cortica-cortical fibers in the frontal, association , limbic and motor cortex of the developing rhesus monkey. Brain Res., 122:393-413,1977. [7] L. R. Squire. Memory and the hippocampus: A synthesis from findings with rats, monkeys, and humans . Psychological Review, 99 :195-231 , 1992. [8] M.V. Tsodyks and M.V. Feigel'man. The enhanced storage capacity in neural networks with low activity level. Europhys. Lett., 6:101 - 105, 1988. [9] D. Horn and E. Ruppin . Synaptic compensation in attractor neural networks: Modeling neuropathological findings in schizophrenia. Neural Computation, page To appear, 1994. [10] W .T. Carpenter and R.W. Buchanan. Schizophrenia. New England Journal of Medicine, 330:10, 1994. [11] P. Seeman. Schizophrenia as a brain disease: The dopamine receptor story. Arch . Neurol., 50:1093-1095, 1993. [12] S. T. DeKosky and S.W. Scheff. Synapse loss in frontal cortex biopsies in alzheimer 's disease: Correlation with cognitive severity. Ann. Neurology, 27(5):457-464, 1990. [13] Y. Jin, S.G. Potkin, D. Rice, and J. Sramek et. al. Abnormal EEG responses to photic stimulation in schizophrenic patients. Schizophrenia Bulletin, 16(4):627634, 1990. [14] R.E. Hoffman and J .A. Rapaport. A psycholoinguistic study of auditory /verbal hallucinations: Preliminary findings. In David A. and Cutting J. , editors, The Neuropsychology of Schizophrenia. Erlbaum, 1993.
902 |@word trial:3 hippocampus:2 rhesus:1 simulation:2 lobe:4 initial:3 denoting:1 tuned:1 current:2 comparing:1 si:2 activation:2 yet:1 aft:4 numerical:1 plasticity:2 analytic:3 motor:1 cue:4 item:1 wijsj:1 berndt:1 persistent:2 behavioral:2 affective:2 buchanan:1 manner:4 acquired:1 indeed:1 behavior:5 frequently:2 growing:1 examine:4 brain:5 goldman:1 overwhelming:1 feigel:2 israel:1 what:1 kind:2 monkey:2 loos:1 astronomy:1 finding:6 temporal:9 unit:2 underlie:1 intervention:2 appear:3 positive:1 accordance:3 tends:1 io:1 receptor:2 establishing:1 firing:3 weakened:1 co:3 limited:1 collapse:1 horn:3 spontaneously:3 block:1 implement:1 area:3 thought:1 projection:11 suggest:1 psychiatric:1 get:1 onto:1 synergistic:1 parahippocampal:1 storage:1 influence:1 demonstrated:1 disorder:3 rule:2 undamaged:2 attraction:1 dominate:1 searching:1 notion:1 autonomous:1 limiting:1 spontaneous:15 trigger:2 enhanced:2 homogeneous:2 hypothesis:2 element:1 updating:1 postmortem:1 predicts:1 role:1 module:7 tsodyks:2 region:1 episode:1 highest:1 neuropsychology:3 disease:4 limbic:1 dynamic:2 basis:1 hopfield:1 emergent:1 various:4 fiber:3 retrieves:1 formation:2 europhys:1 whose:3 apparent:2 modular:1 larger:1 say:1 regeneration:3 otherwise:1 flood:1 emergence:4 itself:1 final:2 associative:2 obviously:1 triggered:1 analytical:2 interaction:1 strengthening:3 macro:1 chaturvedi:1 gen:1 converges:2 help:3 cued:4 recurrent:1 ac:1 measured:4 ij:3 school:1 progress:2 c:2 concentrate:3 direction:1 biopsy:1 stevens:9 consecutively:1 exploration:1 human:1 memorized:3 really:1 investigation:1 preliminary:1 dendritic:1 around:2 exp:1 early:2 consecutive:1 hoffman:1 mil:1 derived:1 focus:1 degenerative:1 psychiatry:1 rothschild:1 baseline:1 dependent:6 weakening:1 relation:1 wij:1 among:1 fairly:1 neuropathological:2 field:2 once:1 represents:1 park:1 peaked:1 connectionist:1 stimulus:5 quantitatively:1 few:3 pathological:7 randomly:1 preserve:3 delayed:2 mll:1 cognitively:1 attractor:8 maintain:2 testify:1 interest:1 investigate:1 highly:1 intra:1 hallucination:12 xaxis:1 continuing:1 re:1 sinha:1 psychological:1 increased:9 modeling:2 subset:1 uniform:2 hundred:1 examining:1 erlbaum:1 stored:8 combined:1 physic:1 vm:1 receiving:2 synthesis:1 central:1 reflect:1 dr:1 external:16 adversely:1 cognitive:5 leading:1 account:2 potential:1 hypothesizes:1 postulated:1 squire:1 onset:2 performed:1 analyze:2 wave:1 decaying:1 parallel:1 contribution:1 il:1 formed:4 characteristic:3 percept:1 ensue:1 gathered:1 published:1 converged:1 stroke:1 history:1 explain:1 synapsis:4 synaptic:23 frequency:2 james:5 associated:2 auditory:1 treatment:1 concentrating:1 stressful:1 enumerates:1 manifest:1 day:1 response:4 synapse:1 symptom:6 just:1 stage:3 arch:2 until:3 correlation:3 hand:2 undergoes:1 aviv:2 orienting:1 building:1 effect:8 functioning:1 evolution:1 illustrated:2 during:2 branching:1 degenerating:4 self:1 sprouting:1 maintained:1 recurrence:1 rat:1 hippocampal:1 passive:1 ruppin:9 invoked:1 recently:1 fi:1 sigmoid:1 functional:1 stimulation:1 volume:1 linking:2 association:1 refer:1 significant:1 pathology:1 delusion:10 stable:1 cortex:5 add:2 recent:4 retrieved:4 involvement:1 store:1 binary:1 continue:1 additional:3 eo:1 converge:1 period:1 hebbian:4 fie:1 england:2 plausibility:1 clinical:2 long:1 retrieval:24 concerning:1 post:1 schizophrenia:22 regenerative:2 schematic:1 prediction:2 variant:1 anomalous:1 basic:1 patient:3 dopamine:1 iteration:1 represent:3 achieved:1 receive:1 preserved:2 fellowship:1 background:1 decreased:1 wealth:1 macroscopic:2 biased:5 meaningless:1 umd:2 invited:1 tend:3 undergo:1 alzheimer:3 whether:1 motivated:1 deep:1 generally:1 neocortex:2 delta:1 correctly:1 threshold:1 year:1 sum:1 master:1 place:2 ob:1 bound:1 hi:3 hom:4 followed:1 abnormal:2 display:1 activity:9 strength:2 occur:2 sharply:1 diffuse:3 generates:1 formulating:1 department:1 developing:1 remain:2 evolves:2 biologically:1 equation:1 remains:1 eventually:3 aphasia:1 end:1 schizophrenic:9 reggia:7 occurrence:2 dyslexia:1 slower:2 neuroanatomical:1 denotes:2 include:2 maintaining:1 elucidation:1 calculating:1 medicine:2 testable:1 realized:1 md:1 microscopic:2 maryland:1 capacity:6 extent:1 reorganization:1 modeled:4 index:1 illustration:1 concomitant:1 reinnervation:1 difficult:1 fe:1 claiming:1 negative:1 neuron:10 compensation:3 jin:1 displayed:1 severity:1 david:6 compensatory:8 connection:8 photic:1 seeman:1 adult:1 beyond:1 suggested:1 proceeds:1 pattern:17 including:2 tau:1 memory:32 unrealistic:1 event:1 overlap:5 representing:2 axis:1 review:2 acknowledgement:1 wander:1 loss:2 mixed:3 subcortical:1 versus:1 rapaport:1 degree:1 basin:1 morphometric:1 story:1 editor:1 share:1 placed:1 last:1 supported:1 verbal:1 bias:1 face:3 taking:2 emerge:1 absolute:1 bulletin:1 distributed:1 curve:1 lett:1 cortical:5 qualitatively:1 scheff:1 cutting:1 global:3 active:1 incoming:2 handbook:1 quiescent:2 neurology:1 sk:1 why:2 reviewed:1 activitydependent:1 tel:2 eeg:2 pathogenesis:1 expansion:1 investigated:1 necessarily:1 noise:11 carpenter:1 neuronal:1 site:2 slow:1 theme:2 perceptual:1 atrophic:1 late:2 third:1 young:1 externally:1 specific:4 showing:1 er:1 neurol:1 physiological:1 magnitude:3 occurring:3 columnar:2 pathophysiology:1 binding:1 corresponds:1 environmental:1 wane:2 rice:1 month:1 presentation:1 ann:2 absence:3 man:2 change:16 adverse:1 eytan:5 tendency:2 college:1 internal:11 support:2 latter:2 reactive:4 frontal:15 tested:1 phenomenon:1
7,139
903
A Silicon Axon Bradley A. Minch, Paul Hasler, Chris Diorio, Carver Mead Physics of Computation Laboratory California Institute of Technology Pasadena, CA 91125 bminch, paul, chris, carver@pcmp.caltech.edu Abstract We present a silicon model of an axon which shows promise as a building block for pulse-based neural computations involving correlations of pulses across both space and time. The circuit shares a number of features with its biological counterpart including an excitation threshold, a brief refractory period after pulse completion, pulse amplitude restoration , and pulse width restoration. We provide a simple explanation of circuit operation and present data from a chip fabricated in a standard 2Jlm CMOS process through the MOS Implementation Service (MOSIS). We emphasize the necessity of the restoration of the width of the pulse in time for stable propagation in axons. 1 INTRODUCTION It is well known that axons are neural processes specialized for transmitting information over relatively long distances in the nervous system. Impulsive electrical disturbances known as action potentials are normally initiated near the cell body of a neuron when the voltage across the cell membrane crosses a threshold. These pulses are then propagated with a fairly stereotypical shape at a more or less constant velocity down the length of the axon . Consequently, axons excel at precisely preserving the relative timing of threshold crossing events but do not preserve any of the initial signal shape. Information , then , is presumably encoded in the relative timing of action potentials. 740 Bradley A. Minch, Paul Hasler, Chris Diorio, Carver Mead The biophysical mechanisms underlying the initiation and propagation of action potentials in axons have been well studied since the seminal work of Hodgkin and Huxley on the giant axon of Loligo. (Hodgkin & Huxley, 1952) Briefly, when the voltage across a small patch of the cell membrane increases to a certain level, a population of ion channels permeable to sodium opens, allowing an influx of sodium ions, which, in turn, causes the membrane voltage to increase further and a pulse to be initiated. This population of channels rapidly inactivates, preventing the passage of additional ions. Another population of channels permeable to potassium opens after a brief delay causing an efflux of potassium ions, restoring the membrane to a more negative potential and terminating the pulse. This cycle of ion migration is coupled to neighboring sections of the axon, causing the action potential to propagate. The sodium channels remain inactivated for a brief interval of time during which the affected patch of membrane will not be able to support another action potential. This period of time is known as the refractory period. The axon circuit which we present in this paper does not attempt to model the detailed dynamics of the various populations of ion channels, although such detailed neuromimes are both possible (Lewis, 1968; Mahowald & Douglas, 1991) and useful for learning about natural neural systems. Nonetheless, it shares a number of important features with its biological counterpart including having a threshold for excitation and a refractory period. It is well accepted that the amplitude of the action potential must be restored as it propagates. It is not as universally understood is that the width of the action potential must be restored in time if it is to propagate over any appreciable distance. Otherwise, the pulse would smear out in time resulting in a loss of precise timing information, or it would shrink down to nothing and cease to propagate altogether. In biological axons, restoration of the pulse width is accomplished through the dynamics of sodium channel inactivation and potassium channel activation. In our silicon model, the pulse width is restored through feedback from the successive stage. This feedback provides an inactivation which is similar to that of the sodium channels in biological axons and is also the underlying cause of refractoriness in our circuit . In the following section we provide a simple description of how the circuit behaves. Following this, data from a chip fabricated in a standard 2p.m CMOS process through MOSIS are presented and discussed. 2 THE SILICON AXON CmCUIT An axon circuit which is to be used as a building block in large-scale computational systems should be made as simple and low-power as possible , since it would be replicated many times in any such system. Each stage of the axon circuit described below consists of five transistors and two small capacitors, making the axon circuit very compact. The axon circuit uses the delay through a stage to time the signal which is fed back to restore the pulse width, thus avoiding the need for an additional delay circuit for each section. Additonally, the circuit operates with low power; during typical operation (a pulse of width 2ms travelling at 10 3 stages/s), pulse propagation costs about 4pJ / stage of energy. Under these circumstances, the circuit consumes about 2nW/stage of static power. A Silicon Axon 741 Figure 1: Three sections of the axon circuit. Three stages of the axon circuit are depicted in Figure 1. A single stage consists of two capacitors and what would be considered a pseudo-nMOS NAND gate and a pseudo-nMOS inverter in digital logic design. These simple circuits are characterized by a threshold voltage for switching and a slew rate for recharging. Consider the inverter circuit. If the input is held low for a sufficiently long time, the pull-up transistor will have charged the output voltage almost completely to the positive rail. If the input voltage is ramped up toward the positive rail, the current in the pull-down transistor will increase rapidly. At some input voltage level, the current in the pull-down transistor will equal the saturation current of the pull-up transistor; this voltage is known as the threshold. The output voltage will begin to discharge at a rapidly increasing rate as the input voltage is increased further . After a very short time, the output will have discharged almost all the way to the negative rail. Now, if the input were decreased rapidly, the output voltage would ramp linearly in time (slew) up toward the positive rail at a rate set by the saturation current in the pull-up transistor and the capacitor on the output node . The NAND gate is similar except both inputs must be (roughly speaking) above the threshold in order for the output to go low. If either input goes low , the output will charge toward the positive rail. Note also that if one input of the NAND gate is held high, the circuit behaves exactly as an inverter. The axon circuit is formed by cascading multiple copies of this simple five transistor circuit in series. Let the voltage on the first capacitor of the nth stage be denoted by Un and the voltage on the second capacitor by v n . Note that there is feedback from U n +1 to the lower input of the NAND gate of the nth stage. Under quiescent conditions, the input to the first stage is low (at the negative rail) , the U nodes of all stages are high (at the positive rail), and all of the v nodes are held low (at the negative rail). The feedback signal to the final stage in the line would be tied to the positive rail. The level of the bias voltages 71 and 72 determine whether or not a narrow pulse fed into the input of the first stage will propagate and, if so, the width and velocity with which it does. In order to obtain a semi-quantitative understanding of how the axon circuit behaves, we will first consider the dynamics of a cascade of simple inverters (three sections of which are depicted in Figure 2) and then consider the addition of feedback. Under most circumstances, discharges will occur on a much faster time scale Bradley A. Minch, Paul Hasler, Chris Diorio, Carver Mead 742 than the recharges, so we make the simplifying assumption that when the input of an inverter reaches the threshold voltage, the output discharges instantaneously. Additionally, we assume that saturated transistors behave as ideal current sources (i.e., we neglect the Early effect) so that the recharges are linear ramps in time. TTT TTT TTT TTT TTT TTT Figure 2: A cascade of pseudo-nMOS inverters. Let hand h be the saturation currents in the pull-up transistors with bias voltages and 1"2, respectively. Let 6 1 and 62 be the threshold voltages of the first and second inverters in a single stage, respectively. Also, let u = IdC and v = I2/C be the slew rates for the U and v nodes, respectively. Let a 1 = 6dv and a 2 = 6 2 /u be the time required for Un to charge from the negative rail up to E>1 and for Vn to charge from the negative rail up to 6 2 , respectively. Finally, let v~ denote the peak value attained by the v signal of the nth stage. 1"1 V? v?71 vn IE ~IE 671 IE ~I a2 (a) II > h Figure 3: Geometry of the idealized ~IE 671 Un and Vn ~I a2 (b) h < 12 signals under the bias conditions (a) h > 12 and (b) II < 12, Consider what would happen if V n - l exceeded 6 1 for a time 671 , In this case, Un would be held low for 671 and then released. Meanwhile , Vn would ramp up to v6 n . Then , Un would begin to charge toward the positive rail while Vn continues to charge. This continues for a time a2 at which point Un will have reached 6 2 causing A Silicon Axon 743 to be discharged to the negative rail. Now, Un+l is held low while Vn exceeds 8 1 this interval of time is precisely On+!. Figures 3a and 3b depict the geometry of the Un and Vn signals in this scenario under the bias conditions h > 12 and h < 12, respectively. Simple Euclidean geometry implies that the evolution of On will be governed by the first-order difference equation Vn which is trivially solved by Thus, the quantity a2-al determines what happens to the width of the pulse as it propagates. In the event that a 2 < aI, the pulse will shrink down to nothing from its initial width. If a 2 > aI, the pulse width will grow without bound from its initial width. The pulse width is preserved only if a 2 = a l . This last case, however, is unrealistic. There will always be component mismatches (with both systematic and random parts), which will cause the width of the pulse to grow and shrink as it propagates down the line, perhaps cancelling on average. Any systematic offsets will cause the pulse to shrink to nothing or to grow without bound as it propagates. In any event, information about the detailed timing of the initial pulse will have been completely lost. Now, consider the action of the feedback in the axon circuit (Figure 1). If Un were to be held low for a time longer than a l (i.e., the time it takes Vn to charge up to 8d, Un+1 would come back and release Un, regardless of the state of the input. Thus, the feedback enforces the condition On ::; a l . If 11 > h (i.e., a 2 < ad, a pulse whose initial width is larger than a 1 will be clipped to a 1 and then shrink down to nothing and disappear. In the event that II < h (i.e., a 2 > ad, a pulse whose initial duration is too small will grow up until its width is limited by the feedback. The axon circuit normally operates under the latter bias condition. The dynamics of the simple inverter chain cause a pulse which is to narrow to grow and the feedback loop serves to limit the pulse width; thus, the width of the pulse is restored in time . The feedback is also the source of the refractoriness in the axon; that is, until U n +l charges up to (roughly) 8 1 , V n -l can have no effect on Un. 3 EXPERIMENTAL DATA In this section, data from a twenty-five stage axon will be shown. The chip was fabricated in a standard 2f.lm p-well (Orbit) CMOS process through MOSIS. Uniform Axon A full space-time picture of pulse (taken at the v nodes of the circuit) propagation down a uniform axon is depicted on the left in Figure 4. The graph on the right in Figure 4 shows the same data from a different perspective. The lower sloped curve represents the time of the initial rapid discharge of the U node at each successive stage-this time marks the leading edge of the pulse taken at the v node of that stage. 744 Bradley A. Minch. Paul Hasler. Chris Diorio. Carver Mead The upper sloped curve marks the time of the final rapid discharge of the v node of each stage-this time is the end of the pulse taken at the v node of that stage. The propagation velocity of the pulse is given (in units of stages/s) by the reciprocal of the slope of the lower inclined curve. The third curve is the difference of the other two and represents the pulse width as a function of position along the axon. The graph on the left of Figure 5 shows propagation velocity as a function of the 72 bias voltage- so long as the pulse propagates, the velocity is nearly independent of 71. Two orders of magnitude of velocity are shown in the plot; these are especially well matched to the time scales of motion in auditory and visual sensory data. The circuit is tunable over a much wider range of velocities (from about one stage per second to well in excess of 10 4 stages/ s). The graph on the right of Figure 5 shows pulse width as a function of 71 for various values of 72-the pulse width is mainly determined by 71 with 72 setting a lower limit. Tapered Axon In biological axons, the propagation velocity of an action potential is related to the diameter of the axon- the bigger the diameter, the greater the velocity. If the axon were tapered, the velocity of the action potential would change as it propagated. If the bias transistors in the axon circuit are operated in their subthreshold region, the effect of an exponentially tapered axon can be simulated by applying a small voltage difference to the ends of each ofthe 71 and 72 bias lines. (Lyon & Mead , 1989) These narrow wires are made with a relatively resistive layer (polysilicon); hence, putting a voltage difference across the ends will linearly interpolate the bias voltages for each stage along the line. In subthreshold, the bias currents are exponentially related to the bias voltages. Since the pulse width and velocity are related to the bias currents , we expect that a pulse will either speed up and get narrower or slow down and get wider (depending on the sign of the applied voltage) exponentially as a function of position along the line. The graph on the left of Figure 6 depicts the boundaries of a pulse as it propagates along of the axon circuit for a positive (*'s) and negative (x 's) voltage difference applied to the 7 lines. The graph on the right of Figure 6 shows the corresponding pulse width for each applied voltage difference. Note that in each case, the width changes by more than an order of magnitude, but the pulse maintains its integrity. That is, the pulse does not disappear nor does it split into multiple pulses-this behavior would not be possible if the pulse width were not restored in time. 4 CONCLUSIONS In this paper we have presented a low-power, compact axon circuit, explained its operation , and presented data from a working chip fabricated through MOSIS. The circuit shares a number of features with its biological counterpart including an excitation threshold , a brief refractory period after pulse completion, pulse amplitude restoration , and pulse width restoration. It is tunable over orders of magnitude in pulse propagation velocity-including those well matched to the time scales of auditory and visual signals- and shows promise for use in synthetic neural systems which perform computations by correlating events which occur over both space and time such as those presented in (Horiuchi et ai, 1991) and (Lazzaro & Mead, 1989). A Silicon Axon 745 Acknowledgements This material is based upon work supported in part under a National Science Foundation Graduate Research Fellowship, the Office of Naval Research, DARPA, and the Beckman Foundation. References A. 1. Hodgkin and A. K. Huxley, (1952). A Quantitative Description of Membrane Current and its Application to Conduction and Excitation in Nerve. Journal of Physiology, 117:6, 500-544. T. Horiuchi, J. Lazzaro , A. Moore , and C. Koch, (1991) . A Delay-Line Based Motion Detection Chip. Advances in Neural Information Processing Systems 3. San Mateo, CA: Morgan Kaufmann Publishers, Inc . 406-412 . J. Lazzaro and C. Mead, (1989). A Silicon Model of Auditory Localization. Neural Computation, 1:1 , 47-57. R . Lyon and C. Mead, (1989). Electronic Cochlea. Analog VLSI and Neural Systems. Reading, MA: Addison-Wesley Publishing Company, Inc. 279-302. E. R. Lewis , (1968). Using Electronic Circuits to Model Simple Neuroelectric Interactions. Proceedings of the IEEE, 56:6, 931-949. M. Mahowald and R. Douglas, (1991). A Silicon Neuron. Nature , 354:19 , 515-518. o.o35r-----.,-----.,--------, 0.03 2 0.025 1.5 ~ ., .& '0 0 .02 ~ f= 0.015 0 .5 > 0 -0 . 5 30 0.06 Stage o 0 TilTIe (s) ?0~--~1~0---~2~0---~30 Stage Figure 4: Pulse propagation along a uniform axon. (Left) Perspective view. (Right) Overhead view . *: pulse boundaries , x: pulse width. T1 = 0.720V, T2 = 0.780V . Velocity = 1, 100stages/s , Width = 3.8ms Bradley A. Minch, Paul Hasler, Chris Diorio, Carver Mead 746 10" r----~--~--~--____, II< o o "" """" "" """" 00 II< 'ttl" 3: "" ..."""" Tau2 = 0.700 V ~Xx 10.2 II< II< ~ "" 0 .740 Tau2 = 0.7 0 V ~o T a u2 = ~, 10-3 """" = +d> l!l ? II< Tau2 ~+ ~ II< ~ Tau2 = 0 .660 V ~ 110 11<... .820 V Ta\lp = 0 .860 V 10~~.5-----0~.6-----0~.~7-----0~.8~---0~.9 10~~ . 6-------0~.~7-------0~ . 8-------0~. 9 Tau2 (V) Taul (V) Figure 5: Uniform axon. (Left) Pulse velocity as a function of r2. (Right) Pulse width as a function of rl for various values of r2. 10" r------~--------~------___, 0 . 14r-------~-------~----~ 0 . 12 )0( II< 0 .1 il:! 0 .08 E::: II< )0( )O()O( x 0.06 x )0( )0( x )0( )0( )0( x 0.0 4#. . x )0( .... )0( )0( x )O()O( )0( )0( ><"XxX )0()0( X X X 1 O?3~------,'---------_'_------__=' 0 . 020~------~1~0~------2~0~------370 Stage o 10 20 30 Stage Figure 6: (Left) Pulse propagation along a tapered axon. (Right) Pulse width as 0.770V, r;ight 0.600V , a function of position along a tapered axon. *: ri eft eft r~eft 0.B20V , r;ight 0.650. x : ri 0.600V , r;ight = 0.770V, r~eft = 0.650V , r;ight = 0.B20V. = = = = =
903 |@word briefly:1 open:2 pulse:55 propagate:4 simplifying:1 initial:7 necessity:1 series:1 bradley:5 current:9 activation:1 additonally:1 must:3 happen:1 shape:2 plot:1 depict:1 nervous:1 reciprocal:1 short:1 provides:1 node:9 neuromimes:1 successive:2 five:3 along:7 consists:2 ramped:1 resistive:1 overhead:1 rapid:2 behavior:1 roughly:2 nor:1 company:1 lyon:2 increasing:1 begin:2 xx:1 underlying:2 matched:2 circuit:29 what:3 ttl:1 giant:1 fabricated:4 jlm:1 pseudo:3 quantitative:2 charge:7 exactly:1 normally:2 unit:1 positive:8 service:1 understood:1 timing:4 t1:1 limit:2 switching:1 mead:9 initiated:2 permeable:2 studied:1 inactivates:1 mateo:1 limited:1 range:1 graduate:1 enforces:1 restoring:1 lost:1 block:2 cascade:2 physiology:1 get:2 applying:1 seminal:1 charged:1 cmcuit:1 go:2 regardless:1 duration:1 recharges:2 stereotypical:1 cascading:1 pull:6 population:4 discharge:5 us:1 velocity:14 crossing:1 continues:2 electrical:1 solved:1 region:1 inclined:1 cycle:1 diorio:5 consumes:1 efflux:1 dynamic:4 terminating:1 upon:1 localization:1 completely:2 darpa:1 chip:5 various:3 horiuchi:2 whose:2 encoded:1 larger:1 ramp:3 otherwise:1 final:2 biophysical:1 transistor:10 interaction:1 cancelling:1 causing:3 neighboring:1 loop:1 rapidly:4 description:2 potassium:3 cmos:3 wider:2 depending:1 completion:2 implies:1 come:1 material:1 biological:6 sufficiently:1 considered:1 koch:1 presumably:1 nw:1 mo:1 lm:1 inverter:8 early:1 a2:4 released:1 beckman:1 instantaneously:1 always:1 inactivation:2 voltage:25 office:1 release:1 naval:1 mainly:1 nand:4 pasadena:1 vlsi:1 denoted:1 fairly:1 equal:1 having:1 represents:2 nearly:1 t2:1 preserve:1 national:1 interpolate:1 geometry:3 attempt:1 detection:1 saturated:1 operated:1 held:6 chain:1 edge:1 carver:6 euclidean:1 orbit:1 increased:1 impulsive:1 restoration:6 mahowald:2 cost:1 uniform:4 delay:4 too:1 conduction:1 minch:5 synthetic:1 migration:1 peak:1 ie:4 systematic:2 physic:1 transmitting:1 leading:1 potential:10 inc:2 idealized:1 ad:2 view:2 reached:1 maintains:1 ttt:6 slope:1 formed:1 il:1 kaufmann:1 subthreshold:2 ofthe:1 discharged:2 reach:1 nonetheless:1 energy:1 static:1 propagated:2 auditory:3 tunable:2 amplitude:3 back:2 nerve:1 exceeded:1 wesley:1 attained:1 ta:1 xxx:1 shrink:5 refractoriness:2 stage:30 correlation:1 until:2 hand:1 working:1 propagation:10 perhaps:1 building:2 effect:3 counterpart:3 evolution:1 hence:1 laboratory:1 moore:1 i2:1 during:2 width:30 excitation:4 m:2 smear:1 motion:2 passage:1 specialized:1 behaves:3 rl:1 refractory:4 exponentially:3 discussed:1 analog:1 eft:4 silicon:9 ai:3 trivially:1 stable:1 longer:1 integrity:1 perspective:2 scenario:1 certain:1 initiation:1 slew:3 accomplished:1 caltech:1 preserving:1 morgan:1 additional:2 greater:1 determine:1 period:5 ight:4 signal:7 semi:1 ii:11 multiple:2 full:1 exceeds:1 faster:1 characterized:1 polysilicon:1 cross:1 long:3 bigger:1 involving:1 idc:1 circumstance:2 cochlea:1 cell:3 ion:6 preserved:1 addition:1 fellowship:1 interval:2 decreased:1 grow:5 source:2 publisher:1 capacitor:5 near:1 ideal:1 neuroelectric:1 split:1 whether:1 speaking:1 cause:5 lazzaro:3 action:10 useful:1 detailed:3 diameter:2 sign:1 per:1 promise:2 affected:1 putting:1 threshold:10 tapered:5 douglas:2 pj:1 hasler:5 mosis:4 graph:5 hodgkin:3 clipped:1 almost:2 electronic:2 vn:9 patch:2 bound:2 layer:1 occur:2 precisely:2 huxley:3 ri:2 influx:1 speed:1 nmos:3 relatively:2 membrane:6 across:4 remain:1 lp:1 making:1 happens:1 dv:1 explained:1 taken:3 equation:1 turn:1 mechanism:1 addison:1 fed:2 serf:1 end:3 travelling:1 operation:3 altogether:1 gate:4 publishing:1 pcmp:1 neglect:1 especially:1 disappear:2 quantity:1 restored:5 distance:2 simulated:1 chris:6 toward:4 sloped:2 length:1 negative:8 implementation:1 design:1 twenty:1 perform:1 allowing:1 upper:1 neuron:2 wire:1 behave:1 precise:1 required:1 california:1 narrow:3 able:1 below:1 mismatch:1 reading:1 saturation:3 including:4 explanation:1 power:4 event:5 unrealistic:1 natural:1 disturbance:1 restore:1 sodium:5 nth:3 technology:1 brief:4 picture:1 excel:1 coupled:1 understanding:1 acknowledgement:1 relative:2 loss:1 expect:1 digital:1 foundation:2 propagates:6 share:3 supported:1 last:1 copy:1 bias:12 institute:1 feedback:10 curve:4 boundary:2 preventing:1 sensory:1 made:2 universally:1 replicated:1 san:1 excess:1 emphasize:1 compact:2 logic:1 correlating:1 quiescent:1 recharging:1 un:12 additionally:1 channel:8 nature:1 ca:2 inactivated:1 meanwhile:1 linearly:2 paul:6 nothing:4 body:1 depicts:1 slow:1 axon:43 position:3 governed:1 rail:13 tied:1 third:1 down:9 offset:1 r2:2 cease:1 magnitude:3 depicted:3 visual:2 v6:1 u2:1 determines:1 lewis:2 ma:1 narrower:1 consequently:1 appreciable:1 change:2 typical:1 except:1 operates:2 determined:1 accepted:1 experimental:1 support:1 mark:2 latter:1 avoiding:1
7,140
904
A solvable connectionist model of immediate recall of ordered lists Neil Burgess Department of Anatomy, University College London London WC1E 6BT, England (e-mail: n.burgessOucl.ac.uk) Abstract A model of short-term memory for serially ordered lists of verbal stimuli is proposed as an implementation of the 'articulatory loop' thought to mediate this type of memory (Baddeley, 1986). The model predicts the presence of a repeatable time-varying 'context' signal coding the timing of items' presentation in addition to a store of phonological information and a process of serial rehearsal. Items are associated with context nodes and phonemes by Hebbian connections showing both short and long term plasticity. Items are activated by phonemic input during presentation and reactivated by context and phonemic feedback during output. Serial selection of items occurs via a winner-take-all interaction amongst items, with the winner subsequently receiving decaying inhibition. An approximate analysis of error probabilities due to Gaussian noise during output is presented. The model provides an explanatory account of the probability of error as a function of serial position, list length, word length, phonemic similarity, temporal grouping, item and list familiarity, and is proposed as the starting point for a model of rehearsal and vocabulary acquisition. 1 Introduction Short-term memory for serially ordered lists of pronounceable stimuli is well described, at a crude level, by the idea of an 'articulatory loop' (AL). This postulates that information is phonologically encoded and decays within 2 seconds unless refreshed by serial rehearsal, see (Baddeley, 1986). It successfully accounts for (i) 52 Neil Burgess the linear relationship between memory span s (the number of items s such that 50% of lists of s items are correctly recalled) and articulation rate r (the number of items that can be said per second) in which s ~ 2r + c, where r varies as a function of the items, language and development; (ii) the fact that span is lower for lists of phonemically similar items than phonemically distinct ones; (iii) unattended speech and articulatory distract or tasks (e.g. saying blah-blah-blah ... ) both reduce memory span. Recent evidence suggests that the AL plays a role in the learning of new words both during development and during recovery after brain traumas, see e.g. (Gathercole & Baddeley, 1993). Positron emission tomography studies indicate that the phonological store is localised in the left supramarginal gyrus, whereas subvocal rehearsal involves Broca's area and some of the motor areas involved in speech planning and production (Paulesu et al., 1993). However, the detail of the types of errors committed is not addressed by the AL idea. Principally: (iv) the majority of errors are 'order errors' rather than 'item errors', and tend to involve transpositions of neighbouring or phonemically similar items; (v) the probability of correctly recalling a list as a function of list length is a sigmoid; (vi) the probability of correctly recalling an item as a function of its serial position in the list (the 'serial position curve') has a bowed shape; (vii) span increases with the familiarity of the items used, specifically the c in s ~ 2r + c can increase from 0 to 2.5 (see (Hulme et al., 1991)), and also increases if a list has been previously presented (the 'Hebb effect'); (viii) 'position specific intrusions' occur, in which an item from a previous list is recalled at the same position in the current list. Taken together, these data impose strong functional constraints on any neural mechanism implementing the AL. Most models showing serial behaviour rely on some form of 'chaining' mechanism which associates previous states to successive states, via recurrent connections of various types. Chaining of item or phoneme representations generates errors that are incompatible with human data, particularly (iv) above, see (Burgess & Hitch, 1992, Henson, 1994). Here items are maintained in serial order by association to a repeatable time-varying signal (which is suggested by position specific intrusions and is referred to below as 'context'), and by the recovery from suppression involved in the selection process - a modification of the 'competitive queuing' model for speech production (Houghton, 1990). The characteristics of STM for serially ordered items arise due to the way that context and phoneme information prompts the selection of each item. 2 The model The model consists of 3 layers of artificial neurons representing context, phonemes and items respectively, connected by Hebbian connections with long and short term plasticity, see Fig. 1. There is a winner-take-all (WTA) interaction between item nodes: at each time step the item with the greatest input is given activation 1, and the others o. The winner at the end of each time step receives a decaying inhibition that prevents it from being selected twice consecutively. During presentation, phoneme nodes are activated by acoustic or (translated) visual input, activation in the context layer follows the pattern shown in Fig. 1, item nodes receive input from phoneme nodes via connections Wij. Connections A Solvable Connectionist Model of Immediate Recall of Ordered Lists A) .....1 - - - - nc 53 ---I.~ ? ? ? ? ? ? 00000 0 0 ? ? ? ? ? ? 0000 0 00 ? ? ? ? ? ? 0 0 0 0 B) context phonemes 0000000 Wij(t) items (WTA + suppression) ~ t=l t=2 t=3 . translated ..,/ visual input 0000000 \ acoustic input buffer [0 output Figure 1: A) Context states as a function of serial position t; filled circles are active nodes, empty circles are inactive nodes. B) The architecture of the model. Full lines are connections with short and long term plasticity; dashed lines are routes by which information enters the model. Wij (t) learn the association between the context state and the winning item, and Wij and Wij learn the association with the active phonemes. During recall, the context layer is re-activated as in presentation, activation spreads to the item layer (via Wij(t)) where one item wins and activates its phonemes (via Wij(t?. The item that now wins, given both context and phoneme inputs, is output, and then suppressed. As described so far, the model makes no errors. Errors occur when Gaussian noise is added to items' activations during the selection of the winning item to be output. Errors are likely when there are many items with similar activation levels due to decay of connection weights and inhibition since presentation. Items may then be selected in the wrong order, and performance will decrease with the time taken to present or recall a list. 2.1 Learning and familiarity Connection weights have both long and short term plasticity: W ij (t) (similarly Wij(t) and Wij(t)) have an incremental long term component Wi~(t), and a oneshot short term component Wl,(t) which decays by a factor b.. per second. The net weight of the connection is the sum of the two components: Wij(t) = Wi~(t)+W/i(t). Learning occurs according to: if Cj(t)ai(t) otherwise, > Wij(t)j 54 Neil Burgess { Wil(t) + eCj(t)Uoi(t) if Cj(t)Uoi(t) > 0; Wij (t) otherwise, (1) where Cj(t) and Uoi(t) are the pre- and post-connection activations, and e decreases with IW/i(t)1 so that the long term component saturates at some maximum value. These modifiable connection weights are never negative. An item's cfamiliarity' is reflected by the size of the long term components wfj and wfj of the weights storing the association with its phonemes. These components increase with each (error-free) presentation or recall of the item. For lists of totally unfamiliar items, the item nodes are completely interchangeable having only the short-term connections w!j to phoneme nodes that are learned at presentation. Whereas the presentation of a familiar item leads to the selection of a particular item node (due to the weights wfj) and, during output, this item will activate its phonemes more strongly due to the weights w! '. Unfamiliar items that are phonemically similar to a familiar item will tend to be represented by the familiar item node, and can take advantage of its long-term item-phoneme weights wfj. Presentation of a list leads to an increase in the long term component of the contextitem association. Thus, if the same list is presented more than once its recall improves, and position specific intrusions from previous lists may also occur. Notice that only weights to or from an item winning at presentation or output are increased. 3 Details There are nw items per list, np phonemes per item, and a phoneme takes time lp seconds to present or recall. At time t, item node i has activation Uoi(t) , context node i has activation Ci(t), Ct is the set of nc context nodes active at time t, phoneme node i has activation bi(t) and Pi is the set of np phonemes comprising item i. Context nodes have activation 0 or J3/2n c , phonemes take activation 0 or 1/ y'n;, so Wij(t) ~ J3/2n c and wlj(t) = Wji(t) ~ 1/h' see (1). This sets the relative effect that context and phoneme layers have on items' activation, and ensures that items of neither few nor many phonemes are favoured, see (Burgess & Hitch, 1992). The long-term components of phoneme-item weights wfj(t) and wji(t) are 0.45/ y'n; for familiar items, and 0.15/ y'n; for unfamiliar items (chosen to match the data in Fig. 3B). The long-term components of context-item weights Wi~(t) increase by 0.15/.Jn; for each ofthe first few presentations or recalls of a list. Apart from the WTA interaction, each item node i has input: (2) where Ii(t) < 0 is a decaying inhibition imposed following an item's selection at presentation or output (see below), TJi is a (0, u) Gaussian random variable added at output only, and Ei(t) is the excitatory input to the item from the phoneme layer during presentation and the context and phoneme layers during recall: during presentation; during recall. During recall phoneme nodes are activated according to bi(t) = 2:j Wij(t)aj(t). (3) A Solvable Connectionist Model of Immediate Recall of Ordered Lists 55 One time step refers to the presentation or recall of an item and has duration nplp. The variable t increases by 1 per time step, and refers to both time and serial position. Short term connection weights and inhibition Ii(t) decay by a factor .6. per second, or .6. nplp per time step. The algorithm is as follows; rehearsal corresponds to repeating the recall phase. Presentation o. Set activations, inhibitions and short term weights to zero, t = 1. 1. Set the context layer to state Ct : Ci(t) = J3/2n c if i E Ct ; Ci(t) = 0 otherwise. 2. Input items, i.e. set the phoneme layer to state 1't : bi(t) = 1/..;n; if i E 1't; bi(t) 0 otherwise. 3. Select the winning item, i.e. ak(t) = 1 where hk(t) = maJC.i{hi(t)}; ai(t) = 0, for i =1= k. 4. Learning, i.e. increment all connection weights according to (1). 5. Decay, i.e. multiply short-term connection weights Wl;(t), w[j(t) and w[j(t), and inhibitions Ii(t) by a factor .6.n plp. 6. Inhibit winner, i.e. set Ik(t) = -2, where k is the item selected in 3. 7. t ---+ t + 1, go to 1. Recall o. t = 1. 1. Set the context layer to state Ct , as above. 2. Set all phoneme activations to zero. 3. Select the winning item, as above. 4. Output. Activate phonemes via Wji(t), select the winning item (in the presence of noise) . 5. Learning, as above. 6. Decay, as above. 7. Inhibit winner, i.e. set Ik(t) = -2, where k is the item selected in 4. 8. t ---+ t + 1, go to 1. = 4 Analysis The output of the model, averaged over many trials, depends on (i) the activation values of all items at the output step for each time t and, (ii) given these activations and the noise level, the probability of each item being the winner. Estimation is necessary since there is no simple exact expression for (ii), and (i) depends on which items were output prior to time t. I define "Y(t, i) to be the time elapsed, by output at time t, since item i was last selected (at presentation or output), i.e. in the absence of errors: . {(t-i)lpnp ifi<t; "Y(t, l) = (nw - (i - t))lpnp if i 2: t. (4) If there have been no prior errors, then at time t the inhibition of item l IS Ii(t) = -2(.6.)7(t,i+l), and short term weights to and from item i have decayed by a factor .6.7(t,i). For a novel list of familiar items, the excitatory input to item i during output at time t is, see (3): Ei(t) = 3.6.7(t,i)IICi n Ct 11/2nc + (0.45 + .6.7(t,i))21I1'i n 1't II/np, (5) 56 Neil Burgess A) 0.90 0.85 0.80 s 0.6 0.75 2 6 4 2 4 6 Figure 2: Serial position curves. Full lines show the estimation, extra markers are error bars at one standard deviation of 5 simulations of 1,000 trials each, see ?5 for parameter values. A) Rehearsal. Four consecutive recalls of a list of 7 digits ('1', .. ,'4'). B) Phonemic similarity. SPCs are shown for lists of dissimilar letters ('d'), similar letters ('s'), and alternating similar and dissimilar letters with the similar ones in odd ('0') and even ('e') positions. C.f. (Baddeley, 1968, expt. V). where IIX II is the number of elements in set X. The probability p(t, i) that item i wins at time t function(Brindle, 1990): .) '" p (t, 1, '" IS estimated by the softmax exp (TrI.i (t)/ 0") (), , Lj=1 exp (mj t /0' ) ntu (6) where TrI.i(t) is hi(t) without the noise term, see (2-3), and 0" = 0.750'. For 0' = 0.5 (the value used below), the r.m.s. difference between p(t, i) estimated by simulation (500 trials) and by (6) is always less than 0.035 for -1 < TrI.i(t) < 1 with 2 to 6 items. Which items have been selected prior to time t affects Ii(t) in hi(t) via "I(t, i). p(t, i) is estimated for all combinations of up to two prior errors using (6) with appropriate values of TrI.i(t), and the average, weighted by the probability of each error combination, is used. The 'missing' probability corresponding to more than two prior errors is corrected for by normalising p(t, i) so that Li p(t, i) = 1 for t = 1, .. , nw' This overestimates the recency effect, especially in super-span lists. 5 Performance The parameter values used are Do = 0.75, nc = 6, 0' = 0.5. Different types of item are modelled by varying (np,lp) : 'digits' correspond to (2,0.15), 'letters' to (2,0.2), and 'words' to (5,0.15-0.3). 'Similar' items all have 1 phoneme in common, dissimilar items have none. Unless indicated otherwise, items are dissimilar and familiar, see ?3 for how familiarity is modelled. The size of 0' relative to Do is set so that digit span ~ 7. np and lp are such that approximately 7 digits can be said in 2 seconds. The model's performance is shown in Figs. 2 and 3. Fig. 2A: the increase in the long-term component of context-item connections during rehearsal brings stability after a small number of rehearsals, i.e. no further errors are committed. Fig. 2B: serial position curves show the correct effect of phonemic similarity among items. 57 A Solvable Connectionist Model of Immediate Recall of Ordered Lists B) r w 4 T 3 u n - 2 0 5 10 0.0 0.5 1.0 1.5 Figure 3: Item span. Full lines show the estimation, extra markers (A only) are error bars at one standard deviation of 3 simulations of 1,000 trials each, see ?5 and ?3 for parameter values. A) The probability of correctly recalling a whole list versus list length. Lists of digits ('d'), unfamiliar items (of the same length, 'u'), and experimental data on digits (adapted from (Guildford & Dallenbach, 1925), 'x') are shown. B) Span versus articulation rate (rate= 1/ipnp, with np = 5 and ip =0.15,0.2, and 0.3). Calculated curves are shown for novel lists of familiar ('f') and unfamiliar ('u') words and lists of familiar words after 5 repetitions ('r'). Data on recall of words ('w') and non-words ('n') are also shown, adapted from (Hulme et al., 1991). Fig. 3A: the probability of recalling a list correctly as a function of list length shows the correct sigmoidal relationship. Fig. 3B: item span shows the correct, approximately linear, relationship to articulation rate, with span for unfamiliar items below that for familiar items. Span increases with repeated presentations of a list in accordance with the 'Hebb effect'. Note that span is slightly overestimated for short lists of very long words. 5.1 Discussion and relation to previous work This model is an extension of (Burgess & Hitch, 1992), primarily to model effects of rehearsal and item and list familiarity by allowing connection weights to show plasticity over different timescales, and secondly to show recency and phonemic similarity effects simultaneously by changing the way phoneme nodes are activated during recall. Note that the 'context' timing signal varies with serial position: reflecting the rhythm of presentation rather than absolute time (indeed the effect of temporal grouping can be modelled by modifying the context representations to reflect the presence of pauses during presentation (Hitch et al., 1995)), so presentation and recall rates cannot be varied. The decaying inhibition that follows an items selection increases the locality of errors, i.e. if item i + 1 replaces item i, then item i is most likely to replace item i + 1 in turn (rather than e.g. item i+ 2). The model has two remaining problems: (i) selecting an item node to form the long term representation of a new item, without taking over existing item nodes, and (ii) learning the correct order of the phonemes within an item - a possible extension to address this problem is presented in (Hartley & Houghton, 1995). The mechanism for selecting items is a modification of competitive queuing 58 Neil Burgess (Houghton, 1990) in that the WTA interaction occurs at the item layer, rather than in an extra layer, so that only the winner is active and gets associated to context and phoneme nodes (this avoids partial associations of a context state to all items similar to the winner, which would prevent the zig-zag curves in Fig. 2B). The basic selection mechanism is sufficient to store serial order in itself, since items recover from suppression in the same order in which they were selected at presentation. The model ma.ps onto the articulatory loop idea in that the selection mechanism corresponds to part of the speech production ('articulation') system and the phoneme layer corresponds to the 'phonological store', and predicts that a 'context' timing signal is also present. Both the phoneme and context inputs to the item layer serve to increase span, and in addition, the former causes phonemic similarity effects and the latter causes recency, position specific intrusions and temporal grouping effects. 6 Conclusion I have proposed a simple mechanism for the storage and recall of serially ordered lists of items. The distribution of errors predicted by the model can be estimated mathematically and models a very wide variety of experimental data. By virtue of long and short term plasticity of connection weights, the model begins to address familiarity and the role of rehearsal in vocabulary acquisition. Many of the predicted error probabilities have not yet been checked experimentally: they are predictions. However, the major prediction of this model, and of (Burgess & Hitch, 1992), is that, in addition to a short-term store of phonological information and a process of sub-vocal rehearsal, STM for ordered lists of verbal items involves a third component which provides a repeatable time-varying signal reflecting the rhythm of the items' presentation. Acknowledgements: I am grateful for discussions with Rik Henson and Graham Hitch regarding data, and with Tom Hartley and George Houghton regarding error probabilities, and to Mike Page for suggesting the use of the softmax function. This work was supported by a Royal Society University Research Fellowship. References Baddeley AD (1968) Quarterly Journal of Ezperimental Pllychology 20 249-264. Baddeley AD (1986) Working Memory, Clarendon Press. Brindle, J S (1990) in: D S Tourebky (ed.) Advancell in Neural Information ProcelJlling Syatemll ! . San Mateo, CA: Morgan Kaufmann. Burgess N & Hitch G J (1992) J. Memory and Language 31 429-460. Gathercole S E & Baddeley A D (1993) Working memory and language, Erlbaum. Guildford J P & Dallenbach K M (1925) American J. of Pllychology 36 621-628. Hartley T & Houghton G (1995) J. Memory and Language to be published. Henson R (1994) Tech. Report, M.R.C. Applied Psychology Unit, Cambridge, U.K. Hitch G, Burgess N, Towse J & Culpin V (1995) Quart. J. of Ezp. Pllychology, submitted. Houghton G (1990) in: R Dale, C Mellish & M Zock (eds.), Current Rellearch in Natural Language Generation 287-319. London: Academic Press. Hulme C, Maughan S & Brown G D A (1991) J. Memory and Language 30685-701. Paulesu E, Frith C D & Frackowiak R S J (1993) Nature 362 342-344. PART II NEUROSCIENCE
904 |@word trial:4 simulation:3 selecting:2 existing:1 current:2 activation:16 yet:1 plasticity:6 shape:1 motor:1 selected:7 item:106 positron:1 short:15 normalising:1 transposition:1 provides:2 node:22 bowed:1 successive:1 sigmoidal:1 ik:2 consists:1 indeed:1 planning:1 nor:1 brain:1 totally:1 stm:2 begin:1 wfj:5 temporal:3 wrong:1 uk:1 unit:1 overestimate:1 timing:3 accordance:1 ak:1 approximately:2 twice:1 mateo:1 suggests:1 bi:4 averaged:1 digit:6 area:2 thought:1 phonemically:4 word:8 pre:1 refers:2 vocal:1 get:1 cannot:1 onto:1 selection:9 recency:3 context:27 storage:1 unattended:1 paulesu:2 imposed:1 missing:1 go:2 starting:1 duration:1 recovery:2 stability:1 increment:1 play:1 exact:1 neighbouring:1 associate:1 element:1 particularly:1 houghton:6 predicts:2 mike:1 role:2 enters:1 ensures:1 connected:1 decrease:2 inhibit:2 zig:1 wil:1 grateful:1 interchangeable:1 serve:1 completely:1 translated:2 frackowiak:1 various:1 represented:1 distinct:1 london:3 activate:2 artificial:1 encoded:1 otherwise:5 neil:5 itself:1 ip:1 advantage:1 net:1 interaction:4 loop:3 gathercole:2 ecj:1 empty:1 p:1 incremental:1 recurrent:1 ac:1 ij:1 odd:1 pronounceable:1 phonemic:7 strong:1 expt:1 involves:2 indicate:1 predicted:2 anatomy:1 hartley:3 correct:4 tji:1 modifying:1 subsequently:1 consecutively:1 human:1 implementing:1 behaviour:1 guildford:2 ntu:1 secondly:1 mathematically:1 extension:2 exp:2 nw:3 major:1 consecutive:1 estimation:3 iw:1 wl:2 repetition:1 successfully:1 weighted:1 iici:1 activates:1 gaussian:3 always:1 super:1 rather:4 varying:4 emission:1 intrusion:4 hk:1 tech:1 suppression:3 am:1 bt:1 lj:1 explanatory:1 relation:1 wij:14 comprising:1 i1:1 among:1 development:2 softmax:2 once:1 phonological:4 never:1 having:1 connectionist:4 stimulus:2 others:1 np:6 few:2 primarily:1 report:1 simultaneously:1 mellish:1 familiar:9 phase:1 recalling:4 multiply:1 hitch:8 activated:5 articulatory:4 partial:1 necessary:1 unless:2 filled:1 iv:2 circle:2 re:1 increased:1 deviation:2 erlbaum:1 varies:2 decayed:1 overestimated:1 receiving:1 together:1 postulate:1 reflect:1 american:1 li:1 account:2 suggesting:1 coding:1 uoi:4 vi:1 depends:2 ad:2 queuing:2 competitive:2 decaying:4 recover:1 phoneme:34 characteristic:1 kaufmann:1 correspond:1 ofthe:1 modelled:3 none:1 published:1 submitted:1 checked:1 ed:2 rehearsal:11 acquisition:2 involved:2 phonologically:1 associated:2 refreshed:1 recall:21 improves:1 cj:3 reflecting:2 clarendon:1 reflected:1 tom:1 strongly:1 wlj:1 working:2 receives:1 ei:2 marker:2 brings:1 aj:1 indicated:1 effect:10 brown:1 former:1 alternating:1 during:18 plp:1 maintained:1 rhythm:2 chaining:2 blah:3 novel:2 sigmoid:1 common:1 functional:1 winner:9 association:6 unfamiliar:6 cambridge:1 ai:2 similarly:1 language:6 henson:3 similarity:5 inhibition:9 recent:1 apart:1 store:5 buffer:1 route:1 wji:3 morgan:1 george:1 ezp:1 impose:1 broca:1 dashed:1 signal:5 ii:12 full:3 hebbian:2 match:1 england:1 reactivated:1 academic:1 long:15 serial:14 post:1 j3:3 prediction:2 basic:1 receive:1 addition:3 whereas:2 fellowship:1 addressed:1 hulme:3 extra:3 tri:4 tend:2 presence:3 iii:1 variety:1 affect:1 burgess:11 psychology:1 architecture:1 ifi:1 reduce:1 idea:3 regarding:2 inactive:1 expression:1 speech:4 trauma:1 cause:2 oneshot:1 involve:1 quart:1 repeating:1 tomography:1 gyrus:1 notice:1 estimated:4 neuroscience:1 correctly:5 per:7 modifiable:1 four:1 changing:1 prevent:1 neither:1 tourebky:1 sum:1 letter:4 saying:1 incompatible:1 graham:1 layer:14 ct:5 hi:3 replaces:1 adapted:2 occur:3 constraint:1 generates:1 span:13 department:1 according:3 combination:2 slightly:1 suppressed:1 wi:3 lp:3 wta:4 modification:2 principally:1 taken:2 previously:1 turn:1 mechanism:6 end:1 quarterly:1 appropriate:1 jn:1 remaining:1 iix:1 wc1e:1 especially:1 society:1 added:2 occurs:3 said:2 amongst:1 win:3 majority:1 mail:1 viii:1 length:6 relationship:3 nc:4 localised:1 negative:1 implementation:1 allowing:1 neuron:1 immediate:4 saturates:1 committed:2 varied:1 prompt:1 connection:18 recalled:2 acoustic:2 learned:1 elapsed:1 address:2 suggested:1 bar:2 below:4 pattern:1 articulation:4 royal:1 memory:10 greatest:1 serially:4 rely:1 natural:1 solvable:4 pause:1 representing:1 prior:5 acknowledgement:1 relative:2 generation:1 versus:2 rik:1 sufficient:1 storing:1 pi:1 production:3 excitatory:2 supported:1 last:1 free:1 verbal:2 wide:1 taking:1 absolute:1 feedback:1 curve:5 vocabulary:2 calculated:1 avoids:1 dale:1 san:1 far:1 approximate:1 active:4 learn:2 mj:1 nature:1 ca:1 frith:1 distract:1 spread:1 timescales:1 whole:1 noise:5 arise:1 mediate:1 repeated:1 fig:9 referred:1 hebb:2 favoured:1 sub:1 position:14 winning:6 crude:1 third:1 familiarity:6 specific:4 repeatable:3 showing:2 list:39 decay:6 virtue:1 evidence:1 grouping:3 ci:3 vii:1 locality:1 likely:2 visual:2 prevents:1 ordered:9 corresponds:3 ma:1 presentation:23 replace:1 absence:1 experimentally:1 specifically:1 corrected:1 experimental:2 zag:1 select:3 college:1 latter:1 dissimilar:4 baddeley:7
7,141
905
oqp r s t u                  ! "#  $ "%  % &   ' )(+*,   #  - QSRUTWVXZYJa1[]\\Z.+[^_`V1R/1RbuW02VaU/4Xmc1vx36dewy\S5 fgR1Yb/8^z|{k79dA7;fPhi:XZ< 0dZ[]^}|:<>j\~=c V XZ[k\Zl]XSc QmRT V XZY aO[k?AZ\Z[^@CnoVRBERpT DFV aXScv/Gd 7Iw?\SfqHJR Y ^?`/ {]d :KMfPL hEXZ/Od []NP^?|rsW@ c VXt[]\ V XSc a1\t[]RT~V?S? ? ? ? ? ?U? ?]? a8\Z[kRT V X~v??8?????k??]?U? ?-????4?8???G? ?? V f XZ?? d fq?m?? { f??4fP? dZf????k[ ?e? ? ? ??V ??XZ^?|\Z? ?OY Vd? \? ^_ ?8Y?[ [kXt[ ??R1?Z\SXZfg[ f?^? R \Z^n?4R4f??xde?kYJ??? [ V??? ?tfP^?|dZ\Z\?8?? l]d?dZdt? ^?|V1r ?? f R ?fP?XSf;?tvG? ^??no?P\t? ^_ f V??R ? fPS?XZX fgV XSX?feT fgV??R1XS? f)fPX ?X Vc r Vk?4? dZfP\ XA? ? ^???\ ? ? ? V XSdZfPded ??o? r f;? ?Pf \??s\^? deV \t? Y XSr V ? v8R8fP?;? fg?d R1V ??T?fP?iXZ\Z?][Ufg^?M?R\Zfg^z|f?R8dfPX [k? Rc8v ?8[k? R1[]v?;? ? XSV f T dZ\S\e^_|Y?[\Z[ ^? de\ \V ?? XZ?Pd?[ ? [kXS^?|RfT V^??RX Y?sRO[^? \Z\S??|f;V??R?vG?n VYbR?f9R\Z?1de?? f?VUR8fP[?]?V ?i? ^?|fPR?R V1\Z??1df?dedZ^?f;? ROROdZ[ f ? d ? ? ?8[]? \ ?1?1fb? fgvGR1[]d \Z? ^??^?R R ? ? ? fP?1dSf? f;?? ?de? ?fP??|YJ^_??][ ?[]? V\\Zf??Z{d d??f;^??gdA??\Zl\eX+?1? ??fgfP? dZr? \Z[][^? ^?? ?eR ? []?? \mX R1fV ??|?\ZXSffg[ ?? ? vs? ?]?? ^z|??\mXt?? f;VfPd?? R1d??4? ^?|R1V? ?lR8V? [ ?S? ? ?n f;? [ v ?O? ? ? ? f ? V V ?? ^? ?1?[ ^??? ???]?Z?fq^nol?XmfPR?1{1? \[9^? lV ?? ?? ??d ? \Zf;? ?8dZfPf?f?^? ? f ?M?? dS\?? ?v8? fP??Y ? V R1? dt\eXZV [k? \Z\Zfe^v? ? [ ? ?n \~? V TAr8[]??? ? ? V ? [ ? [ ? ^n V]???e?1[ ?+??? ? [d r f ? ? ? ? ? {4^? V? d ? c ? ? ?]? ??8?????y?s?8???? ? ?fqR1??\Z? X VdZ?^?|? c [ ? []R? v f ? ? ?1fP[kVUdvs\Sd+? d ?n? R&? [ f?XSdZfP\ZdP^_ Y??[XSfP? ?^n V ? R?^? XZf?\t?8? f V ? XSXZ?^_|lU?WXSd ^?2? ?U? ?R V []??d ? fe??v ? [kfI?^?|l Y ? ? \ZY+?1fJ? ? de^?M? \Z[fg??\Z^??? ?8d ?VU^noV?qv[ ? ? ???X [UVk??s^?fPY Xt\Z? ^_ofPY d l ? T?R8?;\tfg?1?Zfb\ ? f9^_ ? RV \Z?^_ f?f d#R1V1[k? Rd?v de?^z ?? ??8? [ ? dV ?T QSd R\Z[ d? V?? de?b\Z^z fp? ? [ ? ? ?? ? R ?? ^_|V ?g?t[]??\Z^_|?klU\ZR1?? dqV ?sR ? ? V ? ? ?Gf;^{? ?efP?X ? ?8V [ ? d fb? f;^_odyv T?[]? V??f;v[]R ? ??^} R8\ ?g?SfYb? lU???8R f ? ? ^ R ? \S? f X f;? d []\?R4?vF?|R \?Yb?1fb?? ? Xm??f ? ? fP[kXt?f ? ? ???tfP\Zd ?_ J? ?1[ fP? ?XS? V fgR^n|R ? ? ?t?f;f ^? \t? ? ? ? ??O? ?brG[ f ? ? ?nof;f ?T ? ;?8? ? \ ? ?Of? Xt? ?e? d ?]?V ? \ZR1^_ fP? ? ? ? ??d V [ ? ?qR?v ??? ??v&[ ? ??? ?? ? ?? ? VUXZ^} \Z? bd ? ^? ??? rsfY V ?mf?XZlkr ? de\y[]R4v ? ? dZd de??R1de^? \t^? {]f V ? []XZ[kY?fP\mfPX {k[]XZ^n [ ? ^_ V?4d _ ? ? n? ? ? = ? ? n ? ?? ? ? ? ? z ? n ? ? _ _ _ ? ?            vqw  !#"$!&%')(*,+ x 0/1325476,8:9 1 <;>=?A@CB)D7?FE 3 = IJ;>8 3 B O C9 8:=,PJ;08 3 6 G&1H26,8C1H2 , 6 K132ALNM 9 5 6 Q#/R1)2ALNSUT +.? 9 G09 ml O 6 A? n 6 j ? ?8 6 / 1 O L L 1 B`_:a0b [,cUdXe3Y b ) 6>h _:a0b [,c$d,e)Y f a d3Y 4) 6 9j VXW,Y,Z>[ [ Y,f Ug U\^] ] k # Ui ?             !   "#$ %'&)(+*-,/.0 13254 68795:<; =?>@: & >/AB: 0 C#DFEGH 75IKJ J L5MONQPH ERPTSVUW N &YXZH E G\[O])9@(8^_0 13>`bac E def* agih I J j kl E &em2 EnKop ` a q ErE G 0 C W ds(+tuv =KwxHyE ` X E &> ] n Dz=Kw|{FE & > h }J ~j l 7GH P/n<u `` =?&XDz;m?YW.? N 4  > P Dz;B?ySZ= e?E > E q G'4 ?#R P?E D r W ?0  XV? ?/.iG'H E X ?E R P D W d rB? P?P?d n adS4 1 > = ` =?e??'W ?P-E p p ` a ? ?  ? E X P ?)??? W ? DzP?ES ?+?'Sb? uE ` PKS l S W? u X'4 1 W dS<?E `'=?0 1 & > P?P >@?)? W q G0 1 vYE D E & > DzW n?E DzDz; ? ? a q X 4 ? v E DE w?? ? 4 ?#S<? 0 ? & W u ` R 0 1 P c ?<n W? d= n GS?= E ` D?0  = ? N W ?? 4 ? d?Dz= E ` d50 ? & r XH P a ` ;?Xb? r W ` 0 1#G H5v . 7?` =K.p?= n?G 4 6?R = ??; v W ` = ` = n =?&X!0 ?3e?= , 0 1 & ` W ?!2 ? u .XO=K.VG'0 ? vfE X0 C W & 7 E &e?=?? p{xE4 1 &S?NH ;?[])9 E&e ? E n ?qy? a q E r EG0 ?3W d HER<=???W ? & e N 0 ?3ey= E p p5Dz0 ?#??E 2 4  {z0 ?#G ;?? ? d . u n H@E > 0? R = ` .= ` E & r =?W ? p ` W 2y{z= v . ? H=@??W? E ? ?@? Wp5G'4 ? v E ? ?? G ;YW ? 2EnKo p ` Wp ErEG 0  a &?0  v p?z0  =K.?GHyEG?2E no p ` Wp E r E GZ0 ? a& v 0 ??& ? ??? v 0 ?#? =?.?GH=?=?&= ` r ; r E 0  & ?+` W v X'Hy=?ey4 ? .VG u ` 2 E & ? =?.?G W GH5= p`= ey4 ? n G'4 ?#a&?= `Z`W` . 7 W &5D ; 0 C ? GHP ?? &? C G 0 ? E D? W & e0 C G0 C W & ? ? . n ?zW.=?=?&yW u rH XbW XH=G ` u = N =?0 ? r HGO? = n GW ` E &e/0 ? ? GH P e0 ? .X u ` 2 E&nK=?. E? = . vfE ? ?=K& W ur H %'&?XH ? C .\pyE p?P ? Ni=?e?= ? 0 C ?= r ?|a ? E?? ? W q G'4 ?#??E D =?.G0 ? vfE GW ? . GH E Gi? 0? &50 ? v 0 ?3? = G?5P?=<&= ?r ; r E ? 0? & ? ? W v XU=-e0 C .ZG'? ? 2 E &??=?.?GZW?G H5= q` =Ke0  n X 4 ? W d?= ``bW` .?? W ????F? 0 ??&0??G0  E D n W&e?0 ? ? G0 ? W &S E& e)e ? ?3.Gb? ` 2 Edy?K=?.<? Hy= ` =?S'???z?'4 ?#dr?PKSV?'4 C ?fE X'W ? (8r 4 C ?PKdf2 ; H=?a ` =????M?H E S r ? W N 4 ?#& r ? = v W ? ; 7 N Hy0 ? n H NiP ` PV? = ` ? XbW E . ? =?0 1 &yr ? 1 d? d50 1#X'=K? e0 1 ??PKdyS0 ??a &E?'??S0 ? ? d n P u py> E X 0 ? &r X H5P c =?4 ? rHX HyPKd)GH= u d ? R= n Ga ? =K.GZ0 ??v/E G'=T`= ? u 0 6?? P S ? ?W N ?z= e r P W ??E D|? qy` =<R 4 1 W?ySTNQPK4 1 r H X P?. X 0 1#? E GP?S >y=?` Dz;0 ?#& r ?fW > =?D!0 ? S?Dz0 ?#&= E `<? N =?.Ha N G'H E G?XHy0 ?3S?0 ?3dy?5dy0 ?#? PV? e0 ? v =?&.Z4  W& E DOPK.VX'?!0  ? ? E Ga ? ` =K> uynKP . X'WfX H= ? &y0  X=?+e?0 ? ?/=?&S? 1#a &y? D?[O]????D|G P ` ^ H P &?X HyPBu d > = ? {z;? ? & r ?Yae=?D?4 ? Sd5a &yDz? ? &= E `?7 GH P ? ? &?5&y4 ? GZP ?'> 4 ?? = &5S 0 ? W &5E Dz0 C G ;?W?iX HyP-PK.G4 ? ? E1 X'W `?? E ; p ` aHy0 13254 ? G 0 ? GZS p` E n G0 ? n?E DOEp p D|0 ? n E2?D?4 ? G ; 7 E & > W & =?&=?=Key.?G'W n a &5.G ?u n G ? &y0C G=V?+e0 C v =?&.0 ? W& E D Epyp ? W ?0 1 v E G0 ? W &.BG W GH0 1 .?=K.GZ0 C v E X'W ?  ^ = n W &. 0 ? e?= ? G c a/.b? n?)E pp ? a ? 0 ? v ?G 0 ? W &.??y= ? = C W &= ; 0 13=???ey.?GZH=B2yE nKo p ? a pyEr E G06 W &YE w r W ? 0 C?? H v?? E & e?G H= W GZ?y= ? ? ? . E . = n W & e?'W ? e?= ?TEDzr W` 4 ? GHyC v?2 E .=Ke W d?GH =??= N G'W & ?tTE pHy.ZW&?4  G PK?E X'?  W & H= ? =?E ? = ?& W e W? 2G<?<E N 0 C ey= ? E ? 0 ?3=?X ; W ? WGH= ` E pyp ? W?0 1 v E G0 1 a d.?NH4  n H?.VH W ??D#e?2?=N W ? X H ; 1 ?W ? ? ? ?G H= ? . n??u G? ? & ; C ? ??? ??????   '? %'d?=K.X0 1 vYE Xb0 1 W&?p ? W 25Dz= v . W &= E S.? v =K.?E n = ? E 06 d v a e=?DF.E ; Ed % t ??DzGP ` 0 1 &?EeEp X 0 ? ? P ? zD ? = ` 0 1#& r ? W ` E d= u?E Di&=?G c a ?o M ? W2.= ?Z? =?Sf: n W ? u p X PVe ? = ? .0 13W &?a?GH= a Gp u X W ?Hy0  . v W > =?D ?yE &e NQE & G.?G'W)=K.G0 C?vYE GZ=?XH = q E ` E ? =?G'= ` .?E..W n 0 ? EG=?e N 4  GH X H4  . ? WyP?D (8. E ;?X HP N =?0 ? r HGb.!W ? G?y=% tm? DzGb= ? W ? &= uy? E D5&=?G c a ` o ]?W.ZGO=?.VX'4 1 v E ? ? W &?E Dzr W` 0 ??GH ??. v Eo =Q.a v P !? #"$&%('*)*+-,?E 2 W uG X H=?dyE G u? =-W ? G H=Te?0 13.G'? ? 2 ? E&n =?.?7 E &e?GH=K& p ` W n =?= > Ga?PKSG0 ? ? E GZ=GHP py? ? ? v =?G = ? . u .V?  & r .a v =?Wp5G???vfE Dz0 ? G ; n?? 0 1#X'= ? 0 C W& W ? = v W ? =?.p5= n 0 ? ??n 7 NQ=?.H5EDzD n W&.0 ? e= ? ? ? G'Hy= ? WD|D?W c ? ? & r G N W n?E .=?. ? .0/21 3547698: ;=<>6? @BA5?DCE6 F  pypyWS=YGZHyE X N =?a 2yS= `HG PfEd a ? Xp  X?S P ?? P &I PKJMLN OQP ?'H E G-W 2?P ; S X HyP?? W Dz?|aN?4 ? d5r D 4 ? dyP?E ` ?/W > =?D L N RTSUWVXN YZ\[] N O ? N?H = ? = V XN ^`_acb VN dfehg N ^ ? VN ^ ikj  ? S Eml dW N d ? 13d?p u G ? P n G W` ? Z 4  .?G N?=?0 ? r HG ? = n?G W`)X HE G N =?0 1 & G1?=?? & ? > G W =?SXb4 ? ? E X'= 7 E&enJ ] N ^oP 0 1 .@E &  & o &yWN d .'=?? ?y=?& n = [ = Gqpsr dtuwv ( Lx ?yLDe 7  ? LN z M ey=?d a G = GH = =?.X'4 ?#vfE GZ=?W ? GH=-N?= 0 ? r HG ? ?L?{ ^oP ? ? W ? Xb4  v =?k?? q Ga?E & e@0 ? & n DFue?4 ??? r GH=-0 ? & q ?G C . J!V|{o}~P?E d e GZ?=?W u Gp ?y? G.? ? W. X NQ4 ? ey= ? ; u .=em=K.G0  ?YE GZP ps? ? 7 0  . W & =?XH E G?. E X'0  . ? =?.TXZH=?? W DzD W N 0 ? &yr ??? n ` v 0C & ? ? ???? e!? N ^? p ? ? ? e ? ????? L ? z|? V o{ X ? p?? ??? 7 { ^o?? x ?I M H = u&yo da N & e ?? . ? ??'2?E &yn = R= n XZW ??r 0  ? =<& ?'0  v =?? o??? H P 0  ?'= ` 4 ? W& (? M N H = ? = ? 0 1 . E q WS0 ?#Gb4 ? ? = n W d .VX E &XGH5E X ? P?P n G'S?E py`H 1 W ? 0 ? l dW N zw =Ke?r P , 'X WfHa N n D?W.P p 0 ? S X'W GHyP?0 & 0 1#XZ0 C?E D=?. ? 0 ? ? E G P?p&? e H=??E? ?-??? SZW w u XZ0  ? d XW?XH P?E  2 a??= n ` 0 1 P ? 0 ? W d 0 C S r 0 ? ? P & ? ; P ??? F E ?zr a `y 1#??y? ( * E ;? 4 ? ? C ? I ? J J???  X H? ? p N ??? p ? ? ?>e ?w??? N z ( L N ^?? V XN ? p M ? p ? e  (? l ?            ! #"%$'&(*)+ "*, - /. ,    0 1   32 5476       8 9 :   ;=<=> ?'@ACBDA3EFHGJIKAMLKNODAMPRQ @ ATS V U S W#XYSW#Z5[K\Z5[^]KA_ a `b ij kHa5clnh mpo dHe fVagKe k r k gqe d g f ZL I k f g k a g ? qk a d k u d ( G } k u | k o * u y ~  ? d std dw??vyxz  { g ku d ka g ka ? ????J? Z ?Y?Y]?? ? QY?KZ*??? ?V? OY@ ???????????? ??? O @ ? E???E?HGRZL ???%? k a#? Z?YA3?7A?? ?%?Y? AMZL?S?#L ? AM??A?L? ? ? L???Z?=???YS ?VZL???Z? ? N ??? Z5?DS?#Z??[ A ? ? S ? QY? ? Z??S?#Z?K? ? ?J?H?RZ ?=I m5? ?DA???? A??MO S \ ???^? _ Q?@ A?L3XY?KA ? NP?O ? ?=???CODS ?VN ? ?  ? ???? S? P??YS #? ? ? ? OY@=A???P ? ? S Z O ??? ? ??*??? ? ?#? A???S  ?KNN????] L ? X?S?#NL '? @ ?=P?X @ ? A P X?S? ? Z O A ?? ? \ A??T?? ? S LKS? ? S?#? ? ? ? ?R? ??? S? [ [=??AX @ A ? Z? ?  ??? ? ?D? ? V? A ? S @ N N IA?P?ODS  ? Z5XDA'N*?HX @ A ? AMS- ? @ X?\A ??? ODNB(E?? ? ? ??Z ??O??  ? ? ? Z ? ? S O ?C? L?? A?D?KN5?L?O @ Z O'? U ? AC?(O @ AMP ? Z P?PY] ? ??O ? ? N L P _???? ? ? ? ? ? ? ? ? #? ? P ? @ ??? ? ? ? ? O A ? ? ? ?7? ? ? Q ?  N ? ? ? ? ? ? ? a ? ???#?  a  ? ?  E  g  E a   k a  g  X ? X @ Z O?Q?@ ? ?!#" Z ? ? N ??S ? O @ ? S?V??Z? ? ? xg k & ? ? ?(' @ Z5OOD? ??? P?X?S ? Z O ? P ? ? ? ] ? ? ?YZ I?S?#A ? O ? ? ??? ?  E k &*s E k a +-,/. x/0  7/897 Z $=??? ? ? ?  ??Z5XYS? N ? ?YN ??? ? ?@ A??YA ` i % k a SW P3?DA ? ? Z ? A7I  I?ZO ??? Z[?N? ? XD?KAI?S? B?A ? XYS) N U N ?X @ ? S? LPYXYZLXYZLA7N]?P ?65 ? E   k u ? 132 ? g qk a34 ? ? a :<;!=?>A@CB!DFE GHB!=JILKNM<IPOQ= ? ? ??? NP ? LN ? O @ Z O ?J? ? ??P ? BD\ ? ZL N]?XY?K??X PYASR] ? L ? ? ?1 k a ? X @ ZX N5? A ? P?O @ A ? N ? ? N?'S? L ? L N? ? S  L A Z B ? NI?A ? 1 ?[ ? k T3UlWV k f ? E ? . ?X YHZ cb e d ?J@L\ B ? V k ? ? ?) ? Z ? L N?L^] 1 ]J_a` u ] ? ? L ? XYS? N L ? ?'S  O?????N? ?=I A7Igf?BDP?X Z ? I P?A ? N LI ? B IKA?? I ? ??S  \ ZX?S \ f ? ?? ? ? E S P?O?@ A?] ? ? L N? L ? ? ?  ? @ X ? AM? ? N ? ? A ?  LXYA?L=I?? ? AMP?OYS ? Z O A ? ZLIih ? e j ? S? P?Z*? ] U ? ? ?*? ? I ? ? PYO???D??Z5? ? A PYAk?=A?L ? A  ? ??Z ?=AM]K?YZ ? ? L=A%O ?JN ? ? ? NLXYA%?X3O @ A S L=I?A??ml  S?#L V k a ?  ? ? S ? ? ? NB?B ? ? ? ? ? ? O N?O ? \ ? ? L ?^?  ?=A%Z*? ? ?=L ? O S ? ? O ?=Z O Z ? ??O ? A??A ? ? ? @ X\ ? ? X N ? ODN?OY@ A N ] O ?Pn X ?@ ? ? X ? ? l o O ? S? ?K?Pn O ?KZ OYO ? ??L ? ? ? ?YAMPYA L O ? ? ? Sp ? A  ? V k Y ? E ?rqs V ? ? g k f ?? E ? ? @ AMBDA g k a S- P?OD?KAut ? QD@ S L=?? ?KO ? Z?YO ? ?YL w v  ? ? ? ? ? ? ? ? ? ? ????Z ? ? ? A?? 3xD? ?=?+E ? k y sz|{ ?1 ?  ?C? ?  1 k f } ? @J\?A P?XYS  ? ZX?A  N ?pX ? ? ? ? ? ? @ O~ ? ? O ? ??PYS ? ? ? ? Z ?Y??B ? ? L???J? ? O ? Z ? ? S L ? ? ?I?S) ? ? O?S A t  ??? @ A??  ? ?DS ? O AMBYS NL3? ? B ? ?=I?? ? ? ? ?Y@ ??? ? ' ? ? ? Z O ? ? ?#? ka ,?? ? ?  ? o a ? ?? ??  ???  m  1 a s ?? ? ? ? ?x ? ? ? V E ? ?a ?? ? ? ?+? X @ Z? ?  ? O?? ? ? N ????? ? [ ? X @KA I?S  P X ? ? ? Z*? ? A7PTE?? E F , v P S ? L?X @ AR??S ? Z5? ?? PYA _ S ? ?/? A?ZPYP?? ? ? ? ? ? ? X ? ZJL ? ? ? Z? ? ? ?N ? ? A Z5L S ? I A ? A?L=I?A?L=O??RZ*]?PYP?S Z ? ? Z ?KI ? ? \ Z ? ?  Z5? ? A?P ? S ? OY@ \5Z? ? ? Z L ? A?P x ? ZL I BDA7P ? A ? QDS  \A ? ? O @ ACL X @ A ??N P?X???=? ?CO S NL S- ? F? ? ? ?  ??X @ ??? N ?? ??? ?  ? A ? S! @  N I ? ?? ? X?S? ? ??Z?=I ? O @ A ?J? S? ? @ O ?*? ? X ? B XY@ Z Q ? ?? ?KS ^? S  ? ? P(S  O'S P ? @ A ? Z ?S #? ? <? ? [ S: ? A?? S  @KN ? I?A?P?X ?  ? ZX??A ???N ? A \ A?B ?  ? O ?YZ? ? O ? OY?  ? a[ ?WV? ? Z ? ? Z*? ? ? X @ A P ?? ?=X?S ? ? L X ? ? ? } ? S [ [ ] 31 ? _ S? L ? A???A?BYZ ? ? ? ? ? L S? ? S? ? ? OD@ A AC? ? ? ? X ??? ???DA7I?S ? X?S ? ? ? ??B N ? A?L=AC? ? ?  ? ? O @ ? ? ? ? ? ? ? ? ? Z5? ? Z ?YA ? ? ? ?MO?? N? n XYS ? ?K? O ? ? ? ? I ? LN=Q ? ? ?  P?OM_?Z?I OD@ A???Z c? ? ? ?DN ??Z ? Z XYS ? N L ? ??? ? Z ? ? N?YS? O @ ? ? @ A???A ? L ? A ? N??ARX??A ? ??OYS ? ZXYA7P Z ? ?  ?YS? OD??? S) ? Z ? A?? ? ?YZ ? S? ??ZX ?) N? ? ?OY@ ? ? ? ? ? ? ? ? ? ? ? ? Z? ? ? 3Z ? A Z ??N? O @ A3?KA Z XYS )#B ? ? OYS?VN? ? ?(OD@KA ? ?=P?O Z*?=XYZ L=A?N5]??P ??YZCI?S?#A7L=? N ? X @ A ? N ? ? ??S ? A ? S? @KN ?? ? n ? ? ?YS) ? L  a ? ? ? ? ? ? X? ? E X& ? } ? E ? ? r k a ?s E k a !? o x?? ? V E k ? E k f ?  ? ?? 1 k ? ?R? ? ? ? Z ? S ?7Z ?O ?  N??? ? X @ ? ? ?C? Z ??? ? N? ?  O @ ? O ? O @ A?L ? ? ? S ) ? ? Z5?P?A OYO S ! ? ? Z ?D? A ? @ A?P A??7N LI N5? IKAC? ? Z ??Y? ?? ? A ? O N ? ? ? O?@ ? ? P ?             ! " #%$ +%,.-0/21 -436587 9*:(;0:4<>=?:@BA 3 5 ; -C 3 &('*) DFEHGI JLK6MONQPBRTSU%VQWN G KXYI ZL[T\^]_[T[T` aLbcEHdeKSV G Kf[gK6K^h RTi V RjlknmTb o_RTp ZLqsr t K^uRTI ZLqvr Rnwx N E [ RyTr RTK8z p {LK^[ | [TK^K~}F? r E?X }?? ?n?%V ?TK?qHV? ? a q?? q???t*pxL?QK6t?p ZLS4b N X?VcE X?qHp a E w a q?` a?? K RTi_KK??o4K?\6RTK^X?o4y?K^Xp ?L??RT?x NE KyTy?NyK^? K6yT?c????p ??EHRTS?K2? wx EK6V ye\6?sK?^?Y` a?? M K??s[n? qsKfR?i V R?gSK?X_I ?L[TR?]YyT?4r E4? U6[?V y?K ? K6y???q?K V E N M?K8G K6y^??RTS4Kl???K6[TRTI ??N E?RTi4V R??K z u?I ??R unU t P p { EX_K^?KEX4KER??%VQ]4[g[ w a V E?y V EX4Nq G V y?I ??V?Wt*Kn[ ??? p ??[ M i4V R%RT?U?oK?y P N ygq?rQh_?^K?N?P?[T?_? S?K^?nRTp ?? V R?Ny?[?d w?L? tB?4K?` x P R S?KH?s[ ]q?o4RT? ? N E??N E0RTSK>X? JL[F? RT]4yg? V E_?^K^[sV yTK G ` a N ??V RTK^X? N y%I ??? RTSK6ygK?V y?K ? N X4K t*t?` x E?zK6ygy N y?[sp { E0N ]y?q?NX4KtB[TN RgS V R>RTS U X w a [ R?]yT? V E4?8K^[?q??4[TRep aLE? t??XK RTSK?qHbcXK6?*t?p ?LE?%KyTyFNyT[T? D?E NR S K?yfd?N ??X4[ m?p ??[eI x R??YN [T[n?  ? ? K%RTS4r R%?6?>? ???cXp a [TR?yT? V E4?^K8[?V EX?q?N X K ?*?*I ? E?z?K6?Ty N yF?%q?r?????Kr X?R?? ?????? ? K [ RT? ? ? qHV RTp? N E?KyTy?Ny?[ ? ? ? G?w? N ?_[ t ? V E4bcE4y?N??4[gR V t*?cN yT?x R?S4q?M?N ?Y??X ? K N EK P N yfdS4?x ?8SRTi_K?r??N G K p { [ RTy ?K???r E?X?r y?N W?]unR?rQ?*z N?y?` ?LRTS_q?deN ?? X ?4KeN E4K P N ydSp Z ?nS?[qHr ? ?QX? ? [ R ?yT? r E4? K^[?? K V X?R N [q V ? ? K^[nRgI ?L? rcRTp ? N E KyTy?Ny?[ N yK? V qHo ? K?I ??ERTi?K?V?X4V o?RTp ?LG K?? ? R U yTp ? E?so4ygN W4t*K?q??_d?SK?yTK?MeK???u^? qsK^X?rQE0? Dn? ? N X4K6?F?? ?B??RT?SK0?F?^? ? q?N X K t ? V ??W_K????? ? W4?4R?d?Ksh4K6? ? K^\ R?RTS4KsR V I Z ??NQP RTS4Ks? ? R?K6y2[ w a E?6K wx R?u Ncd?K G K6y ? ? E t*K8[T[?N E K ]4[TK^[ V y?N W?_[nRfK?[nRT? a qsV RTp ? N E V ?*z Ny p??RTS4q ? p Z R \^N q??N E?K6E R [?V y?K [nq?V t ? Z?[n? q V t*?4q?N XY?6??? ` aLE ? K ygy N yeq V ? y?K6[?t*Rp ?LE?t*Vy z KK6? RT` ? ? V RgpxLN E?Ky?yTNy?[ p ??[?\8N E \6? ` a?G V ?_t?Kf?gS V ? RnSp ??[B ?L? ?fSK ? yTNc?t?K q b ? y N WY?[TReK^[nRgp { q?VR?p a?N E>` ??[??gS ]4[?V Esp ? ? o N yTR V E R.NEK ? V EXs` ZL[?deN y?R S ? N?P [nR?]4X ? p aLE I R?[HNd?E?yTp??? S?R ?V RSK6y?[T]ygoyTp Z?[p { E4? t ? ? p Z?R?S4V?X??_N R y K6?6K8I ?LG K^X q???\^S V RTR?K?E RTI ?NE ? E R wx t ? V q?K V E[ \?y?? ? RTK?y?` a N E?S ? ?4U?K E?p ? ERTy?NX?4\8K^X } r ? K6[ ? ? ] ` a ? U y?U?\^K6E?RTt?? ?? i_KHj S4I Z?[2p ?[?R S K N P [ R ?X ? p ?LE ? [T?4Z ?8S ? ]4K6[TRgI ? N E?[2` ? E0RTSK?\^N E RTK ? RT[2N?P K^[RT? ?Lq V RTp ??N E V E?X \8N ERTygN ? ? [T]4? ??K6?6R?N ? RgSK?E?K R?[TK^\?RTp ? b E x            !#"%$'&()+* y V Sp a [gK?KnX?j,p ?-  E2?TK6S?[gK?RT`? o4qvy?K V G R?Ip?L? NNE? [.PFN[gK y \ ? RTpx ??N ??E+V .? pa N E?SKepZL[ p a XV K E V V I ? [ RgRTRTK6N?qs\8oN RqHRgKBN]VQo?X_d?X4y` Z RgK S [T[ K [RgRS xw K q y r?N R ? N ? y [ [nRT?TES K?V [?R?[?q???I ? EK^I ? [qsRTp?LpN ? ? E K ? N yBp Z E?RgS Kf[g]4?N ?R p ?- q V0 / \?VQ[TK??N ?  E X ? RTS4K ? V ? p a ? ]q K6E Ky ? ? ? VQp ?LE P y?Nq R?i_UX ` ?L[TR 1_y?WrQE_\^K6[ S` ??[?d?p ??t?t(? ]YV yTV E?R?K K RTS V 2 p Z ? R?i4K?Xp { [TRg]y?W V E\6K8[2rQygK?[nqsV t*t | ` ?L? RTN?? S K2? [TRTp Z ? V RTp {?b EHK?yTy?N y?[ {  ? a K6EK6y?z ? ? R SK E2R S?KK^[ RT` ? V R?` N E?K6ygy N y??ed?` a t*t?4K ? [ ? V t?t ? o4N [T[w ? W4t*K???` ? E K E ? y ? ? ?8?4345 ?2???? ?76 8:9 ?c? ; 9=<?>A@ B ? ?F? 6DC ? 3=E < ? ? 6 < ? ??E N RnS K?S y?K d y N N W4y X ?_[ [TRnRTE S K?K [T[?q N V P ? Rnp? q S?K?? q j K E K?K6y [n? R??`??qv? VQV Rg` a N h???[?p F [?V qHy?` a?[ pZ EK [ p{ qs? y I? ?N K q X N G K6ys? ???HG 5 ?n? @ I CJ < X4I x [n??]yg?4V E??6K^[ p a E? K RTS?K6? ? V ?cK?E??K N  ? [n?4qsoRTp a N E V W b ? R2RTS U X_I x ? 2 ] y??rcE?^K6??? 2ON K6?~i V G KsRTN RTS4` ? [>PFV ??R M ? L V ?^\^Nq?N?X4V R?K P N y V ???_?6N E \^K6` a G V Wt U Xpx ?R?]y?W V E??K^[ ? V E X V y?K?R?i]_[N G K6yTP \NE[gK6y G V R p  G K ? Q KvN E4?8?q?NyFK0? ?R1 qHKsRTS V R d KNW[TK6y G K V E Nc]RTo_?R?[K6? ]_K6E\?KMSUT V WYvX RTiV R N ?_U ? [ RTS U ? [ K6t P?N t*t b d p LZ E ? E N Et*p ??E?K V yqHN 4 ? ? > V IA\]_^ V W ?O?` ?4acb7d ?e ? x ? 6 G E Z  S ? K ? y K p [  ? E  N ? d E ? ? ?  E  \ ? ? p k L N E p ? [ V ? E  ? 4 E ? ?  E N  d ? E ? d 6 K p c ?  S R 6 K 6 \ ? R ? N y V , X S d V *f ? ? V 34503=gih j 3 < P ?=` ? b V lnm ` x [2V E ? ^ E ? a E Ncd E?Xp aL?TRg?4y?W V E4?8KH?TK ? ]4K6E_?^KlRTS V R?? ZLE\?t?]4X_K6[?E?N p ??[nK V E=[pQo Ny ? N X4Kt t*I ??E??K6yTygN y?[ r Rts V Rvu? E V EKw41 yTV t E ? R?dy x yg??\6Nc?R?K? 2'z d { ?| ?~} ^ ?? d { ? ` ? ? d?S_K6y?K ? d { pZ?[RTS?K?? a RgS?pZ?E?o?R a ??K8q V tY? [WK P?N y?K?M?K?[S V ??? X4KE N R?K?W? ` d ?A???? ? >%? ? ?V R ? K6yTE ? d { ? RTS K?K6[nRTp ? q V RgK?b P RTS4K?d?K6p ? ? S R Z h > Z V E X?R S4K?o y KnXY` a \ RTI { N EHK yTyTb yW ? G K6\6??N y ? Fy[ p?ZLE ??qHK ? ? y K6q K E RT[?4o R N VcEXlp ?LE \6t?? X p ??E???Rgp Z?aq a6K ? V ? ? ? ? ? ? d W4?? ^ `??:? ^ d ? `'d B??H? Z  iK%N o4Rgp?LqHV ? j  K?[TRTI??q V RTp{LbcE?o?y?N Wt Kq ? ' V ??ENcd?WYK?[nRTVcR?K6X ? PFN t*t Ncd [ a ??? :Y@?{ ?? -=???????? 567 ?????=? j ?? 3 567 ?Y?K4? 5?7 :?; ??? :@??? -=???? @ l 3  T ? 3 j{ ?? ? 5 G?? @ ? ? ? g ??? ? {? ? ? ?2? ? ? 5A3 ? ? 6 ?? ?0? ` V W ? } > ? ? > ? ? Z { a ? > V I ? ; 9 ?0>; ? @ I 3 @ ? ? @ ? ??< ?? 9A? ???7? @ ??? ? 3 ?7F6 ? ? ? ? @ 3 H ? 6 5 ? ? 9=< T @ j ??<F; ? 6 ^C ? 3?? ? ]? ? `??? |?? ? 3%?_SU?7V W X ? 5 ? 9 <????> h I ? ? @ I 5 3 < 6?6 5 6 ??? < V Bv?? z V I ? ` ? ? f^ ? ? `?V ??? ? ? X0?? ? 3 T 5 C ?F?h ? 3 ? 9%< 6 < ?6? g ?; @ { 3 ? V { ???? { ? `???? ? | ? k k{ V ??? ^ ? ? ? ^ ? j V { ? jn? ? ? k ? E I [ ? 1 ? ? o  ? ? ? ? ? ? * p a E [ ? ? o  ? ? ? ? ? A ? ? ? ? P ? ? ? ? ? ? ? ?A???????? ??P ?  k ?0? |K? `?? ?? ky?c?n? ? ??? d f ? ? ? ? ??? `??c` ? ? k a ? l  ? ? b ? { k ?? ? ?          !#" $&%')( *+,- . / -  #$10 2  3$547698:<; =   > #$ ? >@ ACBED FGHJIKH7LNMPOQIKRTS HKUWVYX[Z+\ I^] _#I^` ab#c_ F)d HKe>f Hg_KhiG _ Fkj d _ X^l+m ` n X+o _ V G lqp r cEs oTp t Z d l<X o ] u Zv H mwxzy Z c{ F}| l IKH~Ei` n X o ~ l X?\Z>U H?_ h Z dY? j Z ?X<Z d X^? ?#Z I l<? X ? u u Z? d H X^l ?W? l c U H X<??C?Y?W? p ? XV GH X _>??j Z d d H {+u p r c p ? u ? u???? c _#I^?i? ??#?????K?E??? ? ?i????E?+?E???? ???Y?? ??&? ? ?? ?^? ?>? ?W? ??1? j Z?X^Z d ?^???????+? ??? ?????q???? ?K?C??9??g?K?? ? ? ?^? ? ??? ?? ?? ? ?+? ?K??E??? ???9? ???<? ? ?K? ? ? ?z? ?? ?g? ? ??????&? ?W???<? ?K?#? ??? ? ? ? ? ?K? ? ??? ? ? ? ? ??? ?W?#??&????? ? ???? ? ??? ? ? ? ? ?K?7? ?#??? ? ?? ? ?7? ? ? ? ?E? ? ? ? ???? ? ? ? ? ? ? ? ?#? ? ? ? ? ? ?C? ?^? ? ?K?E? ?<? ?+???? ? ?#? ?1? ? ? ?W?? ? ?E?? ?K??? ? ????? ? ? ?Y??? ? ? ?5? ?g?K?E? ?? ?? ? ?E? ? ???1?W? ? ? ? ? ??? ?1? ? ?? ? ? ? ? ? ? ?? ? ? ?? ????? ? ? ??#? ? ? ?<? ??? ??}? ??#?? ? ? ? ? ? ? ?#? ?E? ??? ? ? ?<???9?z?<???>?9?1? ??? ? ? ??? ??? ? ? ?E???<???i?? ???K? ? ? ? ? ?? ? ??? ?K? ? ?#? ??? ? ? ? ?K? ? ? ? ???E?^? ? ? ? ?  ?? ? ? ? ? ?? ? ? ? ?>?9? ? ?^ ? ??? ?[?? ? ??? ?<? ?? ?#? ? ?<??? ?W??? ? ? ? ? ? ? ? ?^?  ? ? ?K? ?^? ? ? ?? ? ? ?? ? ?>??? ? ?^?? ?K?K? ??? ? ? ? ? ? ? ? ? ???i?E? ? ? ??? ? ?<?? ? ? K? ? ? ?<????? ? ? ? ? ?? ?K? ?? ? ? ? ???&? ? ? ? ? ? ? ? ? ? ?  ? ?i? ?<??? ? ? ? ?&?? ?K?? ? ? ? ?<? ?? ? ?)? ? ? ? ? ? ?<?? ? K? ?i?<? ? ??? ?+? ?<?  ? ?? ?  a  l ! " ` n o ` # c f $ 2   ( $./10 ? ?% ? ???? ?  a & ? ')( * ,  ? + (- ? 354768:9 6;=<?>^? A ? @CBEDF GIHKJLNM LPO 8 9 F QSR 4 ;1T)U _c X p a V#H I V G l u7_ V l d WYX[Z F G l I l V GH ?#? \ ? ? ? IZ ?H ?z_ ? c { H?\ { Z c e | Z ] H[? _ ? c e H?, { ^ I `X _ Z c{ X<H?j _ cE{ _I e H I V#HJI^` a a Z oTp b c H^X y?Z c e X ?W\ \E_ X H F H F ` ? X G o _+u ` c p d ??` e f H o GH ? Z ! p r C ? g u HWc H<IY f h?fZ ` i cih IK_u o GH ?#cbcE_ F?c Xk m j m m lniZ c e  ]  \po _ q G H \ IKH e ` s " _ ` i _ cqHWI^IK_uI tmv H ? s o N? w h Vr ? ? | ? CO x L x ?? ?? s~? ?y ?}?{? z ?? ?? m ??  ? ?} ? ? o  ? Z c eqo GH ?} ? m?? Z I l H !>U s o ` s cf y?V G R c V GH u7` s cE` ? u ? u ??? c_I<u p a X ?#?T? ? ???? ? ? c o G` a X "KZ>X H Z c _ \o ` s u Z d ? ??H X<o ` ? u Z o _#I?` s X f ` ? c H ??? h o G#H h^_ d d _ F ` a cf X HK? ? HJcj^H_Kh c _#c ? p $ cEH Z I l??>?#ZV ` a _ c X  n ?? m + ?? ? e +N? ? + P? ?m? n????? ? }I? ? m + ? ?? n m ?? n m  n??? ? e + j ? + ? m?? n ?K? ? ? } ? ? m n ??? ? e n?j ? n ? m + ? ? ? ? }P? ????? ? ??? ??? ? r ?? ? { +N? ? + ? m?? n ???? ???? ? ? m ? s ?7??? e n?? ? n ? m +`? ? ? ? ? m  a ? ? } ?? ? ? m  \&? ? ?  ?  ? Y ? ? ? ? ? ? ? n ) ? ? m ?C? ? ?W? ? ? d ? ?? ?? ? ? ?#? ? ?C6 ; J 9 ??? ? ?? ?  ? ? ? ? ? ?? ? ? ? ? ?  ? ? m ?N?? ?  a??P? ? ? ?7n ? ?? ? ? ? ?? ?^? ???E? ???? ? ????E? ? ? ? ? ? ?? ? ? ?? ? ? ?9???K? ? ? ?#? ? X o I ? j o ? IKH ? ???9? ? ?? ? ? ? ` ? ? H ? ?K? ??? ? ? ???????? ? ? ? ????W?? ? ?#? ? ?<? ??? ? ? ? ? ??? ? ? ? m ?P? ? ??? ?#? ? ???<? ? ??K? ?? ?<????1??? ? ?? ?????  ? ?K??? ?. ?<? ? ? ?#?E? ? ? ?? ?? ?  ? ? ? ??? ? ? ? ? ? ? ? ? ?K?? ??? ? ? ?. ? ?? ? ?? ? ? ?? ? ?? ? ?E?  ?? ?? ?E? ?? ? ?5? ?W??? ???? ? ? ? ? ? ?K? ? ? ?#??1? ? ? ? ?#? ?^? ?<? ? ???9?>? ? ? ? ? - ?>? ? ? ?m? ? ?1?<?K?  ? ? ? ?? ? ?Y? ?? ? ?E? ? ? ?E?? ?K? ??? ? ? ?? ?Y? ? ?9??? ?? ?^?^? ] ? ? ???K? ? ? ] ? H- ? ? ? ?<? ? ? ??? ? ? ? ????? ? ?% ?i? ???9? ? ? ?9? ? ? ? ? ?#? ???? ? ?K??K?? ??? ?<? ? ?K? ? ??#? ? ? a ? ??? ? ? ? ?K? n ?K???? ? ?? ?E? sS? ? ? ? ? ? ? ? ?K? ?>? ? ? ? ?? ? ? ? ? ??? ? ? ? ?????? ? ? ?W? ? ? ? ? ?K? ? ? ?? ? ? ?+?<? ? ??? ???? ? ? ? ? ? ???? ? ? ? ? ? ? ? ?? ?<?<? ? ? - ?   ? ??? ?? ? ? ? ? ? ?K? ?3?E? ? ???. ? ? ?. ?#? ?<??? ? ??? ?????? ? ? ? ?J? ? ?#?^?^? ? ?K? ? ? ? ?<? ? ? ? ? ?K? ? ? ? ? ? ? ? ?K? ? ? ?<?^? ?  ? ????? ? ? ? ? ?? ?K?K? ? ?#?? ? ? ? ? ? ?#?K? ? ??<? ? ? ? ?+? ??? ? ? ? ?Y?????? ?  ? ? ? ??? ? ? ? ??? ? ?? ? ? ?E???<? ? ? ?9? ? ? ? ?K? ??? ? ? ? ? ?? ? ? L ?K? ? ?)? ?:? ? ??? ?^?<? ? ? ? ? ? ?[? ?+? ? ?>????? ?<?3?9? ?1? ? ????E? L ? ?  ? ? ????<? ? ?&? ? ? ? ? ?<? ? ? ?p? ?? ?? ?? ? ?-{?K? ?? ? ?#? ? ?. ??? ?K? ? ? L ? ? ??? ?r????? ? ? ? ?E? ? ? ? ? ? ? ? ?>? ?? ?? ? ? ?#??? ?K? ? ??? ???^? ?#??? ? ? ?1? K ? ??? ? ? ? ? ?K???? ? ?5? ?? ? ?>? ?^?? ?E? ? ?^? ??? ?K? ?K? ? ?)?^?????? ??E? ? ? ? ? ? ?[?K? ? ?  ?#? ? ? ? ??? ?E? ?#?E? ? ?? ?? ? ? ?9? ??? ? ???E?? ?K?1? ?. ? ?K? ????? ? ? ? ? ? ? >? ? ? ?? ?? ?? ?K? ?g? ? ? ??? ? ? ? ?            ! " # $%& ')( * +,.-0/ +( 1%2 3546798;: 4=<02?>A@BC04ED=FGH ( IKJLG0MN22POG( Q%RSHT:VU W X Y 2 J 2?:[Z T 4]\^_`+ 2 H /N4/ _ (  J a 2 H <2 b cdHG ( I 4 J o T MdH Tfe g J + 4 hi+2[\ H _`_j <02?k e I 4 l;\ 2 OT ( I R H G02 O UfmnPo B UqW rKsut v 8 e  \ > : H /O G8H wxT 8;2V2[O T ( > H=y 4 <9z@@ {x(  \ ( I / |}/ ( I T 2~?d( ?%R.:[/\(  4 /H? G8 4 c*??p 8.G8( ?KO > H?? j <04 8( I?? ( p G } p 9 ? ? j <H??G( IK??H H j?_ (  ?PHG( ? 4=J \ 4=? T MN( ? \?H _ ? ^ <?( ?KG? > B( ?KG 3f( K? h ? 2 + 2?<???? \?2[? ? _ T ^ \ w0? N??\?N2 a ? e  H? ? H \2P? \?c J 2 < Y?8( 1 ?8?G? 2?2 OG?  R?H T 4 < ? 2 ?[4=>?2[\x? J e  T 2[?  ( ?%RL2?/\( ? ^ J H ? B4?9T 4?| J Z ?J(? G 2?  (  > 2 /\(  4 / H _ H jj < 4? e  RSH G e ? 4 J \??4 <??@@??? v 8;(  \ Y ( * _?_ ? 2?Z 4 / 2L( I / T 8 : J 2 ? T \2 ?[G e  4 J 1 ? ?f?}? ??}???????? ???? ??!????? ?????;?S???? ? ????? ??J G82 h ( I J 2?H< ???S2 T 82?R 4 Z 2 ? Y 2??[4 J \( ? d2 <f8dH \ ? W r ??? ??????;?W r U?? \ ^ T?HG? ?P?? ?0UV????5?? r _ G?8;4 ? ?M.G8;2V?`( *%? 2 H <?6 ? /? T e ? 4 J]? ? ?`z U { ?? ?;?W r U Z ^ 2 \ J 4 T OH? e I O???? G 8: ? 4 c J 2  J;2 \\9?[4 /( * T?(I? 4 ? JE4 ? 78;2 4 <?? >??? ??:?wq? Ox( 1 J;k :[OT( * ? HT 2 T 8 : ? ^J O:?N? 2 J ? : 4 6}H ?N j _ ?;( ? / ? H _ ? 4 <?( ?%G? >?z@ ??? G 4 T08N( I \ ???L2?x7f8=cO?G?? 2L?)? [? ???0w0??2[??H T (? 4 J (  J @=@ { ? 2 ?[4=>?:O  @ ? r  '  m  m  U  { ?}m z ??  Ufm ,  ?  ?  ? z  r V? ?W  U W X "! { ? W r$# ?&%  W r(' ? ?W r U W )$*  { ? W r   U W I ? + , -l Gq? < 4;R @ T/. 210 2 G8?: b ?;H G( g 4 J  U W 3 4 657 K ?8 m:9 ? ? m U;  ? ?m=< z  :> ? ?  U m ? ??@<BA \ ^ w08NHG ??? 8=W $CDFE ? ? ? ;  U ? G$HI ? ? W r J  < 7 @ UxW r$4  < z =W PRQ  ? SW X ? W r$T D ? ? W PVU ? K MXWY 3x? e  ?81Z IK\xG082 [\^]?H _ ? ^ <( 1%G?MR 8z _ a? ` v 8cbdq(  J T 8 2 _ ( I%? : H <! ? HO : G0MN2?:[\ w0e I%R?H=w 4 < 4?}79?;: e <:fhg ` \ j 2 ? e 1 H _ ( 1ji 2?\?G ^ G8;2lk"\m]?H ? ?4 <?( g G8R v 8 (  \(  \ : ? j2 ?PT:[Z \( * ?? : 3?2fM H k:V\8 ^3 / (  J zV n HOO e Fpo  2PG H _  B ?q q _ { G8H G G08 2 [rs] H _ut ^ <0( I T 8 > e  \wv^x ^j C/o  R H ?  v 8:]<0:[O c _ T ^ ? G?H( ? / :  T MN: <:?(  O H \ ? ^R?y _`^ Y\ L, M U ? N7O 1 { ? ?|~}? ? z ? ?=???????8? {"??? ??? ? |????:?R???? r ?X?????c??? ? ? ????/?X?:? ?  ? ?????? ?&???&? 0 ) ??? ? ? r ? ? 0 0 ? ?10 ? ? ????? ??? ??? 0 N ? ? ? ? ? ???8?R? ? ? 0 ? ??? ? ? ??? ? ? ???R?p? ? ? ? U?? U ? ?? ? ?? W r1? ? ? ?p? ??? ?/ r ? ? ) ???? ? ? ? ? ? ? ? ?  ? ? ? ? ? c ? ? ) ? ? r ) ?? ? ? ? Xu? ? ?? 0 0 ?@??? ? ? ? W ? ? ?/?l? ? 0 ? 0 ? ? ?&? K ? ? ? ?? ? ?c? ? ? ? ??? 0 ? ? 0 ) ? ? ? v x p ? ? ? ? ?  ? 0?? ? ? ? l ? ??^? ????? ? 0 ?? ?? ? U W X?7 (  /rW 6@? @ ? ? r ! ?????X?:?7 ?? ? ??? 0) R@ ? ? ? 0? ? ? ? ? ? 0? ? ? ? 0? ? ? ? ? ?? K ? ? U ? r ? t ? , ? W ru?  W 3?? ? ?W r U W X 4?D ? ? @?? ? ? ? ? ?R? ? ? ? 0 ? ? ?????^? X0 ? ? B ? ? ? 0 ? ? ?? ? ? 0 ) ? U?@ ? @??? ? 4G2?G 8HGV(  J G 8;2 _ (  /;2PH < P? H\2]T8 2?2[\ T (?%> H T 4??o  Of|/d( *%G:???? ( ? RS:[/ \ e 1K4/ H _ O ( * N / ? : T 4L?N/  UqW ? Y: 4 J ? ? <?2[b? e  ? 2 J;^3 ? 2[d?2 4 fUxW r 4  ? }  ?  !? { ? ? ? ? ?  ? \>?2 J T ??4 2[ HT G8 : :[/Z 46 ] :PT( * 4 J T8: v ^ ? Te I?H ? 2[\ T ( Q%> ?( ? ? | ( I"?: ? Zd( I%R J J 2 J \( ?K4 J H ? B ( * J T ?;2fO?: /O:9??H Gw ^?^ ? T?H ( I J C8:$# O T  ?&%?H C?2(' : \ C( * > HG: O U m*) BU W r,+ D - : %?H0/ ^ FNTH( 1 / |/N( ?KT 23254( * >S:[/ \ ( I 4 /H ? H ? ?[4 J OT <0. 2[\ G( ? > H G04 < z @@ { I?? I??? cN? G e 2 / ? H ? j < 4 ? e ? >?HG e  ^ /;O T e?T8N: / ^ / ? ( g J 2 H H w 4 < Wr q 3 2 j j <e <?: b ? ^ 6 M : e ?2 > J [2 2??H ??? ? <: + ? Z I f?HC e I 4 /O?T e H T ( ? 4J;\ H jj 2PH ?( ? ! (? ^ ? ! Co 68 # <eJ ?         "!#$  &% '( )* )%+, - #/. +  0  1 ! 32&!54 "687  # ! 9 ":; "2 < !  =>? @A&BDCFE"GIH8JKMLNPO,QO,R,STVUWOXR$OZYV[]i \_^V`a Bcbi O,Y TV`-&d3e TVLR$YfM` hgji BWL g L"Blk O,mnoOqpsr LWtWTu`v&w e \"`a&B e k \xL"y e i Y O B YVO"z|{l} T N(~ i_i? B \??3U e Q"O L? e L t T$? &d3e \ ~ A&? ? ? UcO d O TVUWL??TVU e ?Xf O??uU e \_\ } Y O?T L?L yoT e `A B Y } ??U e tct mVL"?` d3e TVOMO?YVTV` d e TVLm$Y`???T L?? YV? d O TVU e TN O U e QO??L)? B ?sTVUO O Y T `? d?e ? O??? ? ????9? B ??- T L } YuO `? T e Y e B ` B `  TV` ? ?)\ k? e O Y Y e TVL?Y L \ Q O T UWO C? ,? EI? T???O8? } e ?$`  L"??? ? ??? E"E,?Xf U L YVO ?VL"? } ?$` ? L"??` ???M? ? 1?? OZt]OI?"?]` ? BWk?L B NMU ? \ k"LmV?? ?VUWw N(? } YVO?T L Y L"\_Q)OTVUlO C?? ? e? E ? H ? UsO ? ? e TV`  L)? N ? ? T$??`A B `A ? `1?e \ k ?oOZY Y ? ? ? ?P? ? N?O?YuU e \ ? k O ? e ?]`- ? O?m$O BW? e tot m L ?`A w e T$OO,YuTV` ? d ? T L ? E"E ??? ? L??VUW? ? Y O B ? S Y } tot L YVOT$? e T?N?OU e QOY L \_Q)OZ??TFUWOX? ? ? U?O ? } e T$?v L B `v B C?? E ? e B ??? e Q ?(L y9T e `? B O ? ? ? ? ? ? S ??? ?j?,? ? ? ?? ? E ? ? ? ? ??? C??c??????"C ?X? ?8? H,? ? ?,? ? C ? ? ? ? ? ? ? ? ? ? , ? ? C C ? ?"? ? C ? ? ? ? ? ?F? ? ? ? ? c ? ? ? ?  ? ? ? ? ? ? | ? ? ?? ?I? ? ? ? O B L"N?` B TuO B ??T L Y L \j?IO TVU O C? ?j? E,? TuU?O ? [ e Tu`  LB ? ?&B C E E ? ??Lm??? ? P? L T$OT U e ?PTVUl`A Y?O ? ? e T$? ?&L B `? Y L)?9?VUWO?? L m d??????? C ? ??? NMUcOIm$O ??e ?? ? ? ? ? ?- ? ? N ? ? ? YVO L"?oOYVTVO t L Tue UWO  ?   ? ?  ?   ?   ?   d O,TuU L ? ?  ?  ?  C !? ? ? N `?&TVU?`? BW~ ? ? ` ? #\ "ZL"?c?9` TV` L ? %? ?$ ? ?  ? '? &9N O U QO ? 5 - ? C?? ? 7 8 0 ( ? ? *9? C ?:?#; ? ?< ? ?>? = ? ? ? ? ?? ? ??@ E'A ? ? ???W?- ?. ?%/10 ?$H'2 ? ? 1? 3 ? ? ?4 0 ? 6 ( ) ? ? *,+ ? ? j ? ? ? g h i ? f BDCFEHGJI K ? CML ? ?>N ? ? ?_C ? ? ?O ?8?V?'P ? ?!Q R C ?S:?? ? C ? ?? )UT ?  ? ? C E'a ? VW  X ? ? ? ?ZY ? C L ? ?\[ ? ? ? C ? ? ? ??U]^]'_ ? ?? C ? ? ? ` ? ? ? N ?o`? ?b?`v?YTV?lO?n e ?Uctlm$Lt e k e Tu`v LB e \ k)Lm ` ?VU d MC d J ]??fe L T$O?eTuU e e T?YV` B  ?8O N e O  B \ ? ?lYVO?? ? ?g ? T e L e i B `? T$UO k ?9` d O B YV` L ? \ tlt m ?"` ? w Tlv L B T$L?TVUoO k \ L y \ ? L d t } T ? ? ? h ? n e ? c t m L t e k e TV`A&L B `? Y j e ? d e d pnm L t To \|OZYFTV`p ?$L"m C E ? } O?T L?`J ? Y e q t m?L?"`v d e TVO B e T$?9mVO ? n e ?Ur t m L q9e k e TF`? L"??U e ?  B \ ?ts \u  s p m L v ? `? d ?;\_` ? ? ? ) t m L? t9O,mVTu`w O,Y ? e Y N O?U e Q O?Y U L NM??`J B CK e YVYV? - nl`- O ? e \ ? ?Uz A { UoO?m$OZY } \_T?` | ???uT e TVO8? y O i Lf S N?UWO,mVO ? UcO3? L \ } d ? QOZ?,T  m$n Y } ? :~ e m ? ? e \x\ ?U ???? ? ? ) x!? x> y ???? ?  u ? ?  ? ? >? `? ? \_` ??? ? ? m?? ?? ??? ? ? ???? ? ?  ? ? ? ?????1?J?  m?  w?O ? ??? ? ????D???J?????U??? ??? ? p m????Z?'? ? ? ?? ? ?????1???  ?Z?  ? ? ? ?c? ? ? ? s???U?     ? ???U ??  U??? ? ??? > ? ?,?  ?  ? ?W? ? ? ??C ? C ? ? ? ? ? ? >??? ? ? ? ??Z? ? ?U???   s ? ?  ?I ? ? ? ? ? ? ?>? ?   L ? ?   ? ?  ? ?  ? ?  ? ? ?'?M? ?   ?   ?U???? ? ? ? ?  ? ? ? ?  ? ? -?? ???   ?  u ??? ?6???  ? ?V??? )? ? ? ? ? ; ?6? ?U? uU? ?   >? ? ?? ? ? %? ? ? ?>? ? ?   ?  s s#????????? ? ?  ?n? ? ? ? ? ? ? ; ? ? ? ?  ? ? ??? ? ? ??; ? ? k? ? ? ? ? ? ? ? ? l ? ? ?U????? ? ; E ?  L ??O$? e ? ? L ?WTVm e m$? ?$L ? U ? k \ L y e \ ? ?oO Lm?O,w ? y e ? t m  t e k e T$? z L B " e B ?  ? e ?ZUc` ? ? ?O   E e B??? U e T?`? ? y L ? BJ ? UWOO BW? m k ? k e ~ ? ??y? ) E ? ?  B \ ? ?L ? ? ?  s?s ?J ? ? ?  `A ? ? ?9m?y ?D" ? ? ? "! ? #%?$ ?D??'&)('* ? ?+ ???,# ?.- ??? ? /?? ? ? e  YuTVO t L0 ?WO2]1 ?"O ?43 t L`A ?o?`? T$O ??e ? `? LB5T$L Y L?_Q O ?  m??: ? e Y N e Y/?  B O ? ? ?5 ?|O,`?,? B TuY`A&L ? O ? ? L ? }f YVO ` ? B } k Y ? ?oL O(B O5 Y $ O t L> TVUWO ?27 NM?$L B k 8 e t UcY L"B d O,TFU L"?sfM`- TuU?`J B `? TV`? e i ? LB? ? A T `? L"? y 6 e ? ? ?9 ?W: fO L ylT ~? B??<; O ?L>_= ?_L"N `A B k e i k LmV`?@? U d e Y e B e t q m$L"?`? d?e TV` ? LB??$L C X ? ? A ? ? ?C?B ? ? ? ` ? ? ? C?? ? ?E? ?L ? D ?"? ? C ? ? ?F ? ?V?HMG ? ? ? ? C<IKJ ? ? ? 9? ? ? C EON ? ? M I ? G C ? ^ ? ] ?  ? U ? ;  ? C ? C V ?  ? V ? C ? P ? ? ? ? ; ? ? G ? ? G ? Q ?R ? ? ?S ?)T ? ? ? ?W XZY\[ M ? ? a _ Y n ? ??Lm$O S?C E N ? ? e ?  B \ ? \  ? e ?/L t ??? w e \ ?a ? ? q m L q ? mVT?`?&? Ya` KL)N O Q 7 m S Y ` ?o?8O ? U ? ? O N T  B 3 8 e t ?oY L ? d O,TVU L ?no ? Y i ? ??Yq? ? }  OcVb U e ? ? ? ? |1 ?O ed t L"` ? ? ? ? p T ? m e ? `?  B ? `1 ? `A Y(OI? q O,? T ? ? ?$L?U e ? ?                 "! #$&%%'$&(*),+-/. 0213(465 $&( %'7 89$;:<%'=. >?# .@BA 5 (4 5 . C . %'7 D94 >FEHG = $I-+J 3 5 K $ML N O9%'PI+ Q%R$S. 0,T +('U V9% =3W N X9: a #3. - A 1 e +h5i C . g 7 X +> b R"7  -BR ('$&jk N O ($B: YIZ\[^]&_ 5$/(`N V % $&(. %N  + aSb = 7 D -BR 7 O :c+Q @ +6d"e :$ fR 7 D T6R3$ eg = . I +6> 0 PSYIZl[nm5 $&( N  %'$&(. %N V9+ a V o prq`sut^vwyxz{'q`| } $ =3.~ $H3$ ('U ? ~ $???,? ?B???^???? E +653%'7 ?9J . 0 $&:%7 ?9W . %+6(':?Q + ( g (. ? 89>7 O9a T >$ k e . ? >$&%'??+ ('? : ? ?6? -BR ????+653%N O9W?. ? . 0?C + ( 7 ?9g R3W?: b ? 0?0 #$SW?+ :%y. 535? U ?9-&. #?),$ 7  a ? >3-B$ ( % .h?X9>?$&a ~ U ?,('+6>3W?$ a %: b R$&( $ %R $&( $<WS. P??3$<W +  $ ) 0 N 89> C $ (( +6(': c . >6 b R$ ('$<%R$?:?%?%U ?9: g 7 ? @ : . ? ??+6( U V9:%('7 ?,?3k g U X +6>:+Q % = $ ^?V9:?? ('#. > -B$ :?.($<> + % A > + b >?Z + (. ( $H%46+ $ L6? $&>:N O ~ $<%+*+6#3%? N ?9>??  G R$ ( $B:? 0 %N V > C ? ??+653% ? 89WS. 0. 0 C +6E ('N X % R?J 4 Q? R $&4 ('$MW??y7 V :7 ?9>? >N ? % $B?\N O WS$&> :7 O9+ >". 0'? : +?% R3. % @ 46WI1 ?%N V > C g R$ WS+: % ('$ @ $&>% b $B7 C R6% ~ $&-&% +( $&:%'7 X J?. % $?($ jk67  ( $B: A > + ? 0 $B? C ? + E Q`. ??? 53('$ ~ N 8 ? +6k: b $&N ? C R%$&: g ? ?9WS. %'$&: } $ ?&+ a3:N ? $&( $; % b +<? > 7 O %'?? U ? W ? a:? O +6a . 0 . 535( + L6N X J ? % N + >:`% + %RN  : $B:%? 8 J??% + (?? + >$ + Q b = E D? O9-?R b? ?.-? 5 (+ 53. C . % N O94 > ? b N V %R?%R$?53('+65?$&(%'PS% = . %-&4ha:% e ?-&% N X > C %?$ J + : %?($ -&$&>6% b $ 7 C R"% $&:% N V W?. %$I( $&j6?37 89($ +6a30,P % R$SN ? J W $N V . %$&0?P?53('$&-B$&??U O9> C b $ 7 D C R% + b $ ~ $&( ?`% = $?$&: %'N  JS. %'+6(?+ Q G = $ 46( $&W?? R ? . ~ $&(?P U ?,> %$&($B:% 7 ?,> C :% (k- g ? ('$ $B:%N 8 JS. %? %R3. % :R+ ? ??0  ? . 0?0 + b Q\+6(?. b ? 8 3$ ~ . ('7  $ %?P + Qy. 5 5( + LN  J . %N O + > : ? :+ W?$ + ?yb R3N ? @ R?WS.P?P ??9$ 0  . 0 % $ ( a3. %U O ~ $ : %+S% = $?# . - A 5( + 53. C . %N X + a?. 0 C +( 7  %R3W >?53.6('?N  ?B? 0?. ( ? 7 D % b + ? 0 ?#?$<7 89>%'$ ('$ :% ? 7 8 > C % + :% ? ?P % ?$S1+ ::N 89??0 ??,%\P?+ Q -B+6> :?% e'?3-&%? V9> C $&:%X?? X?W ? . % +6(': b = $ ( $S?5 . % N  ? C?%?R $ b $&7  C R% $B: g N ? JI. %'$B:y(?$ j k ?X e $B: W +6(?$y%R3. > + >$I? # k3%+> ) P ?>N V9%?$ 0 P WS. > P?5 e ?&~ U O + ?3: $B:%N  ? ? %'$&: ? G R$ $B:?7 X W . % +6( : @ +6>: %(k -&%$B N X > g = ? O :`5 . 5$&(2? : $  53(?$ N X -&%'7  + >I$ (( + (`. : % = E $&N V9(`- ( N ??%$&( 7 894 > .6>6 :R34? 0 *% R$&(?$ Q +6($yR3. ~ $ C ++6 C $&> $ (. 0?N D9? . %N  +6><1 ( + 13$ (?%N O $B: ? %2N O : . ? :?+y5+ ::N #0 $y%'+ - + > :%( k- ? :7 ? JSN 8 0?. ($&:%'7 8 W . %'+6(':`k:N O9> C?? )?% ? ('$BS+ (?:?WI++6%? 7 D > C $ ((?+ (`. :? %R$ - (N D % $&(N X + > c % R3+ ? C? %R7 ? : ? ? a +6%  + >$ ?k $?% + 0?. - A + Q :53. -B$  ?I? ?F?F??<?l????? ?3? ? ?2?&? ? R U ?9: ?46( A ? . :I:? 5 546('%?$B?? 89>?53. (%*# P %R ?r? N ?9(*??4 (-B$????-B$?+ ? ? - U  $ >6%N X ?-??$B:$&. ( - RF? ? U ? (I?3+6('-&$ ? P : %$&WI:??+6W?WS. > ?k> ? (??`+6>%( . -B% ??? ????  . >  #6P g R $ ? ( J $B:$ . -BR ? &- $ ? >6 $&( - +6>6%( .?&%? ?   ?     ?  ! " ? ?$#? ? %  ? ?& .' 0 .h> % ? } O)( $*'?% + ? O Z + , m.- +6> 0 NO >$ . (`??? -B+ >%( + 0 %R"$&+6('P Q +6(`:%. ? ',$?5 0 . > % : O / ? 0214365 ?7 098 ??:<; = >73??@? :BA ?C0EDF ? ?4G ??IH   ? H <J & X ( ? :N  #3N ? ? ? O( K .P ? n??. > ML X ?. 7?,) . %R VON ?  P Qy? ? + 5 % U 89W?. 0 NV %\PI- ('N  %'$ (?V . Ql+( SRUT . a #3. @BA 5(?+ 53. C . %'7 D + > G + . 5 53$ . ( N X >WVXY?74Z D ?[; = 7]\ D^)_ ? ?a`7cb ? _ F<? 0Ed e ? 7gf_?ChjiC?? d e 7> : ACk 0 D F? c4`l + 0 O c4S m + ( C V . >  ?. k6Q\WS. > > ? K ?4( ? P A N? > 8 N ?  & m Von ?jpBqE; r Y is d t ?20 D _ u 1 D ? _ A ?wv ($B>%?7 X ? @ $ ( . 0?0 c4x > C ' $ b + + zy 0 NVI{ : c - % X  ? R ?S . C65". ) O Z   & m ? N )?%$&(??,> C?? ? ?:WS+6+6%R3NO9> C 7 X9>?. > ? ? v:$B8 %% ?v N? >XSC |a} `9? ? ( ?C + ? > $ u<A 8 ? ?e 7 ? ? >~ ? ?7 Va^ 0l? F ?0Ed ? h 5 ?7 0 _ ??2?? ?.  )? ??? ?  ? ??? ? D  ? x? ??W $ 0,= . (% ? % ? $? mO? y )?$&)?)?? > ? .h> % =3$ v?? v ?$ :$ . ('? }]? (46d 5 8 Z  ? f ? _ ? ? ? i&? ? d ? ?jq _ d = ? ^)0Ei?apB?  hiC?j? d = 7 >O? O i??jp<???_?0 d = ?7 ? d t 7z0 1)i F d r h _ ? k 0 _C^ hq ^ _i ? bo? ? > 7 d r q d = ? 7 ?`? .hW*# ('U?9 C $ c m ? :  ?O?? ?9? v ('$B:: ? & O } N? ( + b . > m X)? ? ( +? c % ( O Z ?  _ ?  ? 5 g NX@? $ : ?7 D9% @ R3??,> C - NV ( @ kNV % :  `9?o?????w? 5? \ 5 ? 7 Y?a? D h 3 ? v % X P EX ? ?  ?P E ? . J $C?  Z? < ??$&$B ? . - A 4 53% 7 X9W . 0 ? $&> :N? %7? ~ N  % P? +63$ 0 1 ( $Q\?&(??&a @ $S%(. a : Ql+"(JS. %N8 +>F? W ?3X) ? %'7 ? 5 0 7 ? -&. g 7 O ~ $ ??&J N ? >+6( J :y. a . 553(+ Lh7 ?9W . % ? N ? > ~ $ X?(?:? $B: ` ? ????? _? 7)? 3 ?7gV ^)0 ? F<? 0 d t h O 5 ? 740E_ ? ? ? ? ? ? ?X?? <?  ,  ?
905 |@word d2:1 cml:1 r:3 pg:1 tr:3 ytn:1 mxt:1 n8:1 xb0:1 dzp:1 amp:2 xz0:2 ka:9 od:8 bd:2 tot:2 wx:3 gv:2 yiz:1 v:1 yr:2 v1r:2 nq:2 rts:9 fpr:1 tvl:3 tvo:1 c6:1 dn:2 vxw:1 ik:2 nrtp:2 tuy:1 x0:2 xz:7 rvu:1 uz:1 ol:1 trg:2 pf:1 mek:1 kg:2 sut:1 tvg:1 ehk:2 acbed:1 otp:2 xd:2 k2:2 zl:5 dfu:1 uo:1 pfe:1 yn:3 sd:1 xv:1 esp:1 io:1 qmr:1 wyk:1 niz:1 kml:1 acl:1 au:1 k:1 uoo:1 r4:1 co:2 nol:1 unu:1 kfi:1 uy:1 yj:1 vu:3 eml:1 w4:2 pnm:1 oqp:2 pxl:1 py:4 dz:11 zfe:1 yt:5 go:1 ke:3 q:2 d5:1 u6:1 dw:1 fkj:1 dzd:3 pt:2 rty:2 pa:1 std:1 yvo:6 p5:1 rq:1 byz:1 pd:1 ui:2 po:2 k0:1 x4k:5 unr:1 cod:1 h0:1 vyx:1 s:1 gi:1 klu:1 ip:1 kxt:2 gq:1 tu:2 tvu:5 oz:2 kh:3 ky:2 qr:1 p:2 r1:5 dsf:1 tk:6 tct:1 oo:2 ac:3 qso:1 ij:1 op:2 vfe:5 uu:2 nof:1 qd:1 vc:1 vx:3 ytr:2 hx:1 ao:1 asb:1 ic:1 cb:5 k3:1 bj:1 mo:3 r3n:1 wi1:1 gihkj:1 iw:1 yr3:1 vz:1 tf:1 qv:2 g04:1 dfv:1 ck:1 nmu:1 pn:2 ej:1 tte:1 yo:2 d50:2 vk:1 fq:2 hk:1 ylt:1 am:4 rgp:1 j_a:1 nn:1 sb:1 vl:1 lbce:1 w:10 jq:1 fhg:1 k6:17 uc:1 x4:1 kw:1 knx:1 np:1 pyp:4 bw:4 ab:1 hyp:3 nl:1 hg:2 xb:1 kt:1 xy:9 dzg:1 vuw:1 dev:1 ar:1 tp:1 ytp:3 lqp:1 gr:1 acb:1 kn:4 sv:1 gd:1 rqs:1 yl:1 e_:3 iy:1 yg:3 d9:2 b8:1 von:2 nm:3 x9:3 hn:1 ek:1 yp:1 f6:1 de:8 rsf:1 mv:2 bg:1 nqp:1 yv:4 oi:2 ni:1 qk:2 pya:2 v9:7 zy:3 lu:2 xyi:1 bdcfe:1 mc:1 j6:1 fo:1 ed:4 rsh:2 ty:2 npo:2 e2:1 di:1 ut:2 yyt:1 hmg:1 dt:1 y7:1 yb:2 ox:1 xa:1 d:4 xzy:3 ei:3 su:1 fehg:1 e0n:1 yf:1 ye:4 gz0:3 qe:1 fet:1 m5:1 x4i:1 ack:1 tn:1 fj:1 ef:1 a30:1 ug:1 ji:2 blk:1 jp:1 jl:2 xsx:1 s5:1 uv:1 u7:2 z4:1 l3:1 pye:2 v0:1 j:3 wxt:1 dye:1 hy0:4 xzw:1 wv:4 rgk:2 lpo:1 xe:1 ey:5 arx:1 ale:3 d7:1 o5:1 xf:2 rti:6 y:6 a1:1 va:1 z5:8 ko:2 n5:3 qy:6 dfehg:2 zw:3 sr:1 nv:3 cdh:1 r1v:2 vw:1 mw:1 mvo:2 trek:1 qho:1 b7:1 fm:2 mvl:1 tm:1 cn:2 nrt:2 br:4 gb:2 f:1 fnt:1 hke:1 v8:1 yw:4 s4:1 a34:1 ytk:2 ph:3 gih:2 ken:1 rw:1 pfn:2 pks:2 rtk:4 wr:1 rb:1 hji:1 zd:3 lnm:2 iz:1 qhy:1 key:1 fpo:1 k4:3 pj:1 ce:3 v1:1 xbw:2 g02:1 uw:1 ncd:4 vqv:1 hwc:1 h5:3 vn:2 p3:1 dy:4 vf:1 sfm:1 ki:1 hi:1 g:3 bv:1 k8:3 hy:4 qsd:1 px:2 xsc:3 tv:8 rnp:1 cbed:1 em:1 y0:1 ur:1 wi:2 n4:1 s1:1 ikj:3 mpo:1 dv:1 den:2 xo:1 gda:1 vq:2 r3:5 xyz:2 a0b:2 ozy:1 v2:1 mvt:1 rl2:1 jn:2 rz:2 sw:3 rce:1 qsr:1 xw:1 dfe:3 l6:1 eon:1 yz:4 xlp:1 yty:4 rt:11 nr:1 prq:2 mx:1 tue:1 w0:2 nx:1 y5:1 fy:1 ru:1 fgr:1 y4:1 xzf:1 nc:2 lg:2 kn5:1 ql:2 fe:5 gk:3 ba:6 dyp:1 zf:2 nvi:1 t:1 bwk:1 dc:1 rn:2 rtt:1 uwo:3 wyx:2 x_i:2 tuo:1 iih:1 pvu:1 c4:1 hgji:1 nip:1 wy:1 xm:1 fp:13 qsk:1 tb:2 rf:1 ia:1 zr:1 dz0:3 mn:2 ozt:1 yq:1 lk:2 rtn:1 gz:5 gf:2 sn:1 vh:1 l2:1 amz:1 kf:3 cih:1 oy:8 vg:3 lv:2 h2:1 gzh:1 o8:1 bdp:1 dq:1 lpn:1 pi:2 tfp:2 lo:1 zle:2 fg:2 tuu:3 xn:3 kyj:1 ytv:2 fb:3 kz:3 ig:1 lz:1 tlt:1 qx:1 vce:2 nov:1 em2:1 sz:1 ml:4 ucy:1 sk:1 ku:1 lks:1 z_:1 da:2 sp:2 pk:2 s2:1 n2:1 ika:1 xu:1 f9:1 rtr:1 vr:1 n:1 uxw:1 jsn:1 pv:2 sf:1 lq:1 xl:1 r6:1 rtp:8 ix:1 hw:1 z0:3 e4:7 xt:3 er:1 r8:2 x:2 pz:3 a3:4 ih:1 kr:1 te:3 o9:5 nk:2 mf:1 rg:5 cx:1 bo:1 a8:1 brg:1 dzr:3 fw:1 zr1:2 kxs:1 enj:1 ya:4 pys:1 zg:1 wq:1 kvn:1 c9:1 ex:2
7,142
906
An Alternative Model for Mixtures of Experts Lei Xu Dept. of Computer Science, The Chinese University of Hong Kong Shatin, Hong Kong, Emaillxu@cs.cuhk.hk Michael I. Jordan Dept. of Brain and Cognitive Sciences MIT Cambridge, MA 02139 Geoffrey E. Hinton Dept. of Computer Science University of Toronto Toronto, M5S lA4, Canada Abstract We propose an alternative model for mixtures of experts which uses a different parametric form for the gating network. The modified model is trained by the EM algorithm. In comparison with earlier models-trained by either EM or gradient ascent-there is no need to select a learning stepsize. We report simulation experiments which show that the new architecture yields faster convergence. We also apply the new model to two problem domains: piecewise nonlinear function approximation and the combination of multiple previously trained classifiers. 1 INTRODUCTION For the mixtures of experts architecture (Jacobs, Jordan, Nowlan & Hinton, 1991), the EM algorithm decouples the learning process in a manner that fits well with the modular structure and yields a considerably improved rate of convergence (Jordan & Jacobs, 1994). The favorable properties of EM have also been shown by theoretical analyses (Jordan & Xu, in press; Xu & Jordan, 1994). It is difficult to apply EM to some parts of the mixtures of experts architecture because of the nonlinearity of softmax gating network. This makes the maximiza- 634 Lei Xu, Michael!. Jordan, Geoffrey E. Hinton tion with respect to the parameters in gating network nonlinear and analytically unsolvable even for the simplest generalized linear case. Jordan and Jacobs (1994) suggested a double-loop approach in which an inner loop of iteratively-reweighted least squares (IRLS) is used to perform the nonlinear optimization. However, this requires extra computation and the stepsize must be chosen carefully to guarantee the convergence of the inner loop. We propose an alternative model for mixtures of experts which uses a different parametric form for the gating network. This form is chosen so that the maximization with respect to the parameters of the gating network can be handled analytically. Thus, a single-loop EM can be used, and no learning stepsize is required to guarantee convergence. We report simulation experiments which show that the new architecture yields faster convergence. We also apply the model to two problem domains. One is a piecewise nonlinear function approximation problem with smooth blending of pieces specified by polynomial, trigonometric, or other prespecified basis functions. The other is to combine classifiers developed previously-a general problem with a variety of applications (Xu, et al., 1991, 1992). Xu and Jordan (1993) proposed to solve the problem by using the mixtures of experts architecture and suggested an algorithm for bypassing the difficulty caused by the softmax gating networks. Here, we show that the algorithm of Xu and Jordan (1993) can be regarded as a special case of the single-loop EM given in this paper and that the single-loop EM also provides a further improvement. 2 MIXTURES OF EXPERTS AND EM LEARNING The mixtures of experts model is based on the following conditional mixture: K P(ylx,6) I: 9j(X, lI)P(ylx, OJ), P(ylx, OJ) where x ERn, and 6 consists of 1I,{Oj}f, and OJ consists of {wj}f,{rj}f. The vector Ij (x, Wj) ~s the output ofthe j-th expert net. The scalar 9j (x, 1I), j 1, ... , K is given by the softmax function: = 9j(X,1I) = e.B;(x,v)/I:e.B;(X,v). (2) In this equation, Pj (x, 1I), j = 1, ... ,K are the outputs of the gating network. The parameter 6 is estimated by Maximum Likelihood (ML), where the log likelihood is given by L = Lt In P(y(t) Ix(t), 6). The ML estimate can be found iteratively using the EM algorithm as follows. Given the current estimate 6(1~), each iteration consists of two steps. (1) E-step. For each pair {x(t),y(t)}, compute h~k)(y(t)lx(t?) = PUlx(t),y(t?), and then form a set of objective functions: Qj(Oj) I:h;k)(y(t)lx(t?) InP(y(t)lx(t), OJ), j = 1,,, ',K; An Alternative Model for Mixtures of Experts Qg(V) = 635 L L h)k) (y(t) Ix(t?) lngt) (x(t) ,v(k?). t (3) j (2). M-step. Find a new estimate e(k+l) = {{ot+l)}f=I,V(k+l)} with: OJk+l) = argmax Qj(Oj), j = 1, ... , K; ~ v(k+l) = arg max Qg(v). (4) v In certain cases, for example when I; (x, Wj) is linear in the parameters Wj, maXe j Qj(Oj) can be solved by solving 8Qj/80j = O. When l;(x,Wj) is nonlinear with respect to Wj, however, the maximization cannot be performed analytically. Moreover, due to the nonlinearity of softmax, maXv Qg(v) cannot be solved analytically in any case. There are two possibilities for attacking these nonlinear optimization problems. One is to use a conventional iterative optimization technique (e.g., gradient ascent) to perform one or more inner-loop iterations. The other is to simply find a new estimate such that Qj(ot+l?) ~ Qj(Ojk?), Qg(v(k+l?) ~ Qg(v(k?). Usually, the algorithms that perform a full maximization during the M step are referred as "EM" algorithms, and algorithms that simply increase the Q function during the M step as "GEM" algorithms. In this paper we will further distinguish between EM algorithms requiring and not requiring an iterative inner loop by designating them as double-loop EM and single-loop EM respectively. Jordan and Jacobs (1994) considered the case of linear {3j(x, v) = vJ[x,l] with v = [VI,???, VK] and semi-linear I; (wn x , 1]) with nonlinear 1;(.). They proposed a double-loop EM algorithm by using the IRLS method to implement the inner-loop iteration. For more general nonlinear {3j(x, v) and I;(x, OJ), Jordan and Xu (in press) showed that an extended IRLS can be used for this inner loop. It can be shown that IRLS and the extension are equivalent to solving eq. (3) by the so-called Fisher Scoring method. 3 A NEW GATING NET AND A SINGLE-LOOP EM To sidestep the need for a nonlinear optimization routine in the inner loop of the EM algorithm, we propose the following modified gating network: gj(x, v) = CkjP(xlvj)/ L:i CkiP(xIVi) , L:j Ckj = 1, Ckj ~ 0, P(xIVj) = aj(vj)-lbj(x) exp{cj(Vj)Ttj(x)} (5) where v = {Ckj,Vj,j = 1,???,K}, tj(x) is a vector of sufficient statistics, and the P(xlvj)'s are density functions from the exponential family. The most common example is the Gaussian: (6) In eq. (5), gj(x, v) is actually the posterior probability PUlx) that x is assigned to the partition corresponding to the j-th expert net, obtained from Bayes' rule: gj(x, v) = PUlx) = CkjP(xIVj)/ P(x, v), P(x, v) = L CkiP(xlvi). (7) 636 Lei Xu, Michael I. Jordan, Geoffrey E. Hinton Inserting this 9j(X, v) into the model eq. (1), we get P(ylx, 8) ?) = "L-. a?P(xlv ~( ) P(ylx, OJ). X,V (8) 3 If we do ML estimation directly on this P(ylx,8) and derive an EM algorithm, we again find that the maximization maXv Q9(v) cannot be solved analytically. To avoid this difficulty, we rewrite eq. (8) as: P(y, x) = P(ylx, 8)P(x, v) = L ajP(xlvj)P(ylx, OJ). (9) j This suggests an asymmetrical representation for the joint density. We accordingly perform ML estimation based on L' = 2:t In P(y(t), x(t?) to determine the parameters a j , Vj, OJ of the gating net and the expert nets. This can be done by the following EM algorithm: (1) E-step. Compute h(k)(y(t) Ix(t?) _ j - a\k) P( x(t) Iv~k ?)P(y(t) Ix(t) O<k?) 3 3 ' 3 . 2:i a~k) P(x(t)lv??)P(y(t)lx(t), 0)"?)' (10) Then let Qj(Oj),j = 1,? .. , K be the same as given in eq. (3), and decompose Q9(v) further into L h)k '(y(t) Ix(t?) In P( x(t) IVj), j = 1, ... , K; t LLh;k)(y(t)lx(t?)lnaj, with a= {al, .. ?,aK}. (11) j (2). M-step. Find a new estimate for j = 1,???, K O;k+l) = argmaXS j Qj(Oj), V]"+l) = argmaxllj QJ(Vj)' a(k+1) = arg maXa, QlX, s.t. 2:j aj = 1. (12) The maximization for the expert nets is the same as in eq. (4). However, for the gating net the maximization now becomes analytically solvable as long as P(xlvj) is from the exponential family. That is, we have: V~k+l) = 2:t h)")(y(t) Ix(t?)tj(x(t?) 3 2:t h)k)(y(t)lx(t?) , a;k+l) = ~ L h)")(y(t) Ix(t?). t In particular, when P(xIVj) is a Gaussian density, the update becomes: (13) An Alternative Model for Mixtures of Experts 637 Two issues deserve to be emphasized further: (1) The gating nets eq. (2) and eq. (5) become identical when f3j(x, v) = lnaj + In bj (x) +Cj (Vj)T tj (x) -In aj(vj) . In other words, the gating net in eq. (5) explicitly uses this function family instead of the function family defined by a multilayer feedforward network. (2) It follows from eq. (9) that max In P(y, xiS) = max [In P(ylx, S) + In P(xlv)]. So, the solution given by eqs. (10) through (14) is actually different from the one given by the original eqs. (3) and (4). The former tries to model both the mapping from x to y and the input x, while the latter only models the mapping from x and y. In fact, here we learn the parameters of the gating net and the expert nets via an asymmetrical representation eq. (9) of the joint density P(y, x) which includes P(ylx) implicitly. However, in the testing phase, the total output still follows eq. (8). 4 PIECEWISE NONLINEAR APPROXIMATION The simple form /j(x, Wj) = wJ[x , 1] is not the only case to which single-loop EM applies. Whenever /j(x, Wj) can be written in a form linear in the parameters: (15) where 4>i,j(X) are prespecified basis functions, maX8 j Qf!(Oj),j = 1"", K in eq. (3) is still a weighted least squares problem that can be soived analytically. One useful special case is when 4>i,j(X) are canonical polynomial terms X~l .. 'X~d, r j ~ 0. In this case, the mixture of experts model implements piecewise polynomial approximations. Another case is that 4>i,j(X) is TIi sini (jll'xt) cosi(jll'xt}, ri ~ 0, in which case the mixture of experts implements piecewise trigonometric approximations. 5 COMBINING MULTIPLE CLASSIFIERS Given pattern classes Ci, i = 1, ... , M, we consider classifiers ej that for each input x produce an output Pj(ylx): Pj(ylx) = [Pj(ll x ), ... ,pj(Mlx)), pj(ilx) ~ 0, LPj(ilx) = 1. (16) The problem of Combining Multiple Classifiers (CMC) is to combine these Pj(ylx)'s to give a combined estimate of P(ylx) . Xu and Jordan (1993) proposed to solve CMC problems by regarding the problem as a special example of the mixture density problem eq. (1) with the Pj(ylx)'s known and only the gating net 9j(X, v) to be learned. In Xu and Jordan (1993), one problem encountered was also the nonlinearity of softmax gating networks, and an algorithm was proposed to avoid the difficulty. Actually, the single-loop EM given by eq. (10) and eq. (13) can be directly used to solve the CMC problem. In particular, when P(xlvj) is Gaussian, eq. (13) becomes eq. (14). Assuming that al = . . . = aK in eq. (7), eq. (10) becomes 638 Lei Xu, Michaell. Jordan, Geoffrey E. Hinton h)k) (y(t) Ix(t)) = P(x(t)lvt))P(y(t)lx(t))/ L:i P(x(t)lvi(k))p(y(t)lx(t)). If we divide L:i P(x(t) Ivi(k)), we get ht)(y(t) Ix(t)) = gj(x, v)P(y(t)lx(t))/ L:i gj(x, v)P(y(t)lx(t)) . Comparing this equation with eq. (7a) in Xu and Jordan (1993), we can see that the two equations are actually the same. Despite the different notation, C?j(x) and Pj(y1.t) Ix(t)) in Xu and Jordan (1993) are the same as gj(x, v) and P(y(t)lx(t)) in Section 3. So the algorithm of Xu and Jordan (1993) is a special case of the single-loop EM given in Section 3. both the numerator and denominator by 6 SIMULATION RESULTS We compare the performance of the EM algorithm presented earlier with the model of mixtures of experts presented by Jordan and Jacobs (1994). As shown in Fig. l(a), we consider a mixture of experts model with K 2. For the expert nets, each P(ylx, OJ) is Gaussian given by eq. (1) with linear !;(x,Wj) = wJ[x, 1] . For the new gating net, each P(x, Vj) in eq. (5) is Gaussian given by eq. (6) . For the old gating net eq. (2), f31(X,V) 0 and f32(X,v) = vT[x, 1]. The learning speeds of the two are significantly different. The new algorithm takes k=15 iterations for the log-likelihood to converge to the value of -1271.8. These iterations require about 1,351,383 MATLAB flops. For the old algorithm, we use the IRLS algorithm given in Jordan and Jacobs (1994) for the inner loop iteration. In experiments, we found that it usually took a large number of iterations for the inner loop to converge. To save computations, we limit the maximum number of iterations by Tmaz = 10. We found that this saved computation without obviously influencing the overall performance. From Fig. 1(b), we see that the outer loop converges in about 16 iterations. Each inner loop takes 290498 flops and the entire process requires 5,312,695 flops. So, we see that the new algorithm yields a speedup of about 4,648,608/1,441,475 = 3.9. Moreover, no external adjustment is needed to ensure the convergence of the new algorithm. But for the old one the direct use of IRLS can make the inner loop diverge and we need to appropriately rescale the updating stepsize of IRLS. = = Figs. 2(a) and (b) show the results of a simulation of a piecewise polynomial approximation problem utilizing the approach described in Section 4. We consider a mixture of experts model with K = 2. For expert nets, each P(ylx, OJ) is Gaussian given by eq. (1) with !;(x,Wj) = W3,jX 3 +W2,jX 2 +Wl,jX+WD,j. In the new gating net eq. (5), each P(x, Vj) is again Gaussian given by eq. (6) . We see that the higher order nonlinear regression has been fit quite well. For multiple classifier combination, the problem and data are the same as in Xu and Jordan (1993). Table 1 shows the classification results. Com-old and Com-new denote the method given in in Xu and Jordan (1993) and in Section 5 respectively. We see that both improve the classification rate of each individual network considerably and that Com - new improves on Com - old. Training set Testing set Classifer el 89.9% 89.2% Classifer el 93.3% 92.7% Com- old Com- new 98.6% 99.4% 98.0% 99.0% Table 1 A comparison of the correct classification rates An Alternative Model for Mixtures of Experts o.s U (a) 2 U 3 U 639 4 (b) Figure 1: (a) 1000 samples from y = a1X + a2 + c, a1 = 0.8, a2 = 0.4, x E [-1,1.5] with prior a1 = 0.25 and y = ai x + a~ + c, ai = 0.8, a2 =' 2.4, x E [1,4] with prior a2 = 0.75, where x is uniform random variable and z is from Gaussian N(O, 0.3). The two lines through the clouds are the estimated models of two expert nets. The fits obtained by the two learning algorithms are almost the same. (b) The evolution of the log-likelihood. The solid line is for the modified learning algorithm. The dotted line is for the original learning algorithm (the outer loop iteration). 7 REMARKS Recently, Ghahramani and Jordan (1994) proposed solving function approximation problems by using a mixture of Gaussians to estimate the joint density of the input and output (see also Specht, 1991; Tresp, et al., 1994). In the special case of linear I;(x, Wj) = wnx,I] and Gaussian P(xlvj) with equal priors, the method given in Section 3 provides the same result as Ghahramani and Jordan (1994) although the parameterizations of the two methods are different. However, the method of this paper also applies to nonlinear l;(x,Wj) = Wn<pj(x) , 1] for piecewise nonlinear approximation or more generally I; (x, Wj) that is nonlinear with respect to Wj, and applies to cases in which P(y, xlvj, OJ) is not Gaussian, as well as the case of combining multiple classifiers. Furthermore, the methods proposed in Sections 3 and 4 can also be extended to the hierarchical mixture of experts architecture (Jacobs & Jordan, 1994) so that single-loop EM can be used to facilitate its training. References Ghahramani, Z., & Jordan, M.I. (1994). Function approximation via density estimation using the EM approach. In Cowan, J.D., Tesauro, G., and Alspector, J., (Eds.), Advances in NIPS 6. San Mateo, CA: Morgan Kaufmann. Jacobs, R.A., Jordan, M.L, Nowlan, S.J., & Hinton, G.E. (1991). Adaptive mixtures of local experts. Neural Computation, 9, 79-87. 640 Lei Xu, Michael!. Jordan, Geoffrey E. Hinton 110' 0 .as -\ I1 -\.5 -2 ~ -2.5 -2 -3 -3.5 -4 0 (a) (b) Figure 2: Piecewise 3rd polynomial approximation. (a) 1000 samples from y = alx3+a3x+a4+c, x E [-1,1.5] with prior 0'1 = 0.4 and y = a~x2+a~x2+a4+c, x E [1,4] with prior 0'2 = 0.6, where x is uniform random variable and z is from Gaussian N(0,0.15). The two curves through the clouds are the estimated models of two expert nets. (b) The evolution of the log-likelihood. Jordan, M.I., & Jacobs, R.A. (1994). Hierarchical mixtures of experts and the EM algorithm. Neural Computation, 6, 181-214. Jordan, M.I., & Xu, L. (in press). Convergence results for the EM approach to mixtures-of-experts architectures. Neural Networks. Specht, D. (1991). A general regression neural network. IEEE Trans. Networks, 2, 568-576. Neural Tresp, V., Ahmad, S., and Neuneier, R. (1994). Training neural networks with deficient data. In Cowan, J.D., Tesauro, G., & Alspector, J., (Eds.), Advances in NIPS 6, San Mateo, CA: Morgan Kaufmann. Xu, L., Krzyzak A., & Suen, C.Y. (1991). Associative switch for combining multiple classifiers. Proc. of 1991 HCNN, Vol. I. Seattle, 43-48. Xu, L., Krzyzak A., & Suen, C.Y. (1992). Several methods for combining multiple classifiers and their applications in handwritten character recognition. IEEE Thans. on SMC, Vol. SMC-22, 418-435. Xu, L., & Jordan, M.I. (1993). EM Learning on a generalized finite mixture model for combining multiple classifiers. Proceedings of World Congress on Neural Networks, Vol. IV. Portland, OR, 227-230. Xu, L., & Jordan, M.I. (1994). On convergence properties of the EM algorithm for Gaussian mixtures. Submitted to Neural Computation.
906 |@word kong:2 f32:1 polynomial:5 simulation:4 jacob:9 solid:1 current:1 comparing:1 wd:1 nowlan:2 com:6 neuneier:1 must:1 written:1 partition:1 update:1 maxv:2 accordingly:1 prespecified:2 provides:2 parameterizations:1 toronto:2 lx:11 direct:1 become:1 consists:3 combine:2 manner:1 alspector:2 brain:1 becomes:4 moreover:2 notation:1 maxa:1 developed:1 guarantee:2 decouples:1 classifier:10 influencing:1 local:1 congress:1 limit:1 despite:1 ak:2 mateo:2 suggests:1 smc:2 testing:2 implement:3 significantly:1 word:1 inp:1 get:2 cannot:3 conventional:1 equivalent:1 rule:1 utilizing:1 regarded:1 us:3 designating:1 recognition:1 updating:1 cloud:2 solved:3 wj:16 ahmad:1 trained:3 solving:3 rewrite:1 classifer:2 basis:2 joint:3 quite:1 modular:1 solve:3 statistic:1 la4:1 associative:1 obviously:1 net:19 took:1 propose:3 inserting:1 argmaxs:1 loop:25 combining:6 trigonometric:2 seattle:1 convergence:8 double:3 produce:1 converges:1 derive:1 rescale:1 ij:1 eq:30 c:1 cmc:3 saved:1 correct:1 require:1 decompose:1 blending:1 extension:1 bypassing:1 considered:1 exp:1 mapping:2 bj:1 jx:3 a2:4 favorable:1 estimation:3 proc:1 wl:1 weighted:1 mit:1 suen:2 gaussian:12 modified:3 avoid:2 ej:1 improvement:1 vk:1 portland:1 likelihood:5 hk:1 el:2 entire:1 i1:1 arg:2 issue:1 overall:1 classification:3 softmax:5 special:5 equal:1 identical:1 report:2 piecewise:8 individual:1 argmax:1 phase:1 lbj:1 possibility:1 mixture:25 tj:3 iv:2 divide:1 old:6 theoretical:1 earlier:2 maximization:6 uniform:2 considerably:2 combined:1 density:7 diverge:1 michael:4 ivj:1 again:2 cognitive:1 external:1 expert:29 sidestep:1 li:1 tii:1 includes:1 caused:1 explicitly:1 vi:1 piece:1 tion:1 performed:1 try:1 bayes:1 square:2 kaufmann:2 yield:4 ofthe:1 handwritten:1 irls:7 m5s:1 submitted:1 whenever:1 ttj:1 ed:2 f3j:1 improves:1 cj:2 routine:1 carefully:1 actually:4 higher:1 improved:1 done:1 cosi:1 furthermore:1 nonlinear:14 jll:2 aj:3 lei:5 facilitate:1 requiring:2 asymmetrical:2 former:1 analytically:7 assigned:1 evolution:2 iteratively:2 reweighted:1 ll:1 during:2 numerator:1 hong:2 generalized:2 recently:1 common:1 cambridge:1 ai:2 rd:1 nonlinearity:3 gj:6 posterior:1 xlv:2 showed:1 tesauro:2 certain:1 vt:1 scoring:1 morgan:2 attacking:1 determine:1 cuhk:1 converge:2 semi:1 full:1 multiple:8 rj:1 smooth:1 faster:2 long:1 a1:2 qg:5 regression:2 multilayer:1 denominator:1 iteration:10 appropriately:1 extra:1 ot:2 sini:1 w2:1 ivi:1 ascent:2 deficient:1 cowan:2 jordan:32 feedforward:1 wn:2 variety:1 switch:1 fit:3 w3:1 architecture:7 ckj:3 inner:11 regarding:1 shatin:1 qj:9 handled:1 krzyzak:2 remark:1 matlab:1 useful:1 generally:1 ylx:17 simplest:1 canonical:1 dotted:1 estimated:3 vol:3 pj:10 ht:1 unsolvable:1 family:4 almost:1 q9:2 distinguish:1 encountered:1 ri:1 x2:2 speed:1 lpj:1 ern:1 speedup:1 combination:2 em:29 character:1 equation:3 previously:2 needed:1 specht:2 gaussians:1 apply:3 hierarchical:2 stepsize:4 save:1 alternative:6 original:2 ensure:1 a4:2 ghahramani:3 chinese:1 objective:1 parametric:2 gradient:2 outer:2 assuming:1 tmaz:1 difficult:1 perform:4 finite:1 flop:3 hinton:7 extended:2 y1:1 michaell:1 canada:1 pair:1 required:1 specified:1 learned:1 nip:2 trans:1 deserve:1 suggested:2 usually:2 pattern:1 oj:18 max:3 difficulty:3 solvable:1 improve:1 tresp:2 prior:5 mlx:1 qlx:1 geoffrey:5 lv:1 sufficient:1 qf:1 f31:1 curve:1 world:1 llh:1 adaptive:1 san:2 implicitly:1 ml:4 gem:1 xi:1 iterative:2 table:2 learn:1 ca:2 domain:2 vj:10 lvi:1 xu:23 fig:3 referred:1 exponential:2 ix:10 xt:2 emphasized:1 gating:19 ci:1 ilx:2 lt:1 simply:2 adjustment:1 scalar:1 applies:3 ma:1 conditional:1 hcnn:1 fisher:1 called:1 total:1 maxe:1 select:1 ojk:2 latter:1 dept:3
7,143
907
Classifying with Gaussian Mixtures and Clusters Nanda Kambhatla and Todd K. Leen Department of Computer Science and Engineering Oregon Graduate Institute of Science & Technology P.O. Box 91000 Portland, OR 97291-1000 nanda@cse.ogi.edu, tleen@cse.ogi.edu Abstract In this paper, we derive classifiers which are winner-take-all (WTA) approximations to a Bayes classifier with Gaussian mixtures for class conditional densities. The derived classifiers include clustering based algorithms like LVQ and k-Means. We propose a constrained rank Gaussian mixtures model and derive a WTA algorithm for it. Our experiments with two speech classification tasks indicate that the constrained rank model and the WTA approximations improve the performance over the unconstrained models. 1 Introduction A classifier assigns vectors from Rn (n dimensional feature space) to one of K classes, partitioning the feature space into a set of K disjoint regions. A Bayesian classifier builds the partition based on a model of the class conditional probability densities of the inputs (the partition is optimal for the given model). In this paper, we assume that the class conditional densities are modeled by mixtures of Gaussians. Based on Nowlan's work relating Gaussian mixtures and clustering (Nowlan 1991), we derive winner-take-all (WTA) algorithms which approximate a Gaussian mixtures Bayes classifier. We also show the relationship of these algorithms to non-Bayesian cluster-based techniques like LVQ and k-Means. The main problem with using Gaussian mixtures (or WTA algorithms thereof) is the explosion in the number of parameters with the input dimensionality. We propose 682 Nanda Kambhatla. Todd K. Leen a constrained rank Gaussian mixtures model for classification. Constraining the rank of the Gaussians reduces the effective number of model parameters thereby regularizing the model. We present the model and derive a WTA algorithm for it. Finally, we compare the performance of the different mixture models discussed in this paper for two speech classification tasks. 2 Gaussian Mixture Bayes (GMB) classifiers Let x denote the feature vector (x E 'Rn ), and {n I , I = 1, ... , K} denote the classes. Class priors are denoted p(nI) and the class-conditional densities are denoted p(x InI). The discriminant function for the Bayes classifier is (1) An input feature vector x is assigned to class I if 6 I(x) > 6J (x) 'VJ:F I . Given the class conditional densities, this choice minimizes the classification error rate (Duda and Hart 1973). We model each class conditional density by a mixture composed of QI component Gaussians. The Bayes discriminant function (see Figure 1) becomes QI I "j,foiTlexp[-~(X-~J{~J-\x-~J)], j'(x)=PW)L ;=1 I~fl (21r)n/2 where ~J and ~f are the mean and the covariance matrix of the ponent for nl. lh (2) mixture com- 0.12 0.1 0.08 0.06 0.04 0.02 5 10 --~)o 20 X 25 Fig. 1: Figure showing the decision rule of a GMB classifier for a two class problem with one input feature. The horizontal axis represents the feature and the vertical axis represents the Bayes discriminant functions. In this example, the class conditional densities are modelled as a mixture of two Gaussians and equal priors are assumed. To implement the Gaussian mixture Bayes classifier (GMB) we first separate the training data into the different classes. We then use the EM algorithm (Dempster Classifying with Gaussian Mixtures and Clusters 683 et al 1977, Nowlan 1991) to determine the parameters for the Gaussian mixture density for each class. 3 Winner-take-all approximations to G MB classifiers In this section, we derive winner-take-all (WTA) approximations to GMB classifiers. We also show the relationship of these algorithms to non-Bayesian cluster-based techniques like LVQ and k-Means. 3.1 The WTA model for GMB The WTA assumptions (relating hard clustering to Gaussian mixtures; see (Nowlan 1991)) are: ? p(x 10 1 ) are mixtures of Gaussians as in (2). ? The summation in (2) is dominated by the largest term. This is "equivalent to assigning all of the responsibility for an observation to the Gaussian with the highest probability of generating that observation" (Nowlan 1991). To draw the relation between GMB and cluster-based classifiers, we further assume that: ? The mixing proportions (oj) are equal for a given class. ? The number of mixture components QI is proportional to p(O/). Applying all the above assumptions to (2), taking logs and discarding the terms that are identical for each class, we get the discriminant function AI [1"2 log IEj!) + "21(x - ILj) 1 Q. (I -y (x) = - ~f I T I-I ~j I ] (x - ILj) . (3) The discriminant function (3) suggests an algorithm that approximates the Bayes classifier. We segregate the feature vectors by class and then train a separate vector quantizer (VQ) for each class. We then compute the means ILl and the covariance each Voronoi cell of each quantizer, and use (3) for classifying new matrices patterns. We call this algorithm VQ-Covariance. Note that this algorithm does not do a maximum likelihood estimation of its parameters based on the probability model used to derive (3). The probability model is only used to classify patterns. E1Jor 3.2 The relation to LVQ and k-Means Further assume that for each class, the mixture components are spherically sym= 0"2 I, with 0"2 identical for all classes. We obtain metric with covariance matrix the discriminant function, El A -y/(x) =- QI r?n )=1 II x - J.t)~ 2 II . (4) 684 Nanda Kamblwtla, Todd K. Leen This is exactly the discriminant function used by the learning vector quantizer (LVQ; Kohonen 1989) algorithm. Though LVQ employs a discriminatory training procedure (i.e it directly learns the class boundaries and does not explicitly build a separate model for each class), the implicit model of the class conditional densities used by LVQ corresponds to a GMB model under all the assumptions listed above. This is also the implicit model underlying any classifier which makes its classification decision based on the Euclidean distance measure between a feature vector and a set of prototype vectors (e.g. a k-Means clustering followed by classification based on (4)) . 4 Constrained rank GMB classifiers In the preceding sections, we have presented a GMB classifier and some WTA approximations to GMB. Mixture models such as GMB generally have too many parameters for small data sets. In this section, we propose a way of regularizing the mixture densities and derive a WTA classifier for the regularized model. 4.1 The constrained rank model In section 2, we assumed that the class conditional densities of the feature vectors x are mixtures of Gaussians (5) where p.J and EJ are the means and covariance matrices for the jth component Gaussian. eJi and AJi are the orthonormal eigenvectors and eigenvalues of I::Ji (ordered such that A}l ~ '" ~ AJn). In (5), we have written the Mahalanobis distance in terms of the eigenvectors. For a particular data point x, the Mahalanobis distance is very sensitive to changes in the squared projections onto the trailing eigen-directions, since the variances are very small in these directions. This is a potential problem with small data sets. When there are insufficient data points to estimate all the parameters of the mixture density accurately, the trailing eigen-directions and their associated eigenvalues are likely to be poorly estimated. Using the Mahalanobis distance in (5) can lead to erroneous results in such cases. We propose a method for regularizing Gaussian mixture classifiers based on the above ideas. We assume that the trailing n - m eigen-directions of each Gaussian component are inaccurate due to overfitting to the training set. We rewrite the class conditional densities (5) retaining only the leading m (0 < m ::; n) eigen-directions Classifying with Gaussian Mixtures and Clusters 685 in the determinants and the Mahalanobis distances p(x IIll) = f j=1 m 2)ft=1 \i I m (21T) / exp [-~(x - ILJ{ (f e;~fiT) (x - ILJ)] i=l Jl (6) We choose the value of m (the reduced rank) by cross-validation over a separate validation set. Thus, our model can be considered to be regularizing or constraining the class conditional mixture densities. If we apply the above model and derive the Bayes discriminant functions (1), we get, 8I (x) = p(nI) ~ j=1 m (21T) / 2 a} m I exp y'ft=l \i [-~(X - ILJ)T (f e}~f/) (x - ILJ)]. i=1 Jl (7) We can implement a constrained rank Gaussian mixture Bayes (GMB-Reduced) classifier based on (7) using the EM algorithm to determine the parameters of the mixture density for each class. We segregate the data into different classes and use the EM algorithm to determine the parameters of the full mixture density (5). We then use (7) to classify patterns. 4.2 A constrained rank WTA algorithm We now derive a winner-take-all (WTA) approximation for the constrained rank mixture model described above. We assume (similar to section 3.1) that ? p(x I nI) are constrained mixtures of Gaussians as in (6). ? The summation in (6) is dominated by the largest term (the WTA assumption). . ? The mixing proportions (a}) are equal for a given class and the number of components QI is proportional to p(nI). Applying these assumptions to (7), taking logs and discarding the terms that are identical for each class, we get the discriminant function (8) It is interesting to compare (8) with (3) . Our model postulates that the trailing m eigen-directions of each Gaussian represent overfitting to noise in the training set. The discriminant functions reflect this; (8) retains only those terms of (3) which are in the leading m eigen-directions of each Gaussian. n- We can generate an algorithm based on (8) that approximates the reduced rank Bayes classifier. We separate the data based on classes and train a separate vector the covariance quantizer (VQ) for each class. We then compute the means matrices ~} for each Voronoi cell of each quantizer and the orthonormal eigenvectors IL}, 686 Nanda Kambhatla. Todd K. Leen Table 1: The test set classification accuracies for the TIMIT vowels data for different algorithms. ALGORITHM MLP (40 nodes in hidden layer) GMB (1 component; full) GMB (1 component; diagonal) GMB-Reduced (1 component; 13-D) VQ-Covariance (1 component) VQ-Covariance-Reduced (1 component; 13-D) LVQ (48 cells) ACCURACY 46.8% 41.4% 46.3% 51.2% 41.4% 51.2% 41.4% eJi and eigenvalues AJ for each covariance matrix EJ. We use (8) for classifying new patterns. Notice that the algorithm described above is a reduced rank version of VQ-Covariance (described in section 3.1). We call this algorithm VQ-CovarianceReduced. 5 Experimental Results In this section we compare the different mixture models and a multi layer perceptron (MLP) for two speech phoneme classification tasks. The measure used is the classification accuracy. 5.1 TIMIT data The first task is the classification of 12 monothongal vowels from the TIMIT database (Fisher and Doddington 1986). Each feature vector consists of the lowest 32 DFT coefficients, time-averaged over the central third of the vowel. We partitioned the data into a training set (1200 vectors), a validation set (408 vectors) for model selection, and a test set (408 vectors). The training set contained 100 examples of each class. The values of the free parameters for the algorithms (the number of component densities, number of hidden nodes for the MLP etc.) were selected by maximizing the performance on the validation set. Table 1 shows the results obtained with different algorithms. The constrained rank models (GMB-Reduced and VQ-Covariance-Reduced 1 ) perform much better than all the unconstrained ones and even beat a MLP for this task. This data set consists of very few data points per class, and hence is particularly susceptible to over fitting by algorithms with a large number of parameters (like GMB). It is not surprising that constraining the number of model parameters is a big win for this task. INote that since the best validation set performance is obtained with only one component for each mixture density, the WTA algorithms are identical to the GMB algorithms (for these results). Classifying with Gaussian Mixtures and Clusters 687 Table 2: The test set classification accuracies for the CENSUS data for different algorithms. ALGORITHM MLP (80 nodes in hidden layer) GMB (1 component; full) GMB (8 components; diagonal) GMB-Reduced (2 components; 35-D) VQ-Covariance (3 components) VQ-Covariance-Reduced (4 components; 38-D) LVQ (55 cells) 5.2 ACCURACY 88.2% 77.2% 70.9% 82.5% 77.5% 84.2% 67.3% CENSUS data The next task we experimented with was the classification of 9 vowels (found in the utterances ofthe days of the week). The data was drawn from the CENSUS speech corpus (Cole et alI994). Each feature vector was 70 dimensional (perceptual linear prediction (PLP) coefficients (Hermansky 1990) over the vowel and surrounding context}. We partitioned the data into a training set (8997 vectors), a validation set (1362 vectors) for model selection, and a test set (1638 vectors). The training set had close to a 1000 vectors per class. The values of the free parameters for the different algorithms were selected by maximizing the validation set performance. Table 2 gives a summary of the classification accuracies obtained using the different algorithms. This data set has a lot more data points per class than the TIMIT data set. The best accuracy is obtained by a MLP, though the constrained rank mixture models still greatly outperform the unconstrained ones. 6 Discussion We have derived WTA approximations to GMB classifiers and shown their relation to LVQ and k-Means algorithms. The main problem with Gaussian mixture models is the explosion in the number of model parameters with input dimensionality, resulting in poor generalization performance. We propose constrained rank Gaussian mixture models for classification. This approach ignores some directions ( "noise") locally in the input space, and thus reduces the effective number of model parameters. This can be considered as a way of regularizing the mixture models. Our results with speech vowel classification indicate that this approach works better than using full mixture models, especially when the data set size is small. The WTA algorithms proposed in this paper do not perform a maximum likelihood estimation of their parameters. The probability model is only used to classify data. We can potentially improve the performance of these algorithms by doing maximum likelihood training with respect to the models presented here. 688 Nanda Kambhatla, Todd K. Leen Acknowledgments This work was supported by grants from the Air Force Office of Scientific Research (F49620-93-1-0253), Electric Power Research Institute (RP8015-2) and the Office of Naval Research (NOOOI4-91-J-1482). We would like to thank Joachim Utans, OGI for several useful discussions and Zoubin Ghahramani, MIT for providing MATLAB code for the EM algorithm. We also thank our colleagues in the Center for Spoken Language Understanding at OGI for providing speech data. References R.A. Cole, D.G. Novick, D. Burnett, B. Hansen, S. Sutton, M. Fanty. (1994) Towards Automatic Collection of the U.S. Census. Proceedings of the International Conference on Acoustics, Speech and Signal Processing 1994. A.P. Dempster, N.M. Laird, and D.B. Rubin. (1977) Maximum Likelihood from Incomplete Data via the EM Algorithm. J. Royal Statistical Society Series B, vol. 39, pp. 1-38. R.O. Duda and P.E. Hart. (1973) Pattern Classification and Scene Analysis. John Wiley and Sons Inc. W.M Fisher and G.R Doddington. (1986) The DARPA speech recognition database: specification and status. In Proceedings of the DARPA Speech Recognition Workshop, p93-99, Palo Alto CA. H. Hermansky. (1990) Perceptual Linear Predictive (PLP) analysis of speech. J. Acoust. Soc. Am., 87(4):1738-1752. T. Kohonen. (1989) Self-Organization and Associative Memory (3rd edition). Berlin: Springer-Verlag. S.J. Nowlan. (1991) Soft Competitive Adaptation: Neural Network Learning Algorithms based on Fitting Statistical Mixtures. CMU-CS-91-126 PhD thesis, School of Computer Science, Carnegie Mellon University.
907 |@word determinant:1 version:1 pw:1 duda:2 proportion:2 covariance:13 thereby:1 series:1 nanda:6 com:1 nowlan:6 surprising:1 assigning:1 written:1 john:1 partition:2 selected:2 quantizer:5 node:3 cse:2 consists:2 fitting:2 multi:1 becomes:1 underlying:1 alto:1 lowest:1 minimizes:1 spoken:1 acoust:1 ajn:1 exactly:1 classifier:22 partitioning:1 grant:1 engineering:1 todd:5 sutton:1 suggests:1 discriminatory:1 graduate:1 averaged:1 acknowledgment:1 implement:2 procedure:1 aji:1 projection:1 zoubin:1 get:3 onto:1 close:1 selection:2 context:1 applying:2 equivalent:1 center:1 maximizing:2 assigns:1 rule:1 orthonormal:2 recognition:2 particularly:1 database:2 ft:2 region:1 highest:1 dempster:2 rewrite:1 predictive:1 darpa:2 surrounding:1 train:2 effective:2 ponent:1 laird:1 associative:1 eigenvalue:3 propose:5 mb:1 fanty:1 adaptation:1 kohonen:2 mixing:2 poorly:1 cluster:7 generating:1 derive:9 school:1 soc:1 c:1 indicate:2 direction:8 generalization:1 summation:2 considered:2 exp:2 week:1 trailing:4 kambhatla:4 estimation:2 hansen:1 palo:1 sensitive:1 cole:2 largest:2 mit:1 gaussian:23 gmb:22 ej:2 office:2 derived:2 naval:1 joachim:1 portland:1 rank:15 likelihood:4 greatly:1 am:1 voronoi:2 el:1 inaccurate:1 hidden:3 relation:3 classification:16 ill:1 denoted:2 retaining:1 constrained:12 equal:3 identical:4 represents:2 hermansky:2 employ:1 few:1 composed:1 vowel:6 organization:1 mlp:6 mixture:41 nl:1 explosion:2 lh:1 incomplete:1 euclidean:1 classify:3 soft:1 retains:1 novick:1 too:1 density:18 international:1 ilj:6 squared:1 postulate:1 reflect:1 central:1 thesis:1 choose:1 rp8015:1 leading:2 potential:1 coefficient:2 inc:1 oregon:1 explicitly:1 lot:1 responsibility:1 doing:1 competitive:1 bayes:11 timit:4 il:1 ni:4 accuracy:7 air:1 variance:1 phoneme:1 ofthe:1 modelled:1 bayesian:3 accurately:1 burnett:1 colleague:1 pp:1 thereof:1 associated:1 noooi4:1 dimensionality:2 day:1 leen:5 box:1 though:2 implicit:2 horizontal:1 aj:1 scientific:1 hence:1 assigned:1 spherically:1 ogi:4 mahalanobis:4 self:1 plp:2 ini:1 ji:1 winner:5 jl:2 discussed:1 approximates:2 relating:2 mellon:1 ai:1 dft:1 automatic:1 unconstrained:3 rd:1 language:1 had:1 specification:1 etc:1 verlag:1 preceding:1 determine:3 signal:1 ii:2 full:4 reduces:2 eji:2 cross:1 hart:2 qi:5 prediction:1 metric:1 cmu:1 represent:1 cell:4 call:2 constraining:3 fit:1 idea:1 prototype:1 speech:10 matlab:1 generally:1 useful:1 listed:1 eigenvectors:3 locally:1 reduced:10 generate:1 outperform:1 notice:1 utans:1 estimated:1 disjoint:1 per:3 carnegie:1 vol:1 drawn:1 draw:1 decision:2 tleen:1 fl:1 layer:3 followed:1 scene:1 dominated:2 department:1 poor:1 em:5 son:1 partitioned:2 wta:17 inote:1 census:4 vq:10 gaussians:7 apply:1 eigen:6 clustering:4 include:1 ghahramani:1 build:2 especially:1 society:1 diagonal:2 win:1 distance:5 separate:6 thank:2 berlin:1 discriminant:10 code:1 modeled:1 relationship:2 insufficient:1 providing:2 susceptible:1 potentially:1 perform:2 vertical:1 observation:2 beat:1 segregate:2 rn:2 acoustic:1 pattern:5 royal:1 oj:1 memory:1 power:1 force:1 regularized:1 improve:2 technology:1 axis:2 utterance:1 prior:2 understanding:1 interesting:1 proportional:2 validation:7 rubin:1 classifying:6 summary:1 supported:1 free:2 sym:1 jth:1 iej:1 perceptron:1 institute:2 taking:2 f49620:1 boundary:1 ignores:1 collection:1 approximate:1 status:1 overfitting:2 corpus:1 assumed:2 iill:1 table:4 ca:1 p93:1 electric:1 vj:1 main:2 big:1 noise:2 edition:1 fig:1 wiley:1 perceptual:2 third:1 learns:1 erroneous:1 discarding:2 showing:1 experimented:1 workshop:1 phd:1 likely:1 ordered:1 contained:1 springer:1 corresponds:1 conditional:11 lvq:10 towards:1 fisher:2 hard:1 change:1 experimental:1 doddington:2 regularizing:5
7,144
908
The Use of Dynamic Writing Information in a Connectionist On-Line Cursive Handwriting Recognition System Stefan Manke Michael Finke University of Karlsruhe Computer Science Department D-76128 Karlsruhe, Germany mankeCO)ira. uka.de, finkem@ira.uka.de Alex Waibel Carnegie Mellon University School of Computer Science Pittsburgh, PA 15213-3890, U.S.A. wai bel CO) cs. cm u.ed u Abstract In this paper we present NPen ++, a connectionist system for writer independent, large vocabulary on-line cursive handwriting recognition. This system combines a robust input representation, which preserves the dynamic writing information, with a neural network architecture, a so called Multi-State Time Delay Neural Network (MS-TDNN), which integrates rec.ognition and segmentation in a single framework. Our preprocessing transforms the original coordinate sequence into a (still temporal) sequence offeature vectors, which combine strictly local features, like curvature or writing direction, with a bitmap-like representation of the coordinate's proximity. The MS-TDNN architecture is well suited for handling temporal sequences as provided by this input representation. Our system is tested both on writer dependent and writer independent tasks with vocabulary sizes ranging from 400 up to 20,000 words. For example, on a 20,000 word vocabulary we achieve word recognition rates up to 88.9% (writer dependent) and 84.1 % (writer independent) without using any language models. 1094 1 Stefan Manke, Michael Finke, Alex Waibel INTRODUCTION Several preprocessing and recognition approaches for on-line handwriting recognition have been developed during the past years . The main advantage of on-line handwriting recognition in comparison to optical character recognition (OCR) is the temporal information of handwriting, which can be recorded and used for recognition . In general this dynamic writing information (i.e. the time-ordered sequence of coordinates) is not available in OCR, where input consists of scanned text. In this paper we present the NPen++ system, which is designed to preserve the dynamic writing information as long as possible in the preprocessing and recognition process. During preprocessing a temporal sequence of N-dimensional feature vectors is computed from the original coordinate sequence, which is recorded on the digitizer. These feature vectors combine strictly local features, like curvature and writing direction [4], with so-called context bitmaps, which are bitmap-like representations of a coordinate's proximity. The recognition component of NPen++ is well suited for handling temporal sequences of patterns, as provided by this kind of input representation. The recognizer, a so-called Multi-State Time Delay Neural Network (MS-TDNN), integrates recognition and segmentation of words into a single network architecture. The MS-TDNN, which was originally proposed for continuous speech recognition tasks [6, 7], combines shift-invariant, high accuracy pattern recognition capabilities of a TDNN [8, 4] with a non-linear alignment procedure for aligning strokes into character sequences. Our system is applied both to different writer dependent and writer independent, large vocabulary handwriting recognition tasks with vocabulary sizes up to 20,000 words. Writer independent word recognition rates range from 92.9% with a 400 word vocabulary to 84.1% with a 20,000 word vocabulary. For the writer dependent system, word recognition rates for the same tasks range from 98.6% to 88.9% [1]. In the following section we give a description of our preprocessing performed on the raw coordinate sequence, provided by the digitizer. In section 3 the architecture and training of the recognizer is presented. A description of the experiments to evaluate the system and the results we have achieved on different tasks can be found in section 4. Conclusions and future work is described in section 5. 2 PREPROCESSING The dynamic writing information, i.e. the temporal order of the data points, is preserved throughout all preprocessing steps. The original coordinate sequence {(x(t), y(t?hE{O ...T'} recorded on the digitizer is transformed into a new temporal sequence X6 = Xo ... XT, where each frame Xt consists of an N-dimensional realvalued feature vector (h(t), . .. , fN(t? E [-1, l]N. Several normalization methods are applied to remove undesired variability from the original coordinate sequence. To compensate for different sampling rates and varying writing speeds the coordinates originally sampled to be equidistant in time are resampled yielding a new sequence {(x(t),y(t?hE{O ...T} which is equidistant in Dynamic Writing Information in Cursive Handwriting Recognition 1095 normalized coordinate (b) writing direction curvature , ,, , x(.?2),y(??2) 'f ~ ,';'1 x(t-l),y('?l) t~, " .... , x('),y(') ' 1" x(t+2),y(t+2) ? x('+1),y('+l) x( ??2),y('.2) , , ~ , : ' /1 ' x(.+2),Y(1+2) x(t-11y('?1). \. ~ " .. ...... x(t+l),y(t+l) x(.),y(') Figure I: Feature extraction for the normalized word "able". The final input representation is derived by c,alculating a 15-dimensional feature vector for each data point, which consists of a context bitmap (a) and information about the curvature and writing direction (b). space. This resampled trajectory is smoothed using a moving average window in order to remove sampling noise. In a final normalization step the goal is to find a representation of the trajectory that is reasonably invariant against rotation and scaling of the input. The idea is to determine the words' baseline using an EM approach similar to that described in [5] and rescale the word such that the center region of the word is assigned to a fixed size. From the normalized coordinate sequence {(x(t), y(t))hE{O .. .T} the temporal sequence 2::r; of N-dimensional feature vectors ~t = (!l(t), ... .IN(t)) is computed (Figure I). Currently the system uses N = 15 features for each data point. The first two features fl(t) = x(t)-x(t-I) and h(t) = y(t)-b describe the relative X movement and the Y position relative to the baseline b. The features /get) to f6(t) are used to desc,ribe the curvature and writing direction in the trajectory [4] (Figure I(b)). Since all these features are strictly local in the sense that they are local both in time and in space they were shown to be inadequate for modeling temporal long range context dependencies typically observed in pen trajectories [2]. Therefore, nine additional features h(t) to !J5(t) representing :3 x:3 bitmaps were included in each feature vector (Figure I(a?. These so-called context bitmaps are basically low resolution, bitmap-like descriptions of the coordinate's proximity, which were originally described in [2]. Thus, the input representation as shown in Figure I combines strictly local features like writing direction and curvature with the c,ontext bitmaps, which are still local 1096 Stefan Manke, Michael Finke, Alex Waibel in space but global in time. That means, each point of the trajectory is visible from each other point of the trajectory in a small neighbourhood. By using these context bitmaps in addition to the local features, important information about other parts of the trajectory, which are in a limited neighbourhood of a coordinate, are encoded. 3 THE NPen++ RECOGNIZER The NPell ++ recognizer integrates recognition and segmentation of words into a single network architecture, the Multi-State Time Delay Neural Network (MSTDNN) . The MS-TDNN, which was originally proposed for continuous speech recognition tasks [6 , 7], combines the high accuracy single character recognition capabilities of a TDNN [8, 4] with a non-linear time alignment algorithm (dynamic time warping) for finding stroke and character boundaries in isolated handwritten words. 3.1 MODELING ASSUMPTIONS Let W = {WI, .. . WK} be a vocabulary consisting of K words. Each of these words Wi is represented as a sequence of characters Wi == Ci l Ci 2 ? ?? Cik where each character Cj itself is modelled by a three state hidden markov model Cj == qj q] qJ. The idea of using three states per character is to model explicitly the imtial, middle and final section of the characters. Thus , Wi is modelled by a sequence of states Wi == qioqi l ? .. qhk . In these word HMMs the self-loop probabilities p(qij/qij) and the transition probabilities p(qij/qij_l) are both defined to be ~ while all other transition probabilities are set to zero. During recognition of an unknown sequence of feature vectors ~'(; = ~o . . . ~T we have to find the word Wi E W in the dictionary that maximizes the a-posteriori probability p( Wi /~a 10) given a fixed set of parameters 0 and the observed coordinate sequence. That means , a written word will be recognized such that Wj = argmaXw,EWp(Wi/Za,O). In our Multi-State Time Delay Neural Network approach the problem of modeling the word posterior probability p( wdz'{; , 0) is simplified by using Bayes' rule which expresses that probability as p(Z'{;/WilO)P(Wi/O) p(z'{;/O) Instead of approximating p( Wi /z'{;, 0) directly we define in the following section a network that is supposed to model the likelihood of the feature vector sequence p(z'{; /Wi, 0). 3.2 THE MS-TDNN ARCHITECTURE In Figure 2 the basic MS-TDNN architecture for handwriting recognition is shown. The first three layers constitute a standard TDNN with sliding input windows in each layer. In the current implementation of the system, a TDNN with 15 input Dynamic Writing InJonnation in Cursive Handwriting Recognition 0 -g :; 0 0 "E i::' '" ..'!l.D .D'"0 1i " mm.o .? ? '" 1097 ~ '" '" '" ::::::: :." .D ] '" '" ~.m.W.J ... -.......-.... E 0 0 N ??? 0 ~ r:'?~ """~L-,;~???i]- ... ~ [~~~::~~~::~ ~~~~~~~~:::~::~:~~~~:J- ... ? ? ? ;r?5_-m~;.?i~::~:~?:?:?:I ; .~ . ",. : ~ __ - _. _____ n ___ ___ . . . __ n n_. n ~. __ nn __ __ __ ~ t. ~-- -- n---:011- ~ 7-n ?? --- -- __ ..~.:~ __ ??-:.____ n --.~-- "_ ........ : ! I I : I I ,I 11 L" ~ X Y ?. --------------- ----.-- -- ~~~: ~ ~~~~~~~:~~~:~~~}:_ ~ z'---_ _ _ _ _ _-l::?-_ _ _ _ _- - ' --------------. time - - - - - - - - - - - - - - - - - - - Figure 2: The Multi-State TDNN architecture , consisting of a 3-layer TDNN to estimate the a posteriori probabilities of the character states combined with word units, whose scores are derived from the word models by a Viterbi approximation of the likelihoods p(x6'IWi). units , 40 units in the hidden layer , and 78 state output units is used . There are 7 time delays both in the input and hidden layer. The softmax normalized output of the states layer is interpreted as an estimate of the probabilities of the states qj given the input window x!~~ = Xt-d .. . Xt+d for each t ime frame t , i.e . exp(1/j (t)) 2::k exp(1/k(t)) ( 1) where 1lj (t) represents the weighted sum of inputs to state unit j at time t . Based on these estimates, the output of the word units is defined to be a Viterbi approximation of the log likelihoods of the feature vector sequence given the word model Stefan Manke, Michael Finke, Alex Waibel 1098 T logp(zrlwi) ~ m~ qo L 10gp(z;~~lqt, Wi) + logp(qtlqt-I, Wi) t=1 T ~ m~ Llogp(qtlz;~~) -logp(qt) + logp(qtlqt-1, Wi). qo t=1 (2) Here, the maximum is over all possible sequences of states q'{; = qo . .. qT given a word model, p(qtlz!~~) refers to the output of the states layer as defined in (1) and p(qt) is the prior probability of observing a state qt estimated on the training data. 3.3 TRAINING OF THE RECOGNIZER During training the goal is to determine a set of parameters 0 that will maximize the posterior probability p( wlzr, 0) for all training input sequences. But in order to make that maximization computationally feasible even for a large vocabulary system we had to simplify that maximum a posteriori approach to a maximum likelihood training procedure that maximizes p(zrlw, 0) for all words instead. The first step of our maximum likelihood training is to bootstrap the recognizer using a subset of approximately 2,000 words of the training set that were labeled manually with the character boundaries to adjust the paths in the word layer correctly. After training on this hand-labeled data, the recognizer is used to label another larger set of unlabeled training data. Each pattern in this training set is processed by the recognizer. The boundaries determined automatically by the Viterbi alignment in the target word unit serve as new labels for this pattern. Then, in the second phase, the recognizer is retrained on both data sets to achieve the final performance of the recognizer. 4 EXPERIMENTS AND RESULTS We have tested our system both on writer dependent and writer independent tasks with vocabulary sizes ranging from 400 up to 20,000 words. The word recognition results are shown in Table 1. The scaling of the recognition rates with respect to the vocabulary size is plotted in Figure 3b. T a bl e 1 W' rlter depen dent an d'In depen dent recogmtIOn resu ts Writer Dependent Writer Independent Vocabulary RecognitIon KecogmtIOn Test Test Task Size Patterns Patterns Rate Rate crtAOO 400 800 98.6% 92.9% 800 wsj_l,OOO 1,000 800 97.8% wsj_7,000 7,000 2,500 89.3% wsj_l0,000 10,000 1,600 92.1% 2,500 87.7% wsj..20,000 20,000 1,600 88.9% 2,500 84.f% Dynamic Writing Information in Cursive Handwriting Recognition 1099 100r-----.------.----~------., ... writer dependent ....-. writer independent .-+- ................... 95 .......... ......-----... 90 ................................... - .................... ............... ....................-.... ----- "'-"'-'. --.. 85 80 75~----~----~----~------~ 5000 10000 15000 20000 size of vocabulary (b) Figure 3: (a) Different writing styles in the database : cursive (top), hand-printed (middle) and a mixture of both (bottom) (b) Recognition results with respect to the vocabulary size For the writer dependent evaluation, the system was trained on 2,000 patterns from a 400 word vocabulary, written by a single writer, and tested on a disjunct set of patterns from the same writer. In the writer dependent case, the training set consisted of 4,000 patterns from a 7,000 word vocabulary, written by approximately 60 different writers. The test was performed on data from an independent set of 40 writers . All data used in these experiments was collected at the University of Karlsruhe, Germany. Only minimal instructions were given to the writers. The writers were asked to write as natural as they would normally do on paper, without any restrictions in writing style. The consequence is , that the database is characterized by a high variety of different writing styles, ranging from hand-printed to strictly cursive patterns or a mixture of both writing styles (for example see Figure 3a). Additionally the native language of the writers was german, but the language of the dictionary is english. Therefore , frequent hesitations and corrections can be observed in the patterns of the database. But since this sort of input is typical for real world applications , a robust recognizer should be able to process these distorted patterns, too . From each of the writers a set of 50-100 isolated words, choosen randomly from the 7,000 word vocabulary, was collected. The used vocabularies CRT (Conference Registration Task) and WSJ (ARPA Wall Street Journal Task) were originally defined for speech recognition evaluations. These vocabularies were chosen to take advantage of the synergy effects between handwriting recognition and speech recognition, since in our case the final goal is to integrate our speech recognizer JANUS [10] and the proposed NPen++ system into a multi-modal system. 1100 5 Stefan Manke, Michael Finke, Alex Waibel CONCLUSIONS In this paper we have presented the NPen++ system, a neural recognizer for writer dependent and writer independent on-line cursive handwriting recognition. This system combines a robust input representation, which preserves the dynamic writing information, with a neural network integrating recognition and segmentation in a single framework. This architecture has been shown to be well suited for handling temporal sequences as provided by this kind of input. Evaluation of the system on different tasks with vocabulary sizes ranging from 400 to 20,000 words has shown recognition rates from 92.9% to 84.1 % in the writer independent case and from 98.6% to 88.9% in the writer dependent case. These results are especially promising because they were achieved with a small training set compared to other systems (e.g. [3]). As can be seen in Table 1, the system has proved to be virtually independent of the vocabulary. Though the system was trained on rather small vocabularies (e.g. 400 words in the writer dependent system), it generalizes well to completely different and much larger vocabularies. References [1] S. Manke and U. Bodenhausen, "A Connectionist Recognizer for Cursive Handwriting Recognition", Proceedings of the ICASSP-94, Adelaide, April 1994. [2] S. Manke, M. Finke, and A. Waibel, "Combining Bitmaps with Dynamic Writing Information for On-Line Handwriting Recognition", Proceedings of the ICPR94, Jerusalem, October 1994. [3] M. Schenkel, I. Guyon, and D. Henderson, "On-Line Cursive Script Recognition Using Time Delay Neural Networks and Hidden Markov Models", Proceedings of the ICASSP-94, Adelaide, April 1994. [4] I. Guyon, P. Albrecht, Y. Le Cun, W. Denker, and W. Hubbard, "Design of a Neural Network Character Recognizer for a Touch Terminal", Pattern Recognition, 24(2), 1991. [5] Y. Bengio and Y. LeCun. "Word Normalization for On-Line Handwritten Word Recognition", Proceedings of the ICPR-94, Jerusalem, October 1994. [6] P. Haffner and A. Waibel, "Multi-State Time Delay Neural Networks for Continuous Speech Recognition", Advances in Neural Information Processing Systems (NIPS-4) , Morgan Kaufman, 1992. [7] C. Bregler, H. Hild, S. Manke, and A. Waibel, "Improving Connected Letter Recognition by Lipreading", Proceedings of the ICASSP-93, Minneapolis, April 1993. [8] A. Waibel, T. Hanazawa, G . Hinton, K. Shiano, and K. Lang, "Phoneme Recognition using Time-Delay Neural Networks", IEEE Transactions on Acoustics, Speech and Signal Processing, March 1989. [9] W. Guerfali and R. Plamondon, "Normalizing and Restoring On-Line Handwriting", Pattern Recognition, 16(5), 1993. [10] M. Woszczyna et aI., "Janus 94: Towards Spontaneous Speech Translation", Proceedings of the ICASSP-94, Adelaide, April 1994.
908 |@word middle:2 instruction:1 score:1 past:1 bitmap:10 current:1 lang:1 written:3 fn:1 visible:1 remove:2 designed:1 npen:6 qij:3 consists:3 combine:7 multi:7 terminal:1 automatically:1 window:3 provided:4 maximizes:2 cm:1 kind:2 interpreted:1 kaufman:1 developed:1 finding:1 temporal:10 unit:7 normally:1 digitizer:3 local:7 consequence:1 path:1 approximately:2 co:1 hmms:1 limited:1 range:3 minneapolis:1 lecun:1 restoring:1 bootstrap:1 procedure:2 printed:2 word:40 integrating:1 refers:1 get:1 unlabeled:1 context:5 writing:21 restriction:1 center:1 jerusalem:2 resolution:1 rule:1 coordinate:14 target:1 spontaneous:1 us:1 pa:1 recognition:43 rec:1 native:1 labeled:2 database:3 observed:3 bottom:1 region:1 wj:1 connected:1 movement:1 asked:1 dynamic:11 trained:2 serve:1 writer:30 completely:1 icassp:4 represented:1 llogp:1 describe:1 whose:1 encoded:1 larger:2 gp:1 itself:1 hanazawa:1 final:5 sequence:24 advantage:2 frequent:1 loop:1 combining:1 achieve:2 supposed:1 description:3 wsj:2 n_:1 rescale:1 qt:4 school:1 c:1 direction:6 crt:1 wall:1 lqt:1 bregler:1 desc:1 strictly:5 dent:2 correction:1 mm:1 proximity:3 hild:1 exp:2 viterbi:3 dictionary:2 recognizer:15 integrates:3 label:2 currently:1 hubbard:1 guerfali:1 weighted:1 stefan:5 rather:1 varying:1 ira:2 derived:2 likelihood:5 baseline:2 sense:1 posteriori:3 dependent:12 nn:1 typically:1 lj:1 hidden:4 transformed:1 germany:2 j5:1 softmax:1 extraction:1 sampling:2 manually:1 represents:1 future:1 connectionist:3 simplify:1 randomly:1 preserve:3 ime:1 phase:1 consisting:2 evaluation:3 adjust:1 alignment:3 henderson:1 mixture:2 yielding:1 offeature:1 plotted:1 isolated:2 minimal:1 arpa:1 disjunct:1 modeling:3 finke:6 logp:4 maximization:1 subset:1 delay:8 inadequate:1 too:1 dependency:1 combined:1 manke:8 michael:5 mstdnn:1 recorded:3 ontext:1 style:4 albrecht:1 f6:1 de:2 wk:1 explicitly:1 performed:2 script:1 observing:1 bayes:1 sort:1 capability:2 iwi:1 accuracy:2 phoneme:1 modelled:2 raw:1 handwritten:2 basically:1 trajectory:7 stroke:2 za:1 wai:1 ed:1 against:1 handwriting:15 sampled:1 proved:1 segmentation:4 cj:2 cik:1 originally:5 x6:2 modal:1 april:4 ooo:1 though:1 hand:3 qo:3 touch:1 karlsruhe:3 effect:1 normalized:4 consisted:1 assigned:1 wdz:1 undesired:1 during:4 self:1 m:7 ranging:4 rotation:1 he:3 mellon:1 ai:1 depen:2 language:3 had:1 moving:1 aligning:1 curvature:6 posterior:2 lipreading:1 seen:1 morgan:1 additional:1 recognized:1 determine:2 maximize:1 signal:1 sliding:1 characterized:1 long:2 compensate:1 bodenhausen:1 basic:1 normalization:3 achieved:2 preserved:1 addition:1 uka:2 virtually:1 bengio:1 variety:1 equidistant:2 architecture:9 idea:2 haffner:1 shift:1 qj:3 speech:8 nine:1 constitute:1 cursive:10 transforms:1 processed:1 estimated:1 per:1 correctly:1 carnegie:1 write:1 express:1 registration:1 year:1 sum:1 letter:1 distorted:1 throughout:1 guyon:2 schenkel:1 scaling:2 fl:1 resampled:2 layer:8 scanned:1 alex:5 speed:1 optical:1 department:1 waibel:9 icpr:1 march:1 ewp:1 em:1 character:11 wi:14 cun:1 argmaxw:1 invariant:2 xo:1 computationally:1 german:1 available:1 generalizes:1 denker:1 ocr:2 neighbourhood:2 original:4 top:1 especially:1 approximating:1 bl:1 warping:1 street:1 hesitation:1 collected:2 october:2 implementation:1 design:1 unknown:1 markov:2 t:1 hinton:1 variability:1 frame:2 smoothed:1 retrained:1 woszczyna:1 bel:1 acoustic:1 nip:1 able:2 recogmtion:1 pattern:14 natural:1 representing:1 realvalued:1 tdnn:13 text:1 prior:1 relative:2 resu:1 integrate:1 translation:1 english:1 choosen:1 boundary:3 vocabulary:23 transition:2 world:1 preprocessing:7 simplified:1 transaction:1 janus:2 synergy:1 global:1 pittsburgh:1 continuous:3 pen:1 table:2 additionally:1 promising:1 reasonably:1 robust:3 improving:1 main:1 noise:1 position:1 xt:4 ognition:1 normalizing:1 ci:2 suited:3 ordered:1 goal:3 towards:1 feasible:1 plamondon:1 included:1 determined:1 typical:1 called:4 adelaide:3 evaluate:1 tested:3 handling:3
7,145
909
Forward dynamic models in human motor control: Psychophysical evidence Daniel M. Wolpert wolpert@psyche .mit .edu Zouhin Ghahramani zoubin@psyche.mit.edu Michael I. Jordan jordan@psyche.mit.edu Department of Brain & Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 Abstract Based on computational principles, with as yet no direct experimental validation, it has been proposed that the central nervous system (CNS) uses an internal model to simulate the dynamic behavior of the motor system in planning, control and learning (Sutton and Barto, 1981; Ito, 1984; Kawato et aI., 1987; Jordan and Rumelhart, 1992; Miall et aI., 1993). We present experimental results and simulations based on a novel approach that investigates the temporal propagation of errors in the sensorimotor integration process. Our results provide direct support for the existence of an internal model. 1 Introduction The notion of an internal model, a system which mimics the behavior of a natural process, has emerged as an important theoretical concept in motor control (Jordan, 1995). There are two varieties of internal models-"forward models," which mimic the causal flow of a process by predicting its next state given the current state and the motor command, and "inverse models," which are anticausal, estimating the motor command that causes a particular state transition. Forward modelsthe focus of this article-have been been shown to be of potential use for solving four fundamental problems in computational motor control. First, the delays in most sensorimotor loops are large, making feedback control infeasible for rapid 44 Daniel M. Wolpert, Zoubin Ghahramani, Michaell. Jordan movements. By using a forward model for internal feedback the outcome of an action can be estimated and used before sensory feedback is available (Ito, 1984; Miall et al., 1993). Second, a forward model is a key ingredient in a system that uses motor outflow ("efference copy") to anticipate and cancel the reafferent sensory effects of self-movement (Gallistel, 1980; Robinson et al., 1986). Third, a forward model can be used to transform errors between the desired and actual sensory outcome of a movement into the corresponding errors in the motor command, thereby providing appropriate signals for motor learning (Jordan and Rumelhart, 1992). Similarly by predicting the sensory outcome of the action, without actually performing it, a forward model can be used in mental practice to learn to select between possible actions (Sutton and Barto, 1981). Finally, a forward model can be used for state estimation in which the model's prediction of the next state is combined with a reafferent sensory correction (Goodwin and Sin, 1984). Although shown to be of theoretical importance, the existence and use of an internal forward model in the CNS is still a major topic of debate. When a subject moves his arm in the dark, he is able to estimate the visual location of his hand with some degree of accuracy. Observer models from engineering formalize the sources of information which the CNS could use to construct this estimate (Figure 1). This framework consists of a state estimation process (the observer) which is able to monitor both the inputs and outputs of the system. In particular, for the arm, the inputs are motor commands and the output is sensory feedback (e.g. vision and proprioception). There are three basic methods whereby the observer can estimate the current state (e.g. position and velocity) of the hand form these sources: It can make use of sensory inflow, it can make use of integrated motor outflow (dead reckoning), or it can combine these two sources of information via the use of a forward model. u(t) Input System X(t) Output y(t) S ensory Moto r Command Fe edback - Observer State estimate x(t) Figure 1. Observer model of state estimation . 45 Forward Dynamic Models in Human Motor Control 2 State Estimation Experiment To test between these possibilities, we carried out an experiment in which subjects made arm movements in the dark. The full details of the experiment are described in the Appendix. Three experimental conditions were studied, involving the use of null, assistive and resistive force fields. The subjects' internal estimate of hand location was assessed by asking them to localize visually the position of their hand at the end of the movement. The bias of this location estimate, plotted as a function of movement duration shows a consistent overestimation of the distance moved (Figure 2). This bias shows two distinct phases as a function of movement duration, an initial increase reaching a peak of 0.9 cm after one second followed by a sharp transition to a region of gradual decline. The variance of the estimate also shows an initial increase during the first second of movement after which it plateaus at about 2 cm 2 . External forces had distinct effects on the bias and variance propagation. Whereas the bias was increased by the assistive force and decreased by the resistive force (p < 0.05), the variance was unaffected. a - c - 1.0 1.5 E 1.0 0 0.5 (J) 0.0 E 0 (J) 0.5 <tI i:i5 CD <] -1 .0 0.0 0.5 E 0 Q) 1.0 1.5 2.0 - 2.5 (\I E 0 2.0 > - :~::- . -' 0.5 1.0 1.5 2.0 2.5 2.0 2.5 Time (5) 2 1 ..... . Q) 1.5 0 c:: c:: 1.0 <tI L... 0.0 d 0 <tI .:... -0.5 2.5 Time (5) b - ~ . -- -1.5 0.0 (\I -- <tI 0 <tI L... 0.5 <tI > <] 0.0 0.0 0.5 1.0 1.5 Time (5) 2.0 2.5 -1 -2 0.0 0.5 1.0 1.5 Time (5) Figure 2. The propagation of the (a) bias and (b) variance of the state estimate is shown, with standard error lines, against movement duration. The differential effects on (c) bias and (d) variance of the external force, assistive (dotted lines) and resistive (solid lines), are also shown relative to zero (dashed line). A positive bias represents an overestimation of the distance moved. 46 3 Daniel M. Wolpert, Zoubin Ghahramani, Michael I. Jordan Observer Model Simulation These experimental results can be fully accounted for only if we assume that the motor control system integrates the efferent outflow and the reafferent sensory inflow. To establish this conclusion we have developed an explicit model of the sensorimotor integration process which contains as special cases all three of the methods referred to above. The model-a Kalman filter (Kalman and Bucy, 1961)-is a linear dynamical system that produces an estimate of the location of the hand by monitoring both the motor outflow and the feedback as sensed, in the absence of vision, solely by proprioception. Based on these sources of information the model estimates the arm's state, integrating sensory and motor signals to reduce the overall uncertainty in its estimate. Representing the state of the hand at time t as x(t) (a 2 x 1 vector of position and velocity) acted on by a force u = [Uint, Uext]T, combining both internal motor commands and external forces, the system dynamic equations can be written in the general form of (1) x(t) = Ax(t) + Bu(t) + w(t), where A and B are matrices with appropriate dimension. The vector w(t) represents the process of white noise with an associated covariance matrix given by Q = E[w(t)w(t)T]. The system has an observable output y(t) which is linked to the actual hidden state x(t) by y(t) = Cx(t) + v(t), (2) where C is a matrix with appropriate dimension and the vector v(t) represents the output white noise which has the associated covariance matrix R = E[v(t)v(t)T]. In our paradigm, y(t) represents the proprioceptive signals (e.g. from muscle spindles and joint receptors). In particular, for the hand we approximate the system dynamics by a damped point mass moving in one dimension acted on by a force u(t). Equation 1 becomes (3) where the hand has mass m and damping coefficient {3. We assume that this system is fully observable and choose C to be the identity matrix. The parameters in the simulation, {3 = 3.9 N ,s/m, m = 4 kg and Uint = 1.5 N were chosen based on the mass of the arm and the observed relationship between time and distance traveled. The external force Uext was set to -0.3, 0 and 0.3 N for the resistive, null and assistive conditions respectively. To end the movement the sign of the motor command Uint was reversed until the arm was stationary. Noise covariance matrices of Q = 9.5 X 10- 5 [ and R = 3.3 x 1O- 4 [ were used representing a standard deviation of 1.0 cm for the position output noise and 1.8 cm s-l for the position component of the state noise. At time t = 0 the subject is given full view of his arm and, therefore, starts with an estimate X(O) = x(O) with zero bias and variance-we assume that vision calibrates the system. At this time the light is extinguished and the subject must rely on the inputs and outputs to estimate the system's state. The Kalman filter, using a Forward Dynamic Models in Human Motor Control 47 model of the system A, Band C, provides an optimal linear estimator of the state given by i(t) = ,Ax(t) + Bu(t).I + ,K(t)[y(t) - Cx(t)] I 'V' V forward model sensory correction where K(t) is the recursively updated gain matrix. The model is, therefore, a combination of two processes which together contribute to the state estimate. The first process uses the current state estimate and motor command to predict the next state by simulating the movement dynamics with a forward model. The second process uses the difference between actual and predicted reafferent sensory feedback to correct the state estimate resulting from the forward model. The relative contributions of the internal simulation and sensory correction processes to the final estimate are modulated by the Kalman gain matrix K(t) so as to provide optimal state estimates. We used this state update equation to model the bias and variance propagation and the effects of the external force. By making particular choices for the parameters of the Kalman filter, we are able to simulate dead reckoning, sensory inflow-based estimation, and forward modelbased sensorimotor integration. Moreover, to accommodate the observation that subjects generally tend to overestimate the distance that their arm has moved, we set the gain that couples force to state estimates to a value that is larger than its veridical value; B = ~ [1~4 1~6] while both A and C accurately reflected the true system. This is consistent with the independent data that subjects tend to under-reach in pointing tasks suggesting an overestimation of distance traveled (Soechting and Flanders, 1989). Simulations of the Kalman filter demonstrate the two distinct phases of bias propagation observed (Figure 3). By overestimating the force acting on the arm the forward model overestimates the distance traveled, an integrative process eventually balanced by the sensory correction. The model also captures the differential effects on bias of the externally imposed forces. By overestimating an increased force under the assistive condition, the bias in the forward model accrues more rapidly and is balanced by the sensory feedback at a higher level. The converse applies to the resistive force. In accord with the experimental results the model predicts no change in variance under the two force conditions. 4 Discussion We have shd-ttn that the Kalman filter is able to reproduce the propagation of the bias and variance of estimated position of the hand as a function of both movement duration and external forces . The Kalman filter also simulates the interesting and novel empirical result that while the variance asymptotes, the bias peaks after about one second and then gradually declines. This behavior is a consequence pf a trade off between the inaccuracies accumulating in the internal simulation of the arm's dynamics and the feedback of actual sensory information. Simple models which do not trade off the contributions of a forward model with sensory feedback, such as those based purely on sensory inflow or on motor outflow, are unable to reproduce the observed pattern of bias and variance propagation. The ability of the Kalman filter to parsimoniously model our data suggests that the processes embodied in the Daniel M. Wolpert, Zoubin Ghahramani, Michaell. Jordan 48 c a - - 1.0 E - E 0 0 en en 0.5 n1 [Ii n1 en <l 0.0 0.5 E 1.5 2.0 2.5 Time (s) b C\I 1.0 0.5 - 2.5 C\I 1.0 1.5 2.0 2.5 2.0 2.5 Time (s) d 2 E 2.0 Q) 1.5 - c:: n1 1.0 .... c:: n1 n1 0.5 n1 > 0.0 0 ~-------:.:...:.:...:,;.;.:.;.. 0.0 0 - 1.5 1.0 0.5 0.0 -0.5 -1.0 -1.5 0 Q) 0 .... 0 -1 > <l -2 0.0 0.5 1.0 1.5 Time (s) 2.0 2.5 0.0 0.5 1.0 1.5 Time (s) Figure 3. Simulated bias and variance propagation, in the same representation and scale as Figure 2, from a Kalman filter model of the sensorimotor integration process. filter, namely internal simulation through a forward model together with sensory correction, are likely to be embodied in the sensorimotor integration process. We feel that the results of this state estimation study provide strong evidence that a forward model is used by the CNS in maintaining its estimate of the hand location. Furthermore, the state estimation paradigm provides a framework to study the sensorimotor integration process in both normal and patient populations. For example, the specific predictions of the sensorimotor integration model can be tested in both patients with sensory neuropathies, who lack proprioceptive reafference, and in patients with damage to the cerebellum, a proposed site for the forward model (Miall et al., 1993). Acknowledgements We thank Peter Dayan for suggestions about the manuscript. This project was supported by grants from the McDonnell-Pew Foundation, ATR Human Information Processing Research Laboratories, Siemens Corporation, and by grant N00014-94-10777 from the Office of Naval Research. Daniel M. Wolpert and Zoubin Ghahramani are McDonnell-Pew Fellows in Cognitive Neuroscience. Michael!. Jordan is a NSF Presidential Young Investigator. Forward Dynamic Models in Human Motor Control 49 Appendix: Experimental Paradigm To investigate the way in which errors in the state estimate change over time and with external forces we used a setup (Figure 4) consisting of a combination of planar virtual visual feedback with a two degree of freedom torque motor driven manipulandum (Faye, 1986). The subject held a planar manipulandum on which his thumb was mounted. The manipulandum was used both to accurately measure the position of the subject's thumb and also, using the torque motors, to constrain the hand to move along a line across the subject's body. A projector was used to create virtual images in the plane of the movement by projecting a computer VGA screen onto a horizontal rear projection screen suspended above the manipulandum. A horizontal semi-silvered mirror was placed midway between the screen and manipulandum. The subject viewed the reflected image of the rear projection screen by looking down at the mirror; all projected images, therefore, appeared to be in the plane of the thumb, independent of head position. Projector '0' \ ,', I I \ : , I' I' \ " .,,>1. '> It; .~ Cursor Image \ \ Screen Finger Semi?silvered mirror Bulb Manlpulandum ~orque motors Figure 4. Experimental Setup Eight subjects participated and performed 300 trials each. Each trial started with the subject visually placing his thumb at a target square projected randomly on the movement line. The arm was then illuminated for two seconds, thereby allowing the subject to perceive visually his initial arm configuration. The light was then extinguished leaving just the initial target. The subject was then required to move his hand either to the left or right, as indicated by an arrow in the initial starting square. This movement was made in the absence of visual feedback of arm configuration. The subject was instructed to move until he heard a tone at which point he stopped. The timing of the tone was controlled to produce a uniform distribution of path lengths from 0-30 cm. During this movement the subject either moved in a randomly selected null or constant assistive or resistive 0.3N force field generated by the torque motors. Although it is not possible to directly probe a subject's internal representation of the state of his arm, we can examine a function of this state-the estimated visual location of the thumb. (The relationship between the state of the arm and the visual coordinates of the hand is known as 50 Daniel M. Wolpert, Zoubin Ghahramani, Michaell. Jordan the kinematic transformation; Craig, 1986.) Therefore, once at rest the subject indicated the visual estimate of his unseen thumb position using a trackball, held in his other hand, to move a cursor projected in the plane of the thumb along the movement line. The discrepancy between the actual and visual estimate of thumb location was recorded as a measure of the state estimation error. The bias and variance propagation of the state estimate was analyzed as a function of movement duration and external forces. A generalized additive model (Hastie and Tibshirani, 1990) with smoothing splines (five effective degrees of freedom) was fit to the bias and variance as a function of final position, movement duration and the interaction of the two forces with movement duration, simultaneously for main effects and for each subject. This procedure factors out the additive effects specific to each subject and, through the final position factor, the position-dependent inaccuracies in the kinematic transformation. References Craig, J. (1986). Introduction to robotics. Addison-Wesley, Reading, MA. Faye, I. (1986). An impedence controlled manipulandum for human movement studies. MS Thesis, MIT Dept. Mechanical Engineering, Cambridge, MA. Gallistel, C. (1980). The organization of action: A new synthesis. Erlbaum, Hilladale, NJ. Goodwin, G. and Sin, K. (1984). Adaptive filtering prediction and control. PrenticeHall, Englewood Cliffs, NJ. Hastie, T. and Tibshirani, R. (1990). Generalized Additive Models. Chapman and Hall, London. Ito, M. (1984). The cerebellum and neural control. Raven Press, New York. Jordan, M. and Rumelhart, D. (1992). Forward models: Supervised learning with a distal teacher. Cognitive Science, 16:307-354. Jordan, M. I. (1995). Computational aspects of motor control and motor learning. In Heuer, H. and Keele, S., editors, Handbook of Perception and Action: Motor Skills. Academic Press, New York. Kalman, R. and Bucy, R. S. (1961). New results in linear filtering and prediction. Journal of Basic Engineering (ASME), 83D:95-108. Kawato, M., Furawaka, K., and Suzuki, R. (1987). A hierarchical neural network model for the control and learning of voluntary movements. Biol.Cybern., 56:117. Miall, R., Weir, D., Wolpert, D., and Stein, J. (1993). Is the cerebellum a Smith Predictor? Journal of Motor Behavior, 25(3):203-216. Robinson, D., Gordon, J., and Gordon, S. (1986). A model of the smooth pursuit eye movement system. Biol.Cybern., 55:43-57. Soechting, J. and Flanders, M. (1989). Sensorimotor representations for pointing to targets in three- dimensional space. J.Neurophysiol., 62:582-594. Sutton, R. and Barto, A. (1981). Toward a modern theory of adaptive networks: expettation and prediction. Psychol.Rev., 88:135-170.
909 |@word trial:2 integrative:1 gradual:1 sensed:1 simulation:7 covariance:3 thereby:2 ttn:1 solid:1 accommodate:1 recursively:1 initial:5 configuration:2 contains:1 daniel:6 current:3 yet:1 written:1 must:1 additive:3 midway:1 motor:29 asymptote:1 update:1 stationary:1 selected:1 manipulandum:6 nervous:1 tone:2 plane:3 smith:1 mental:1 provides:2 contribute:1 location:7 five:1 along:2 direct:2 differential:2 gallistel:2 consists:1 resistive:6 combine:1 rapid:1 behavior:4 planning:1 examine:1 brain:1 torque:3 actual:5 pf:1 becomes:1 project:1 estimating:1 moreover:1 mass:3 null:3 kg:1 cm:5 developed:1 transformation:2 corporation:1 nj:2 temporal:1 fellow:1 ti:6 control:13 converse:1 grant:2 veridical:1 overestimate:2 before:1 positive:1 engineering:3 timing:1 consequence:1 sutton:3 receptor:1 cliff:1 solely:1 path:1 studied:1 suggests:1 practice:1 procedure:1 empirical:1 projection:2 integrating:1 zoubin:6 onto:1 cybern:2 accumulating:1 imposed:1 projector:2 starting:1 duration:7 perceive:1 estimator:1 his:10 population:1 notion:1 coordinate:1 updated:1 feel:1 target:3 us:4 velocity:2 rumelhart:3 predicts:1 observed:3 capture:1 region:1 movement:23 trade:2 balanced:2 overestimation:3 dynamic:9 solving:1 purely:1 neurophysiol:1 joint:1 finger:1 assistive:6 distinct:3 effective:1 london:1 outcome:3 emerged:1 larger:1 efference:1 presidential:1 ability:1 unseen:1 transform:1 final:3 interaction:1 loop:1 combining:1 rapidly:1 moved:4 produce:2 strong:1 predicted:1 anticausal:1 correct:1 filter:9 human:6 virtual:2 anticipate:1 correction:5 hall:1 normal:1 visually:3 predict:1 pointing:2 major:1 estimation:8 integrates:1 soechting:2 create:1 mit:4 reaching:1 barto:3 command:8 office:1 ax:2 focus:1 naval:1 dayan:1 rear:2 dependent:1 integrated:1 hidden:1 reproduce:2 overall:1 smoothing:1 integration:7 special:1 field:2 construct:1 once:1 chapman:1 represents:4 placing:1 cancel:1 uint:3 mimic:2 discrepancy:1 spline:1 extinguished:2 overestimating:2 gordon:2 modern:1 randomly:2 simultaneously:1 parsimoniously:1 phase:2 consisting:1 cns:4 n1:6 freedom:2 organization:1 englewood:1 possibility:1 investigate:1 kinematic:2 analyzed:1 light:2 damped:1 held:2 damping:1 desired:1 plotted:1 causal:1 theoretical:2 stopped:1 increased:2 asking:1 deviation:1 uniform:1 predictor:1 delay:1 erlbaum:1 bucy:2 teacher:1 combined:1 fundamental:1 peak:2 spindle:1 bu:2 off:2 michael:3 together:2 modelbased:1 synthesis:1 prenticehall:1 thesis:1 central:1 recorded:1 choose:1 cognitive:3 dead:2 external:8 suggesting:1 potential:1 coefficient:1 performed:1 view:1 observer:6 linked:1 start:1 contribution:2 square:2 accuracy:1 variance:14 reckoning:2 who:1 thumb:8 accurately:2 craig:2 monitoring:1 unaffected:1 plateau:1 reach:1 against:1 sensorimotor:9 associated:2 efferent:1 gain:3 couple:1 massachusetts:1 formalize:1 actually:1 manuscript:1 wesley:1 higher:1 supervised:1 reflected:2 planar:2 furthermore:1 just:1 until:2 hand:14 horizontal:2 propagation:9 lack:1 indicated:2 effect:7 concept:1 true:1 laboratory:1 proprioceptive:2 white:2 distal:1 cerebellum:3 sin:2 during:2 self:1 whereby:1 m:1 generalized:2 asme:1 demonstrate:1 image:4 novel:2 kawato:2 he:3 cambridge:2 ai:2 pew:2 similarly:1 had:1 moving:1 driven:1 n00014:1 suspended:1 muscle:1 faye:2 paradigm:3 dashed:1 semi:2 signal:3 full:2 ii:1 smooth:1 academic:1 controlled:2 prediction:5 involving:1 basic:2 vision:3 patient:3 inflow:4 accord:1 robotics:1 whereas:1 participated:1 decreased:1 leaving:1 source:4 neuropathy:1 rest:1 subject:21 tend:2 simulates:1 flow:1 proprioception:2 jordan:12 variety:1 fit:1 hastie:2 reduce:1 decline:2 peter:1 york:2 cause:1 action:5 generally:1 heard:1 dark:2 stein:1 band:1 nsf:1 dotted:1 sign:1 estimated:3 neuroscience:1 tibshirani:2 key:1 four:1 monitor:1 localize:1 shd:1 inverse:1 i5:1 uncertainty:1 appendix:2 investigates:1 illuminated:1 followed:1 calibrates:1 constrain:1 aspect:1 simulate:2 performing:1 acted:2 department:1 combination:2 mcdonnell:2 across:1 psyche:3 rev:1 making:2 projecting:1 gradually:1 equation:3 eventually:1 addison:1 end:2 available:1 pursuit:1 eight:1 probe:1 hierarchical:1 appropriate:3 simulating:1 uext:2 existence:2 maintaining:1 ghahramani:6 establish:1 psychophysical:1 move:5 damage:1 distance:6 reversed:1 unable:1 simulated:1 thank:1 atr:1 topic:1 toward:1 kalman:11 length:1 relationship:2 providing:1 setup:2 fe:1 debate:1 allowing:1 observation:1 voluntary:1 looking:1 head:1 sharp:1 michaell:3 goodwin:2 namely:1 required:1 mechanical:1 inaccuracy:2 robinson:2 able:4 dynamical:1 pattern:1 perception:1 appeared:1 reading:1 natural:1 force:21 rely:1 predicting:2 arm:15 representing:2 technology:1 eye:1 started:1 carried:1 psychol:1 embodied:2 traveled:3 acknowledgement:1 relative:2 fully:2 interesting:1 suggestion:1 mounted:1 filtering:2 ingredient:1 validation:1 foundation:1 degree:3 bulb:1 consistent:2 article:1 principle:1 editor:1 cd:1 accounted:1 supported:1 placed:1 copy:1 infeasible:1 bias:18 institute:1 feedback:11 dimension:3 transition:2 sensory:20 forward:24 made:2 instructed:1 projected:3 adaptive:2 suzuki:1 miall:4 approximate:1 observable:2 skill:1 keele:1 handbook:1 learn:1 main:1 arrow:1 noise:5 reafference:1 outflow:5 body:1 site:1 referred:1 en:3 screen:5 position:12 explicit:1 accrues:1 flanders:2 third:1 ito:3 young:1 externally:1 down:1 specific:2 evidence:2 raven:1 importance:1 mirror:3 cursor:2 wolpert:8 cx:2 likely:1 visual:7 applies:1 ma:3 identity:1 viewed:1 weir:1 absence:2 change:2 acting:1 experimental:7 siemens:1 select:1 internal:12 support:1 modulated:1 assessed:1 investigator:1 dept:1 tested:1 biol:2
7,146
91
652 Scaling Properties of Coarse-Coded Symbol Memories Ronald Rosenfeld David S. Touretzky Computer Science Department Carnegie Mellon University Pittsburgh, Pennsylvania 15213 Abstract: Coarse-coded symbol memories have appeared in several neural network symbol processing models. In order to determine how these models would scale, one must first have some understanding of the mathematics of coarse-coded representations. We define the general structure of coarse-coded symbol memories and derive mathematical relationships among their essential parameters: memory 8ize, 8ymbol-8et size and capacity. The computed capacity of one of the schemes agrees well with actual measurements oC tbe coarse-coded working memory of DCPS, Touretzky and Hinton's distributed connectionist production system. 1 Introduction A di8tributed repre8entation is a memory scheme in which each entity (concept, symbol) is represented by a pattern of activity over many units [3]. If each unit participates in the representation of many entities, it is said to be coar8ely tuned, and the memory itself is called a coar8e-coded memory. Coarse-coded memories have been used for storing symbols in several neural network symbol processing models, such as Touretzky and Hinton's distributed connectionist production system DCPS [8,9], Touretzky's distributed implementation of linked list structures on a Boltzmann machine, BoltzCONS [10], and St. John and McClelland's PDP model of case role defaults [6]. In all of these models, memory capacity was measured empirically and parameters were adjusted by trial and error to obtain the desired behavior. We are now able to give a mathematical foundation to these experiments by analyzing the relationships among the fundamental memory parameters. There are several paradigms for coarse-coded memories. In a feature-based repre8entation, each unit stands for some semantic feature. Binary units can code features with binary values, whereas more complicated units or groups of units are required to code more complicated features, such as multi-valued properties or numerical values from a continuous scale. The units that form the representation of a concept define an intersection of features that constitutes that concept. Similarity between concepts composed of binary Ceatures can be measured by the Hamming distance between their representations. In a neural network implementation, relationships between concepts are implemented via connections among the units forming their representations. Certain types of generalization phenomena thereby emerge automatically. A different paradigm is used when representing points in a multidimensional continuous space [2,3]. Each unit encodes values in some subset of the space. Typically the @ American Institute of Physics 1988 653 subsets are hypercubes or hyperspheres, but they may be more coarsely tuned along some dimensions than others [1]. The point to be represented is in the subspace formed by the intersection of all active units. AB more units are turned on, the accuracy of the representation improves. The density and degree of overlap of the units' receptive fields determines the system's resolution [7]. Yet another paradigm for coarse-coded memories, and the one we will deal with exclusively, does not involve features. Each concept, or symbol, is represented by an arbitrary subset of the units, called its pattern. Unlike in feature-based representations, the units in the pattern bear no relationship to the meaning of the symbol represented. A symbol is stored in memory by turning on all the units in its pattern. A symbol is deemed present if all the units in its pattern are active. l The receptive field of each unit is defined as the set of all symbols in whose pattern it participates. We call such memories coarsecoded symbol memories (CCSMs). We use the term "symbol" instead of "concept" to emphasize that the internal structure of the entity to be represented is not involved in its representation. In CCSMs, a short Hamming distance between two symbols does not imply semantic similarity, and is in general an undesirable phenomenon. The efficiency with which CCSMs handle sparse memories is the major reason they have been used in many connectionist systems, and hence the major reason for studying them here. The unit-sharing strategy that gives rise to efficient encoding in CCSMs is also the source of their major weakness. Symbols share units with other symbols. AB more symbols are stored, more and more of the units are turned on. At some point, some symbol may be deemed present in memory because all of its units are turned on, even though it was not explicitly stored: a "ghost" is born. Ghosts are an unwanted phenomenon arising out of the overlap among the representations of the various symbols. The emergence of ghosts marks the limits of the system's capacity: the number of symbols it can store simultaneously and reliably. 2 Definitions and Fundamental Parameters A coarse coded symbol memory in its most general form consists of: ? A set of N binary state units. ? An alphabet of Q symbols to be represented. Symbols in this context are atomic entities: they have no constituent structure. ? A memory scheme, which is a function that maps each symbol to a subset of the units - its pattern. The receptive field of a unit is defined as the set of all symbols to whose pattern it belongs (see Figure 1). The exact nature of the lThis criterion can be generalized by introducing a visibility threshold: a fraction of the pattern that should be on in order for a symbol to be considered present. Our analysis deals only with a visibility criterion of 100%, but can be generalized to accommodate nOise. 654 I I 81 I 82 I 88 I 8 I 85 I 86 I 87 I 88 I 4 Ul U2 U8 U4 U5 U6 ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? Figure 1: A memory scheme (N = 6, Q = 8) defined in terms of units Us and symbols 8;. The columns are the symbols' patterns. The rows are the units' receptive fieldB. memory scheme mapping determines the properties of the memory, and is the central target of our investigation. As symbols are stored, the memory fills up and ghosts eventually appear. It is not possible to detect a ghost simply by inspecting the contents of memory, since there is no general way of distinguishing a symbol that was stored from one that emerged out of overlaps with other symbols. (It is sometimes possible, however, to conclude that there are no ghosts.) Furthermore, a symbol that emerged as a ghost at one time may not be a ghost at a later time if it was subsequently stored into memory. Thus the definition of a ghost depends not only on the state of the memory but also on its history. Some memory schemes guarantee that no ghost will emerge as long as the number of symbols stored does not exceed some specified limit. In other schemes, the emergence of ghosts is an ever-present possibility, but its probability can be kept arbitrarily low by adjusting other parameters. We analyze systems of both types. First, two more bits of notation need to be introduced: Pghost: Probability of a ghost. The probability that at least one ghost will appear after some number of symbols have been stored. k: Capacity. The maximum number of symbols that can be stored simultaneously before the probability of a ghost exceeds a specified threshold. If the threshold is 0, we say that the capacity is guaranteed. A localist representation, where every symbol is represented by a single unit and every unit is dedicated to the representation of a single symbol, can now be viewed as a special case of coarse-coded memory, where k = N = Q and Pghost = o. Localist representations are well suited for memories that are not sparse. In these cases, coarsecoded memories are at a disadvantage. In designing coarse-coded symbol memories we are interested in cases where k ? N ? Q. The permissible probability for a ghost in these systems should be low enough so that its impact can be ignored. 655 3 3.1 Analysis of Four Memory Schemes Bounded Overlap (guaranteed capacity) If we want to construct the memory scheme with the largest possible a (given Nand k) while guaranteeing Pghost = 0, the problem can be stated formally as: Given a set of size N, find the largest collection of subsets of it such that no union of k such subsets subsumes any other subset in the collection. This is a well known problem in Coding Theory, in slight disguise. Unfortunately, no complete analytical solution is known. We therefore simplify our task and consider only systems in which all symbols are represented by the same number of units (i.e. all patterns are of the same size). In mathematical terms, we restrict ourselves to constant weight codes. The problem then becomes: Given a set of size N, find the largest collection of subsets of size exactly L such that no union of k such subsets subsumes any other subset in the collection. There are no known complete analytical solutions for the size of the largest collection of patterns even when the patterns are of a fixed size. Nor is any efficient procedure for constructing such a collection known. We therefore simplify the problem further. We now restrict our consideration to patterns whose pairwise overlap is bounded by a given number. For a given pattern size L and desired capacity k, we require that no two patterns overlap in more than m units, where: _lL -k 1J m- (1) -- Memory schemes that obey this constraint are guaranteed a capacity of at least k symbols, since any k symbols taken together can overlap at most L - 1 units in the pattern of any other symbol - one unit short of making it a ghost. Based on this constraint, our mathematical problem now becomes: Given a set of size N, find the largest collection of subsets of size exactly L such that the intersection of any two such subsets is of size ~ m (where m is given by equation 1.) Coding theory has yet to produce a complete solution to this problem, but several methods of deriving upper bounds have been proposed (see for example [4]). The simple formula we use here is a variant of the Johnson Bound. Let abo denote the maximum number of symbols attainable in memory schemes that use bounded overlap. Then (m~l) (m~l) (2) 656 The Johnson bound is known to be an exact solution asymptotically (that is, when N, L, m -+ 00 and their ratios remain finite). Since we are free to choose the pattern size, we optimize our memory scheme by maximizing the above expression over all possible values of L. For the parameter subspace we are interested in here (N < 1000, k < 50) we use numerical approximation to obtain: < max ( LeII,N] N L - m )m+l < (3) (Recall that m is a function of Land k.) Thus the upper bound we derived depicts a simple exponential relationship between Q and N/k. Next, we try to construct memory schemes of this type. A Common Lisp program using a modified depth-first search constructed memory schemes for various parameter values, whose Q'S came within 80% to 90% of the upper bound. These results are far from conclusive, however, since only a small portion of the parameter space was tested. In evaluating the viability of this approach, its apparent optimality should be contrasted with two major weaknesses. First, this type of memory scheme is hard to construct computationally. It took our program several minutes of CPU time on a Symbolics 3600 to produce reasonable solutions for cases like N = 200, k = 5, m = 1, with an exponential increase in computing time for larger values of m. Second, if CCSMs are used as models of memory in naturally evolving systems (such as the brain), this approach places too great a burden on developmental mechanisms. The importance of the bounded overlap approach lies mainly in its role as an upper bound for all possible memory schemes, subject to the simplifications made earlier. All schemes with guaranteed capacities can be measured relative to equation 3. 3.2 Random Fixed Size Patterns (a stochastic approach) Randomly produced memory schemes are easy to implement and are attractive because of their naturalness. However, if the patterns of two symbols coincide, the guaranteed capacity will be zero (storing one of these symbols will render the other a ghost). We therefore abandon the goal of guaranteeing a certain capacity, and instead establish a tolerance level for ghosts, Pghost. For large enough memories, where stochastic behavior is more robust, we may expect reasonable capacity even with very small Pghost. In the first stochastic approach we analyze, patterns are randomly selected subsets of a fixed size L. Unlike in the previous approach, choosing k does not bound Q. We may define as many symbols as we wish, although at the cost of increased probability of a ghost (or, alternatively, decreased capacity). The probability of a ghost appearing after k symbols have been stored is given by Equation 4: (4) 657 TN,L(k, e) is the probability that exactly e units will be active after k symbols have been stored. It is defined recursively by Equation 5": TN,L(O,O) = 1 TN,L(k, e) 0 for either k 0 and e 1= 0, or k > 0 and e < L TN,L(k, e) = E~=o T(k - 1, c - a) . (N-~-a)) . (~:~)/(~) = = (5) We have constructed various coarse-coded memories with random fixed-size receptive fields and measured their capacities. The experimental results show good agreement with the above equation. The optimal pattern size for fixed values of N, k, and a can be determined by binary search on Equation 4, since Pghost(L) has exactly one maximum in the interval [1, N]. However, this may be expensive for large N. A computational shortcut can be achieved by estimating the optimal L and searching in a small interval around it. A good initial estimate is derived by replacing the summation in Equation 4 with a single term involving E[e]: the expected value of the number of active units after k symbols have been stored. The latter can be expressed as: The estimated L is the one that maximizes the following expression: An alternative formula, developed by Joseph Tebelskis, produces very good approximations to Eq. 4 and is much more efficient to compute. After storing k symbols in memory, the probability Pz that a single arbitrary symbol x has become a ghost is given by: Pz(N,L,k,a) .(L) (N L_i)k / (N)k L L i = f.(-1)' (6) If we now assume that each symbol's Pz is independent of that of any other symbol, we obtain: (7) This assumption of independence is not strictly true, but the relative error was less than 0.1% for the parameter ranges we considered, when Pghost was no greater than 0.01. We have constructed the two-dimensional table TN,L(k, c) for a wide range of (N, L) values (70 ~ N ~ 1000, 7 ~ L ~ 43), and produced graphs of the relationships between N, k, a, and Pghost for optimum pattern sizes, as determined by Equation 4. The 658 results show an approximately exponential relationship between a and N /k [5]. Thus, for a fixed number of symbols, the capacity is proportional to the number of units. Let arl p denote the maximum number of symbols attainable in memory schemes that use random fixed-size patterns. Some typical relationships, derived from the data, are: ~ 0.0086. eO.46S f arlp(Pghost = 0.001) ~ O.OOOS. eO. 47S f arlP(Pghost 3.3 = 0.01) (8) Random Receptors (a stochastic approach) A second stochastic approach is to have each unit assigned to each symbol with an independent fixed probability s. This method lends itself to easy mathematical analysis, resulting in a closed-form analytical solution. After storing k symbols, the probability that a given unit is active is 1 - (1 - s)k (independent of any other unit). For a given symbol to be a ghost, every unit must either be active or else not belong to that symbol's pattern. That will happen with a probability [1 - s . (1 - s)k] N, and thus the probability of a ghost is: (9) Pghost(a, N, k,s) Assuming Pghost simplified to: ? 1 and k ? a (both hold in our case), the expression can be Pghost(a,N,k,s) a? [1- s. (1- s)k]N from which a can be extracted: arr(N, k, 8, Pghost) (10) We can now optimize by finding the value of s that maximizes a, given any desired upper bound on the expected value of Pghost. This is done straightforwardly by solving Ba/Bs = o. Note that 8? N corresponds to L in the previous approach. The solution is s = l/(k + 1), which yields, after some algebraic manipulation: (11) A comparison of the results using the two stochastic approaches reveals an interesting similarity. For large k, with Pghost = 0.01 the term 0.468/k of Equation 8 can be seen as a numerical approximation to the log term in Equation 11, and the multiplicative factor of 0.0086 in Equation 8 approximates Pghost in Equation 11. This is hardly surprising, since the Law of Large Numbers implies that in the limit (N, k -+ 00, with 8 fixed) the two methods are equivalent. 659 Finally, it should be. noted that the stochastic approaches we analyzed generate a family of memory schemes, with non-identical ghost-probabilities. Pghost in our formulas is therefore better understood as an expected value, averaged over the entire family. 3.4 Partitioned Binary Coding (a reference point) The last memory scheme we analyze is not strictly distributed. Rather, it is somewhere in between a distributed and a localist representation, and is presented for comparison with the previous results. For a given number of units N and desired capacity k, the units are partitioned into k equal-size "slots," each consisting of N / k units (for simplicity we assume that k divides N). Each slot is capable of storing exactly one symbol. The most efficient representation for all possible symbols that may be stored into a slot is to assign them binary codes, using the N / k units of each slot as bits. This would allow 2N Jic symbols to be represented. Using binary coding, however, will not give us the required capacity of 1 symbol, since binary patterns subsume one another. For example, storing the code '10110' into one of the slots will cause the codes '10010', '10100' and '00010' (as well as several other codes) to become ghosts. A possible solution is to use only half of the bits in each slot for a binary code, and set the other half to the binary complement of that code (we assume that N/k is even). This way, the codes are guaranteed not to subsume one another. Let Qpbc denote the number of symbols representable using a partitioned binary coding scheme. Then, '"pbc -_ ..... 2NJ2Ic -- eO.847 !:!-.. (12) Once again, Q is exponential in N /k. The form of the result closely resembles the estimated upper bound on the Bounded Overlap method given in Equation 3. There is also a strong resemblance to Equations 8 and 11, except that the fractional multiplier in front of the exponential, corresponding to Pghost, is missing. Pghost is 0 for the Partitioned Binary Coding method, but this is enforced by dividing the memory into disjoint sets of units rather than adjusting the patterns to reduce overlap among symbols. As mentioned previously, this memory scheme is not really distributed in the sense used in this paper, since there is no one pattern associated with a symbol. Instead, a symbol is represented by anyone of a set of k patterns, each N /k bits long, corresponding to its appearance in one of the k slots. To check whether a symbol is present, all k slots must be examined. To store a new symbol in memory, one must scan the k slots until an empty one is found. Equation 12 should therefore be used only as a point of reference. 4 Measurement of DCPS The three distributed schemes we have studied all use unstructured patterns, the only constraint being that patterns are at least roughly the same size. Imposing more complex structure on any of these schemes may is likely to reduce the capacity somewhat. In 660 Memory Scheme Bounded Overlap Random Fixed-size Patterns Random Receptors Partitioned Binary Coding Qbo(N, k) < Result eO.367 t r Q,,!p(Pghost = 0.01) ~ 0.0086. e?.468 Q,,!p(Pghost = 0.001) ~ 0.0008 . e?.473f Q _ P . e N .1og (k+1)"'Tl/((k+l)"'Tl_k"') ,.,. - ghost eO.347r Qpbc -- Table 1 Summary of results for various memory schemes. order to quantify this effect, we measured the memory capacity of DCPS (BoltzCONS uses the same memory scheme) and compared the results with the theoretical models analyzed above. DCPS' memory scheme is a modified version of the Random Receptors method [5]. The symbol space is the set of all triples over a 25 letter alphabet. Units have fixed-size receptive fields organized as 6 x 6 x 6 subspaces. Patterns are manipulated to minimize the variance in pattern size across symbols. The parameters for DCPS are: N = 2000, Q = 25 3 = 15625, and the mean pattern size is (6/25)3 x 2000 = 27.65 with a standard deviation of 1.5. When Pghost = 0.01 the measured capacity was k = 48 symbols. By substituting for N in Equation 11 we find that the highest k value for which Q,.,. ~ 15625 is 51. There does not appear to be a significant cost for maintaining structure in the receptive fields. 5 Summary and Discussion Table 1 summarizes the results obtained for the four methods analyzed. Some differences must be emphasiz'ed: and Qpbc deal with guaranteed capacity, whereas Q,.!p and Q,.,. are meaningful only for Pghost > O. ? Qbo ? Qbo is only an upper bound. ? Q,.!p is based on numerical estimates. ? Qpbc is based on a scheme which is not strictly coarse-coded. The similar functional form of all the results, although not surprising, is aesthetically pleasing. Some of the functional dependencies among the various parameters_ can be derived informally using qualitative arguments. Only a rigorous analysis, however, can provide the definite answers that are needed for a better understanding of these systems and their scaling properties. 661 Acknowledgments We thank Geoffrey Hinton, Noga Alon and Victor Wei for helpful comments, and Joseph Tebelskis for sharing with us his formula for approximating Pghost in the case of fixed pattern sizes. This work was supported by National Science Foundation grants IST-8516330 and EET-8716324, and by the Office of Naval Research under contract number NOOO14-86K-0678. The first author was supported by a National Science Foundation graduate fellowship. References [1] Ballard, D H. (1986) Cortical connections and parallel processing: structure and function. Behavioral and Brain Sciences 9(1). [2] Feldman, J. A., and Ballard, D. H. (1982) Connectionist models and their properties. Cognitive Science 6, pp. 205-254. [3] Hinton, G. E., McClelland, J. L., and Rumelhart, D. E. (1986) Distributed representations. In D. E. Rumelhart and J. L. McClelland (eds.), Parallel Distributed Processing: Explorations in the Microstructure of Cognition, volume 1. Cambridge, MA: MIT Press. [4] Macwilliams, F.J., and Sloane, N.J.A. (1978). The Theory of Error-Correcting Codes, North-Holland. [5] Rosenfeld, R. and Touretzky, D. S. (1987) Four capacity models for coarse-coded symbol memories. Technical report CMU-CS-87-182, Carnegie Mellon University Computer Science Department, Pittsburgh, PA. [6] St. John, M. F. and McClelland, J. L. (1986) Reconstructive memory for sentences: a PDP approach. Proceedings of the Ohio University Inference Conference. [7] Sullins, J. (1985) Value cell encoding strategies. Technical report TR-165, Computer Science Department, University of Rochester, Rochester, NY. [8] Touretzky, D. S., and Hinton, G. E. (1985) Symbols among the neurons: details of a connectionist inference architecture. Proceedings of IJCAI-85, Los Angeles, CA, pp. 238-243. [9] Touretzky, D. S., and Hinton, G. E. (1986) A distributed connectionist production system. Technical report CMU-CS-86-172, Computer Science Department, Carnegie Mellon University, Pittsburgh, PA. [10] Touretzky, D. S. (1986) BoltzCONS: reconciling connectionism with the recursive nature of stacks and trees. Proceedings of the Eighth A nnual Conference of the Cognitive Science Society, Amherst, MA, pp. 522-530.
91 |@word trial:1 version:1 attainable:2 thereby:1 tr:1 accommodate:1 recursively:1 initial:1 born:1 exclusively:1 tuned:2 surprising:2 yet:2 must:5 john:2 ronald:1 numerical:4 happen:1 visibility:2 half:2 selected:1 short:2 coarse:14 mathematical:5 along:1 constructed:3 become:2 qualitative:1 consists:1 behavioral:1 pairwise:1 expected:3 roughly:1 behavior:2 nor:1 multi:1 brain:2 automatically:1 actual:1 cpu:1 becomes:2 estimating:1 notation:1 bounded:6 maximizes:2 developed:1 finding:1 guarantee:1 every:3 multidimensional:1 unwanted:1 exactly:5 unit:45 grant:1 appear:3 before:1 understood:1 limit:3 receptor:3 encoding:2 analyzing:1 approximately:1 resembles:1 examined:1 studied:1 range:2 graduate:1 averaged:1 acknowledgment:1 atomic:1 union:2 recursive:1 implement:1 definite:1 procedure:1 evolving:1 undesirable:1 context:1 optimize:2 equivalent:1 map:1 missing:1 maximizing:1 resolution:1 simplicity:1 unstructured:1 correcting:1 fill:1 deriving:1 u6:1 his:1 handle:1 searching:1 target:1 ooos:1 exact:2 distinguishing:1 designing:1 us:1 agreement:1 pa:2 rumelhart:2 expensive:1 u4:1 role:2 highest:1 mentioned:1 developmental:1 solving:1 ymbol:1 efficiency:1 represented:10 various:5 alphabet:2 reconstructive:1 choosing:1 whose:4 emerged:2 apparent:1 valued:1 larger:1 say:1 rosenfeld:2 emergence:2 itself:2 abandon:1 analytical:3 took:1 arr:1 turned:3 constituent:1 los:1 ijcai:1 empty:1 optimum:1 produce:3 guaranteeing:2 derive:1 alon:1 measured:6 eq:1 strong:1 dividing:1 implemented:1 aesthetically:1 c:2 arl:1 implies:1 quantify:1 closely:1 subsequently:1 stochastic:7 exploration:1 require:1 assign:1 microstructure:1 generalization:1 really:1 investigation:1 connectionism:1 inspecting:1 summation:1 adjusted:1 strictly:3 hold:1 around:1 considered:2 great:1 mapping:1 cognition:1 substituting:1 major:4 agrees:1 largest:5 mit:1 modified:2 rather:2 og:1 office:1 derived:4 naval:1 check:1 mainly:1 rigorous:1 detect:1 sense:1 helpful:1 inference:2 typically:1 entire:1 nand:1 interested:2 among:7 special:1 field:6 construct:3 equal:1 once:1 identical:1 constitutes:1 connectionist:6 others:1 simplify:2 report:3 di8tributed:1 randomly:2 composed:1 simultaneously:2 manipulated:1 national:2 ourselves:1 consisting:1 ab:2 pleasing:1 possibility:1 weakness:2 analyzed:3 capable:1 tree:1 divide:1 desired:4 theoretical:1 increased:1 column:1 earlier:1 disadvantage:1 localist:3 cost:2 introducing:1 deviation:1 subset:13 johnson:2 too:1 front:1 stored:13 straightforwardly:1 dependency:1 answer:1 st:2 hypercubes:1 fundamental:2 density:1 amherst:1 contract:1 participates:2 physic:1 together:1 again:1 central:1 choose:1 disguise:1 cognitive:2 american:1 coding:7 subsumes:2 north:1 explicitly:1 depends:1 later:1 try:1 multiplicative:1 closed:1 linked:1 analyze:3 portion:1 complicated:2 parallel:2 rochester:2 minimize:1 formed:1 accuracy:1 variance:1 yield:1 produced:2 coarsecoded:2 history:1 touretzky:8 sharing:2 ed:2 definition:2 pp:3 involved:1 naturally:1 associated:1 hamming:2 pbc:1 adjusting:2 recall:1 fractional:1 improves:1 organized:1 wei:1 done:1 though:1 furthermore:1 until:1 working:1 replacing:1 resemblance:1 effect:1 concept:7 ize:1 true:1 multiplier:1 hence:1 assigned:1 nooo14:1 semantic:2 deal:3 attractive:1 ll:1 noted:1 oc:1 criterion:2 generalized:2 dcps:6 sloane:1 complete:3 tn:5 dedicated:1 boltzcons:3 meaning:1 consideration:1 ohio:1 common:1 functional:2 empirically:1 volume:1 belong:1 slight:1 approximates:1 mellon:3 measurement:2 significant:1 cambridge:1 imposing:1 feldman:1 mathematics:1 similarity:3 belongs:1 manipulation:1 store:2 certain:2 binary:14 arbitrarily:1 came:1 victor:1 seen:1 greater:1 somewhat:1 eo:5 determine:1 paradigm:3 exceeds:1 technical:3 long:2 coded:15 impact:1 variant:1 involving:1 cmu:2 sometimes:1 achieved:1 cell:1 whereas:2 want:1 fellowship:1 decreased:1 interval:2 else:1 source:1 permissible:1 noga:1 unlike:2 comment:1 subject:1 lisp:1 call:1 exceed:1 enough:2 viability:1 easy:2 independence:1 pennsylvania:1 restrict:2 architecture:1 reduce:2 angeles:1 whether:1 expression:3 ul:1 render:1 algebraic:1 cause:1 hardly:1 ignored:1 involve:1 informally:1 u5:1 mcclelland:4 generate:1 estimated:2 arising:1 disjoint:1 carnegie:3 coarsely:1 group:1 ist:1 four:3 threshold:3 kept:1 asymptotically:1 graph:1 fraction:1 tbe:1 enforced:1 letter:1 place:1 family:2 reasonable:2 summarizes:1 scaling:2 bit:4 bound:10 guaranteed:7 simplification:1 activity:1 constraint:3 encodes:1 tebelskis:2 anyone:1 optimality:1 argument:1 department:4 representable:1 remain:1 across:1 partitioned:5 joseph:2 making:1 b:1 taken:1 computationally:1 equation:16 nnual:1 previously:1 eventually:1 mechanism:1 needed:1 studying:1 naturalness:1 obey:1 appearing:1 alternative:1 symbolics:1 reconciling:1 maintaining:1 somewhere:1 establish:1 approximating:1 society:1 receptive:7 strategy:2 said:1 lends:1 subspace:3 distance:2 thank:1 capacity:23 entity:4 reason:2 assuming:1 code:11 relationship:8 ratio:1 unfortunately:1 stated:1 rise:1 ba:1 implementation:2 reliably:1 boltzmann:1 upper:7 neuron:1 finite:1 subsume:2 hinton:6 ever:1 pdp:2 stack:1 arbitrary:2 david:1 introduced:1 complement:1 required:2 specified:2 connection:2 conclusive:1 sentence:1 able:1 pattern:36 eighth:1 appeared:1 ghost:26 program:2 max:1 memory:59 hyperspheres:1 overlap:12 turning:1 representing:1 scheme:30 imply:1 deemed:2 understanding:2 relative:2 law:1 expect:1 bear:1 interesting:1 proportional:1 geoffrey:1 triple:1 foundation:3 degree:1 storing:6 share:1 land:1 production:3 row:1 summary:2 supported:2 last:1 free:1 allow:1 institute:1 wide:1 emerge:2 sparse:2 abo:1 distributed:10 tolerance:1 default:1 dimension:1 stand:1 depth:1 evaluating:1 cortical:1 jic:1 collection:7 made:1 coincide:1 simplified:1 author:1 far:1 emphasize:1 eet:1 active:6 reveals:1 pittsburgh:3 conclude:1 alternatively:1 continuous:2 search:2 lthis:1 table:3 nature:2 ballard:2 robust:1 ca:1 complex:1 constructing:1 noise:1 tl:1 depicts:1 ny:1 wish:1 exponential:5 lie:1 formula:4 minute:1 symbol:75 list:1 pz:3 essential:1 burden:1 importance:1 suited:1 intersection:3 simply:1 appearance:1 likely:1 forming:1 expressed:1 u2:1 holland:1 corresponds:1 determines:2 extracted:1 ma:2 slot:9 viewed:1 goal:1 u8:1 content:1 hard:1 shortcut:1 determined:2 typical:1 contrasted:1 except:1 called:2 experimental:1 meaningful:1 formally:1 internal:1 mark:1 latter:1 scan:1 tested:1 phenomenon:3
7,147
910
Unsupervised Classification of 3D Objects from 2D Views Satoshi Suzuki Hiroshi Ando ATR Human Information Processing Research Laboratories 2-2 Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-02, Japan satoshi@hip.atr.co.jp, ando@hip.atr.co.jp Abstract This paper presents an unsupervised learning scheme for categorizing 3D objects from their 2D projected images. The scheme exploits an auto-associative network's ability to encode each view of a single object into a representation that indicates its view direction. We propose two models that employ different classification mechanisms; the first model selects an auto-associative network whose recovered view best matches the input view, and the second model is based on a modular architecture whose additional network classifies the views by splitting the input space nonlinearly. We demonstrate the effectiveness of the proposed classification models through simulations using 3D wire-frame objects. 1 INTRODUCTION The human visual system can recognize various 3D (three-dimensional) objects from their 2D (two-dimensional) retinal images although the images vary significantly as the viewpoint changes. Recent computational models have explored how to learn to recognize 3D objects from their projected views (Poggio & Edelman, 1990). Most existing models are, however, based on supervised learning, i.e., during training the teacher tells which object each view belongs to. The model proposed by Weinshall et al. (1990) also requires a signal that segregates different objects during training. This paper, on the other hand, discusses unsupervised aspects of 3D object recognition where the system discovers categories by itself. 950 Satoshi Suzuki, Hiroshi Ando This paper presents an unsupervised classification scheme for categorizing 3D objects from their 2D views. The scheme consists of a mixture of 5-layer auto-associative networks, each of which identifies an object by non-linearly encoding the views into a representation that describes transformation of a rigid object. A mixture model with linear networks was also studied by Williams et al. (1993) for classifying objects under affine transformations. We propose two models that employ different classification mechanisms. The first model classifies the given view by selecting an auto-associative network whose recovered view best matches the input view. The second model is based on the modular architecture proposed by Jacobs et al. (1991) in which an additional 3-layer network classifies the views by directly splitting the input space. The simulations using 3D wire-frame objects demonstrate that both models effectively learn to classify each view as a 3D object. This paper is organized as follows. Section 2 describes in detail the proposed models for unsupervised classification of 3D objects. Section 3 describes the simulation results using 3D wire-frame objects. In these simulations, we test the performance of the proposed models and examine what internal representations are acquired in the hidden layers. Finally, Section 4 concludes this paper. 2 THE NETWORK MODELS This section describes an unsupervised scheme that classifies 2D views into 3D objects. We initially examined classical unsupervised clustering schemes, such as the k-means method or the vector quantization method, to see whether such methods can solve this problem (Duda & Hart, 1973). Through simulations using the wire-frame objects described in the next section, we found that these methods do not yield satisfactory performance. We, therefore, propose a new unsupervised learning scheme for classifying 3D objects. The proposed scheme exploits an auto-associative network for identifying an object. An auto-associative network finds an identity mapping through a bottleneck in the hidden layer, i.e., the l network approximates functions . F and F- 1 such that R n ~Rm r ) R n where m < n. The network, thus, compresses the input into a low dimensional representation by eliminating redundancy. If we use a five-layer perceptron network, the network can perform nonlinear dimensionality reduction, which is a nonlinear analogue to the principal component analysis (Oja, 1991; DeMers & Cottrell, 1993). The proposed classification scheme consists of a mixture of five-layer auto-associative networks which we call the identification networks, or the I-Nets. In the case where the inputs are the projected views of a rigid object, the minimum dimension that constrains the input variation is the degree of freedom of the rigid object, which is six in the most general case, three for rotation and three for translation. Thus, a single I-Net can compress the views of an object into a representation whose dimension is its degree of freedom. The proposed scheme categorizes each view of a number of 3D objects into its class through selecting an appropriate I-Net. We present the following two models for different selection and learning methods. Model I: The model I selects an I-Net whose output best fits the input (see Fig. 1). SpecIfically, we assume a classifier whose output vector is given by the softmax function of a negative squared difference between the input and the output of the I-Nets, i.e., (1) Unsupervised CLassification of 3D Objects from 2D Views I-Net I-Net ??? 951 I-Net I-Net I-Net... I-Net 2D Projected Images of 3D Objects 2D Projected Images of 3D Objects Modell Model II Figure 1: Model I and Model II. Each I-Net (identification net) is a 5-layer auto-associative network and the C-Net (classification net) is a 3-layer network. where Y * and Yi denote the input and the output of the i th I-Net, respectively. Therefore, if only one of the I-Nets has an output that best matches the input, then the output value of the corresponding unit in the classifier becomes nearly one and the output values of the other units become nearly zero. For training the network, we maximize the following objective function: Lexp[-ally * _YiI1 2 ] In-'~'--~------~ exp[ Yi 112 ] L -Ily *- (2) i where a (>1) denotes a constant. This function forces the output of at least one I-Net to fit the input, and it also forces the rest of I-Nets to increase the error between the input and the output. Since it is difficult for a single I-Net to learn more than one object, we expect that the network will eventually converge to the state where each I-Net identifies only one object. Model II: The model II, on the other hand, employs an additional network which we call the classification network or the C-Net, as illustrated in Fig. 1. The C-Net classifies the given views by directly partitioning the input space. This type of modular architecture has been proposed by Jacobs et al. (1991) based on a stochastic model (see also Jordan & Jacobs, 1992). In this architecture, the final output, Y, is given by (3) where Yi denotes the output of the i th I-Net, and gi is given by the softmax function gi = eXP[SiVtexP[Sj] where Si is the weighted sum arriving at the (4) i th output unit of the C-Net. For the C-Net, we use three-layer perceptron, since a simple perceptron with two layers did not provide a good performance for the objects used for our simulations (see Section Satoshi Suzuki, Hiroshi Ando 952 3). The results suggest that classification of such objects is not a linearly separable problem. Instead of using MLP (multi-layer perceptron), we could use other types of networks for the C-Net, such as RBF (radial basis function) (Poggio & Edelman, 1990). We maximize the objective function In LgjO'-1 exp[-lly*-yJ /(20'2)] (5) j where 0'2 is the variance. This function forces the C-Net to select only one I-Net, and at the same time, the selected I-Net to encode and decode the input information. Note that the model I can be interpreted as a modified version of the model II, since maximizing (2) is essentially equivalent to maximizing (5) if we replace Sj of the C-Net in (4) with a ne&ative s~uared difference between the input and the output of the i th I-Net, i.e., Sj = -Ily * -yj Although the model I is a more direct classification method that exploits auto-associative networks, it is interesting to examine what information can be extracted from the input for classification in the model II (see Section 3.2). Ir . 3 SIMULATIONS We implemented the network models described in the previous section to evaluate their performance. The 3D objects that we used for our simulations are 5-segment wire-frame objects whose six vertices are randomly selected in a unit cube, as shown in Fig. 2 (a) (see also Poggio & Edelman, 1990). Various views of the objects are obtained by orthographically projecting the objects onto an image plane whose position covers a sphere around the object (see Fig. 2 (b?. The view position is defined by the two parameters, 8 and fj). In the simulations, we used x, y image coordinates of the six vertices of three wire-frame objects for the inputs to the network. The models contain three I-Nets, whose number is set equal to the number of the objects. The number of units in the third layer of the five-layer I-Nets is set equal to the number of the view parameters, which is two in our simulations. We used twenty units in the second and fourth layers. To train the network efficiently, we initially limited the ranges of 8 and fj) to 1r /8 and 1r /4 and gradually increased the range until it covered the whole sphere. During the training, objects were randomly selected among the three and their views were randomly selected within the view range. The steepest ascent method was used for maximizing the objective functions (2) and (5) in our simulations, but more efficient methods, such as the conjugate gradient method, can also be used. (a) (b) z View y Figure 2: (a) 3D wire-frame objects. (b) Viewpoint defined by two parameters, 8 and fj). 953 Unsupervised Classification of 3D Objects from 2D Views 3.1 SIMULATIONS USING THE MODEL I This section describes the simulation results using the model!. As described in Section 2, the classifier of this model selects an I-Net that produces minimum error between the output and the input. We test the classification performance of the model and examine internal representations of the I-Nets after training the networks. The constant a in the objective function (2) was set to 50 during the training. Fig. 3 shows the output of the classifier plotted over the view directions when the views of an object are used for the inputs. The output value of a unit is almost equal to one over the entire range of the view direction, and the outputs of the other two units are nearly zero. This indicates that the network effectively classifies a given view into an object regardless of the view directions. We obtained satisfactory results for classification if more than five units are used in the second and fourth layers of the I-Nets. Fig. 4 shows examples of the input views of an object and the views recovered by the corresponding I-Net. The recovered views are significantly similar to the input views, indicating that each auto-associative I-Net can successfully compress and recover the views of an object. In fact, as shown in Fig. 5, the squared error between the input and the output of an I-Net is nearly zero for only one of the objects. This indicates that each I-Net can be used for identifying an object for almost entire view range. UNIT 1 UNIT 2 UNIT 3 Figure 3: Outputs of the classifier in the model I. The output value of the second unit is almost equal to one over the full view range, and the outputs of the other two units are nearly zero for one of the 3D objects. Recovered views Input views Figure 4: Examples of the input and recovered views of an object. The recovered views are significantly similar to the input views. 954 Satoshi Suzuki, Hiroshi Ando We further analyzed what information is encoded in the third layer of the I-Nets. Fig. 6 (a) illustrates the outputs of the third layer units plotted as a function of the view direction ( (}, ?) of an object. Fig. 6 (b) shows the view direction ( (} , ?) plotted as a function of the outputs of the third layer units. Both figures exhibit single-valued functions, i.e. the view direction of the object uniquely determines the outputs of the hidden units, and at the same time the outputs of the hidden units uniquely determine the view direction. Thus, each I-Net encodes a given view of an object into a representation that has one-to-one correspondence with the view direction. This result is expected from the condition that the dimension of the third layer is set equal to the degree of freedom of a rigid object. Object 1 Object 2 Object 3 Figure 5: Error between the input view and the recovered view of an I-Net for each object. The figures show that the I-Net recovers only the views of Object 3. (a) unit! unit2 (b) a Wlit2 unit2 Figure 6: (a) Outputs of the third layer units of an I-Net plotted over the view direction ( (}, ?) of an object. (b) The view direction plotted over the outputs of the third layer units. Figure (b) was obtained by inversely replotting Figure (a). 3.2 SIMULATIONS USING THE MODEL n In this section, we show the simulation results using the model II. The C-Net in the model learns to classify the views by splitting the input space nonlinearly. We examine internal representations of the C-Net that lead to view invariant classification in its output. 955 Unsupervised Classification of 3D Objects from 2D Views In the simulations, we used the same 3 wire-frame objects used in the previous simulations. The C-Net contains 20 units in the hidden layer. The parameter cr in the objective function (5) was set to 0.1. Fig. 7 (a) illustrates the values of an output unit in the C-Net for an object. As in the case of the model I, the model correctly classified the views into their original object for almost entire view range. Fig. 7 (b) illustrates the outputs of two of the hidden units as examples, showing that each hidden unit has a limited view range where its output is nearly one. The C-Net, thus, combines these partially invariant representations in the hidden layer to achieve full view invariance at the output layer. To examine a generalization ability of the model, we limited the view range in the training period and tested the network using the images with the full view range. Fig. 8 (a) and (b) show the values of an output unit of the C-Net and the error of the corresponding I-Net plotted over the entire view range. The region surrounded by a rectangle indicates the range of view directions where the training was done. The figures show that the correct classification and the small recovery error are not restricted within the training range but spread across this range, suggesting that the network exhibits a satisfactory capability of generalization. We obtained similar generalization results for the model I as well. We also trained the networks with a sparse set of views rather than using randomly selected views. The results show that classification is nearly perfect regardless of the viewpoints if we use at least 16 training views evenly spaced within the full view range. Figure 7: (a) Output values of an output unit of the C-Net when the views of an object are given (cf. Fig.3). (b) Output values of two hidden units ofthe C-Net for the same object. OUTPUT ERROR Figure 8: (a) Output values of an output unit of the C-Net. (b) Errors between the input views and the recovered views of the corresponding I-Net. The region surrounded by a rectangle indicates the view range where the training was done. 956 Satoshi Suzuki, Hiroshi Ando 4 CONCLUSIONS We have presented an unsupervised classification scheme that classifies 3D objects from their 2D views. The scheme consists of a mixture of non-linear auto-associative networks each of which identifies an object by encoding an input view into a representation that indicates its view direction. The simulations using 3D wire-frame objects demonstrated that the scheme effectively clusters the given views into their original objects with no explicit identification of the object classes being provided to the networks. We presented two models that utilize different classification mechanisms. In particular, the model I employs a novel classification and learning strategy that forces only one network to reconstruct the input view, whereas the model II is based on a conventional modular architecture which requires training of a separate classification network. Although we assumed in the simulations that feature points are already identified in each view and that their correspondence between the views is also established, the scheme does not, in principle, require the identification and correspondence of features, because the scheme is based solely on the existence of non-linear mappings between a set of images of an object and its degree of freedom. Therefore, we are currently investigating how the proposed scheme can be used to classify real gray-level images of 3D objects. Acknowledgments We would like to thank Mitsuo Kawato for extensive discussions and continuous encouragement, and Hiroaki Gomi and Yasuharu Koike for helpful comments. We are also grateful to Tommy Poggio for insightful discussions. References DeMers, D. and Cottrell, G. (1993). Non-linear dimensionality reduction. In Hanson, S. 1., Cowan, 1. D. & Giles, C. L., (eds), Advances in Neural Information Processing Systems 5. Morgan Kaufmann Publishers, San Mateo, CA. 580-587. Duda, R. O. and Hart, P. E. (1973). Pattern Classification and Scene Analysis. John Wiley & Sons, NY. Jacobs, R. A., Jordan, M. I., Nowlan, S. 1. and Hinton, G. E. (1991). Adaptive mixtures of local experts. Neural Computation, 3,79-87. Jordan, M. I. and Jacobs, R. A. (1992). Hierarchies of adaptive experts. In Moody, J. E., Hanson, S. J. & Lippmann, R. P., (eds), Advances in Neural Information Processing Systems 4. Morgan Kaufmann Publishers, San Mateo, CA. 985-992. Oja, E. (1991). Data compression, Feature extraction, and autoassociation in feedforward neural networks. In Kohonen, K. et al. (eds), Anificial Neural Networks. Elsevier Science publishers B.V., North-Holland. Poggio, T. and Edelman, S. (1990). A network that learns to recognize three-dimensional objects. Nature, 343, 263. Weinshall, D., Edelman, S. and Btilthoff, H. H. (1990). A self-organizing multiple-view representation of 3D objects. In Touretzky, D. S., (eds), Advances in Neural Information Processing Systems 2. Morgan Kaufmann Publishers, San Mateo, CA. 274-281. Williams, C. K. I., Zemel, R. S. and Mozer, M. C. (1993). Unsupervised learning of object models. AAAI Fall 1993 Symposium on Machine Learning in Computer Vision.
910 |@word version:1 eliminating:1 compression:1 duda:2 simulation:19 jacob:5 reduction:2 contains:1 selecting:2 existing:1 recovered:9 nowlan:1 si:1 john:1 cottrell:2 selected:5 plane:1 steepest:1 five:4 direct:1 become:1 symposium:1 edelman:5 consists:3 combine:1 tommy:1 acquired:1 expected:1 seika:1 examine:5 multi:1 becomes:1 provided:1 classifies:7 what:3 weinshall:2 interpreted:1 transformation:2 rm:1 classifier:5 partitioning:1 unit:29 local:1 encoding:2 solely:1 studied:1 examined:1 mateo:3 co:2 autoassociation:1 limited:3 range:16 acknowledgment:1 yj:2 significantly:3 radial:1 suggest:1 onto:1 selection:1 equivalent:1 conventional:1 demonstrated:1 maximizing:3 williams:2 regardless:2 splitting:3 identifying:2 recovery:1 variation:1 coordinate:1 hierarchy:1 decode:1 recognition:1 yasuharu:1 region:2 mozer:1 constrains:1 trained:1 grateful:1 segment:1 basis:1 various:2 train:1 hiroshi:5 zemel:1 tell:1 whose:9 modular:4 encoded:1 solve:1 valued:1 reconstruct:1 ability:2 gi:2 itself:1 final:1 associative:11 net:57 propose:3 kohonen:1 organizing:1 achieve:1 cluster:1 produce:1 perfect:1 object:76 implemented:1 direction:13 correct:1 stochastic:1 human:2 require:1 generalization:3 around:1 exp:3 mapping:2 vary:1 currently:1 successfully:1 weighted:1 modified:1 rather:1 cr:1 categorizing:2 encode:2 ily:2 indicates:6 helpful:1 elsevier:1 rigid:4 entire:4 initially:2 hidden:9 selects:3 classification:25 among:1 softmax:2 cube:1 equal:5 categorizes:1 extraction:1 btilthoff:1 unsupervised:13 nearly:7 employ:4 randomly:4 oja:2 recognize:3 ando:6 freedom:4 mitsuo:1 mlp:1 mixture:5 analyzed:1 poggio:5 plotted:6 hip:2 classify:3 increased:1 giles:1 cover:1 vertex:2 teacher:1 cho:1 moody:1 squared:2 aaai:1 expert:2 japan:1 suggesting:1 retinal:1 north:1 view:83 recover:1 capability:1 ative:1 ir:1 variance:1 kaufmann:3 efficiently:1 yield:1 spaced:1 ofthe:1 satoshi:6 koike:1 identification:4 classified:1 gomi:1 touretzky:1 ed:4 recovers:1 demers:2 dimensionality:2 organized:1 supervised:1 done:2 until:1 hand:2 ally:1 nonlinear:2 gray:1 contain:1 laboratory:1 satisfactory:3 illustrated:1 during:4 self:1 uniquely:2 demonstrate:2 fj:3 image:10 discovers:1 novel:1 rotation:1 kawato:1 jp:2 approximates:1 encouragement:1 modell:1 recent:1 belongs:1 yi:3 morgan:3 minimum:2 additional:3 converge:1 maximize:2 determine:1 period:1 signal:1 ii:8 full:4 multiple:1 kyoto:1 match:3 sphere:2 hart:2 essentially:1 vision:1 lly:1 whereas:1 publisher:4 rest:1 ascent:1 comment:1 cowan:1 effectiveness:1 jordan:3 call:2 feedforward:1 fit:2 architecture:5 identified:1 bottleneck:1 whether:1 six:3 orthographically:1 soraku:1 covered:1 category:1 correctly:1 redundancy:1 utilize:1 rectangle:2 sum:1 fourth:2 almost:4 layer:24 correspondence:3 scene:1 encodes:1 aspect:1 hiroaki:1 separable:1 conjugate:1 describes:5 across:1 son:1 projecting:1 gradually:1 invariant:2 restricted:1 discus:1 lexp:1 mechanism:3 eventually:1 appropriate:1 existence:1 original:2 compress:3 denotes:2 clustering:1 cf:1 exploit:3 hikaridai:1 classical:1 objective:5 already:1 strategy:1 exhibit:2 gradient:1 separate:1 thank:1 atr:3 gun:1 evenly:1 difficult:1 negative:1 twenty:1 perform:1 wire:9 segregate:1 hinton:1 frame:9 nonlinearly:2 extensive:1 hanson:2 established:1 pattern:1 analogue:1 force:4 scheme:16 inversely:1 ne:1 identifies:3 concludes:1 auto:11 expect:1 interesting:1 degree:4 affine:1 principle:1 viewpoint:3 classifying:2 surrounded:2 translation:1 arriving:1 perceptron:4 fall:1 sparse:1 dimension:3 suzuki:5 adaptive:2 projected:5 san:3 sj:3 lippmann:1 investigating:1 assumed:1 continuous:1 learn:3 nature:1 ca:3 anificial:1 did:1 spread:1 linearly:2 whole:1 fig:13 ny:1 wiley:1 position:2 explicit:1 third:7 learns:2 showing:1 insightful:1 explored:1 quantization:1 effectively:3 illustrates:3 visual:1 partially:1 holland:1 determines:1 extracted:1 identity:1 rbf:1 replace:1 change:1 specifically:1 principal:1 invariance:1 indicating:1 select:1 internal:3 evaluate:1 tested:1
7,148
911
A Study of Parallel Perturbative Gradient Descent D. Lippe? J. Alspector Bellcore Morristown, NJ 07960 Abstract We have continued our study of a parallel perturbative learning method [Alspector et al., 1993] and implications for its implementation in analog VLSI. Our new results indicate that, in most cases, a single parallel perturbation (per pattern presentation) of the function parameters (weights in a neural network) is theoretically the best course. This is not true, however, for certain problems and may not generally be true when faced with issues of implementation such as limited precision. In these cases, multiple parallel perturbations may be best as indicated in our previous results. 1 INTRODUCTION Motivated by difficulties in analog VLSI implementation of back-propagation [Rumelhart et al., 1986] and related algorithms that calculate gradients based on detailed knowledge of the neural network model, there were several similar recent papers proposing to use a parallel [Alspector et al., 1993, Cauwenberghs, 1993, Kirk et al., 1993] or a semi-parallel [Flower and Jabri, 1993] perturbative technique which has the property that it measures (with the physical neural network) rather than calculates the gradient. This technique is closely related to methods of stochastic approximation [Kushner and Clark, 1978] which have been investigated recently by workers in fields other than neural networks. [Spall, 1992] showed that averaging multiple parallel perturbations for each pattern presentation may be asymptotically preferable in the presence of noise. Our own results [Alspector et al., 1993] indicated ?Present address: Dept. of EECSj MITj Cambridge, MA 02139; dalippe@mit.edu 804 D. Lippe, 1. Alspector that multiple parallel perturbations are also preferable when only limited precision is available in the learning rate which is realistic for a physical implementation. In this work we have investigated whether multiple parallel perturbations for each pattern are non-asymptotically preferable theoretically (without noise). We have also studied this empirically, to the limited degree that simulations allow, by removing the precision constraints of our previous work. 2 GRADIENT ESTIMATION BY PARALLEL WEIGHT PERTURBATION Following our previous work, one can estimate the gradient of the error, E( w), with respect to any weight, Wi, by perturbing Wi by 6w 1 and measuring the change in the output error, 6E, as the entire weight vector, W, except for component Wi is held constant. 6E 6w1 E(w + 6;1) - E(w) 6Wi We now consider perturbing all weights simultaneously. However, we wish to have the perturbation vector, 6w, chosen uniformly on a hypercube. Note that this requires only a random sign multiplying a fixed perturbation and is natural for VLSI using a parallel noise generator [Alspector et al., 1991J. This leads to the approximation (ignoring higher order terms) w 6E - -+ 8E 2:(8E) -6w? -8w? (6Wi) -6w?? - 8w? , 'i?1 ] (1 ) , The last term has expectation value zero for random and independently distributed 6w1 ? The weight change rule where 1] is a learning rate, will follow the gradient on the average but with considerable noise. For each pattern, one can reduce the variance of the noise term in (1) by repeating the random parallel perturbation many times to improve the statistical estimate. If we average over P perturbations, we have where p indexes the perturbation number. A Study of Parallel Perturbative Gradient Descellf 3 3.1 805 THEORETICAL RELATIVE EFFICIENCY BACKGROUND Spall [Spall, 1992] shows in an asymptotic sense that multiple perturbations may be faster if only a noisy measurement of E( tV) is available, and that one perturbation is superior otherwise. His results are asymptotic in that they compare the rate of convergence to the local minimum if the algorithms run for infinite time. Thus, his results may only indicate that 1 perturbation is superior close to a local minimum. Furthermore, his result implicitly assumes that P perturbations per weight update takes P times as long as 1 perturbation per weight update. Experience shows that the time required to present patterns to the hardware is often the bottleneck in VLSI implementations of neural networks [Brown et al., 1992]. In a hardware implementation of a perturbative learning algorithm, a few perturbations might be performed with no time penalty while waiting for the next pattern presentation. The remainder of this section sketches an argument that multiple perturbations may be desirable for some problems in a non-asymptotic sense, even in a noise free environment and under the assumption of a multiplicative time penalty for performing multiple perturbations. On the other hand, the argument also shows that there is little reason to believe in practice that any given problem will be learned more quickly by multiple perturbations. Space limitations prevent us from reproducing the full argument and discussion of its relevance which can be found in [Lippe, 1994]. The argument fixes a point in weight space and considers the expectation value of the change in the error induced by one weight update under both the 1 perturbation case and the multiple perturbation case. [Cauwenberghs, 1994] contains a somewhat related analysis of the relative speed of one parallel perturbation and weight perturbation as described in [Jabri and Flower, 1991]. The analysis is only truly relevant far from a local minimum because close to a local minimum the variance of the change of the error is as important as the mean of the change of the error. 3.2 Calculations If P is the number of perturbations, then our learning rule is ~Wi -'TJ ~ 6E(p) =P If W is the number of weights, then ~E, 1 w 8E (2) L.J~. P=1 6wi calculated to second order in 'TJ, is W W 82E ~E = 'L.J " -8 W ? ~Wi + -2 '" L.J '" L.J 8 W? 8 W ? ~Wi~Wj. i=l' Expanding 6E(p) i=l j=l to second order in (j (where W 6E(p) 8E = ' " -6w~P) L.J 8w' j=l J J W 6Wi W ? = ?(j), we obtain 82 + !2 "'" E 6w~P)6w(P). L.J L.J 8w' k J k j=l k=l (3) J J 8w (4) 806 D. Lippe, 1. Alspector [Lippe, 1994] shows that combining (2)-(4), retaining only first and second order terms, and taking expectation values gives 2 < l:1E >= -TJX + ~ (Y + PZ) (5) where x w L i=l (8E)2 8w' ' ' Z Y Note that first term in (5) is strictly less than or equal to 0 since X is a sum of squares l . The second term, on the other hand, can be either positive or negative. Clearly then a sufficient condition for learning is that the first term dominates the second term. By making TJ small enough, we can guarantee that learning occurs. Strictly speaking, this is not a necessary condition for learning. However, it is important to keep in mind that we are only focusing on one point in weight space. If, at this point in weight space, < l:1E > is negative but the second term's magnitude is close to the first term's magnitude, it is not unlikely that at some other point in weight space < l:1E > will be positive. Thus, we will assume that for efficient learning to occur, it is necessary that TJ be small enough to make the first term dominate the second term. Assume that some problem can be successfully learned with one perturbation, at learning rate TJ(I). Then the first order term in (5) dominates the second order terms. Specifically, at any point in weight space we have, for some large constant J1., TJ(I)X ~ J1.TJ(I)2IY + ZI In order to learn with P perturbations, we apparently need TJ( P)X ~ J1. TJ(~)2 IY + P ZI (6) The assumption that the first order term of (5) dominates the second order terms implies that convergence time is proportional to Thus, learning is more efficient in the multiple perturbation case if ,.lp), J1.TJ(P) > J1.TJ(I) P (7) It turns out, as shown in [Lippe, 1994] that the conditions (6) and (7) can be met simultaneously with multiple perturbations if ~ 2. =f lIf we are at a stationary point then the first term in (5) is O. A Study of Parallel Perturbative Gradient Descent 807 It is shown in [Lippe, 1994], by using the fact that the Hessian of a quadratic function with a minimum is positive semi-definite, that if E is guadratic and has < 2). Any well a minimum, then Y and Z have the same sign (and hence behaved function acts quadratically sufficiently close to a stationary point. Thus, we can not get < flE > more than a factor of P larger by using P perturbations near local minima of well behaved functions. Although, as mentioned earlier, we are entirely ignoring the issue of the variance of flE, this may be some indication of the asymptotic superiority of 1 perturbation. =f 3.3 Discussion of Results -i The result that multiple perturbations are superior when ~ 2 may seem somewhat mysterious . It sheds some light on our answer to rewrite (5) as Y < flE >= -"IX + "I2(p + Z). For strict gradient descent, the corresponding equation is < flE >= flE = -"IX + "I2Z. The difference between strict gradient descent and perturbative gradient descent, on average, is the second order term "I2~. This is the term which results from not following the gradient exactly, and it Obviously goes down as P goes up and the gradient measurement becomes more accurate. Thus, if Z and Y have different signs, P can be used to make the second order term disappear. There is no way to know whether this situation will occur frequently. Furthermore, it is important to keep in mind that if Y is negative and Z is positive, then raising P may make the magnitude of the second order term smaller, but it makes the term itself larger. Thus, in general, there is little reason to believe that multiple perturbations will help with a randomly chosen problem. An example where multiple perturbations help is when we are at a point where the error surface is convex along the gradient direction, and concave in most other directions. Curvature due to second derivative terms in Y and Z help when the gradient direction is followed, but can hurt when we stray from the gradient. In this case, Z < 0 and possibly Y > 0, so multiple perturbations might be preferable in order to follow the gradient direction very closely. 4 4.1 SIMULATIONS OF SINGLE AND MULTIPLE PARALLEL PERTURBATION CONSTANT LEARNING RATES The second order terms in (5) can be reduced either by using a small learning rate, or by using more perturbations, as discussed briefly in [Cauwenberghs, 1993]. Thus, if "I is kept constant, we expect a minimum necessary number of perturbations in order to learn. This in itself might be of importance in a limited precision implementation. If there is a non-trivial lower bound on "I, then it might be necessary to use multiple perturbations in order to learn. This is the effect that was noticed in [Alspector et al., 1993]. At that time we thought that we had found empirically 808 P 1 1 1 1 1 7 7 7 7 7 D. Lippe, J. Alspector TJ .0005 .001 .002 .003 .004 .00625 .008 .0125 .025 .035 Table 1: Running times for the first initial weight vector Time for < . 1 Time for < .5 1,121,459 831 , 684 784, 768 4 94,029 1,695,974 707,840 583,654 922,880 1,010,355 Not tested 32,179 18,534 11,008 9,933 9,728 23,834 16,845 13,261 12,006 17,024 that multiple perturbations were necessary for learning. The problem was that we failed to decrease the learning rate with the number of perturbations. 4.2 EMPIRICAL RELATIVE EFFICIENCY OF SINGLE AND MULTIPLE PERTURBATION ALGORITHMS Section 3 showed that, in theory, multiple perturbations might be faster than 1 perturbation. We investigated whether or not this is the case for the 7 input hamming error correction problem as described in [Biggs, 1989]. This is basically a nearest neighbor problem. There exist 16 distinct 7 bit binary code words. When presented with an arbitrary 7 bit binary word, the network is to output the code word with the least hamming distance from the input. After preliminary tests with 50, 25, 7, and 1 perturbation, it seemed that 7 perturbations provided the fastest learning, so we concentrated on running simulations for both the 1 perturbation and the 7 perturbation case. Specifically, we chose two different (randomly generated) initial weight vectors, and five different seeds for the pseudo-random function used to generate the bWi. For each of these ten cases, we tested both 1 perturbation and 7 perturbations with various learning rates in order to obtain the fastest possible learning. The 128 possible input patterns were repeatedly presented in order. We investigated how many pattern presentations were necessary to drive the MSE below .1 and how many presentations were necessary to drive it below .5. Recalling the theory developed in section 3, we know that multiple perturbations can be helpful only far away from a stationary point. Thus, we expected that 7 perturbations might be quicker reaching .5 but would be slower reaching .1. The results are summarized in tables 1 and 2. Each table summarizes information for a different initial weight vector. All of the data presented are averaged over 5 runs, one with each of the different random seeds. The two columns labeled "Time for < .5" and "Time for < .1" are adjusted according to the assumption that one weight update at 7 perturbations takes 7 times as long as one weight update at 1 perturbation. In each table, the following four numbers appear in italics: the shortest time to reach .1 with 1 perturbation, the shortest time to reach .1 with 7 perturbations, the shortest time to reach .5 with 1 perturbation, and the shortest time to reach .5 with 7 perturbations. 7 perturbations were a loss in three out of four of the experiments. Surprisingly, A Study of Parallel Perturbative Gradient Descent l' 'T/ 1 1 1 1 1 1 1 1 1 .001 .002 .003 .004 .00625 .008 .0125 .025 .035 809 Table 2: Running times for the second initial weight vector TIme for < .1 TIme for < .5 928,236 719 , 078 154,139 1,603,354 629 , 530 611,610 912,333 1,580,442 Not tested 22,133 12,817 10,675 11,150 21,059 19,112 15,949 14,515 11,141 the one time that multiple perturbations helped was in reaching .1 from the second initial weight vector. There are several possible explanations for this. To begin with, these learning times are averages over only five simulations each, which makes their statistical significance somewhat dubious. Unfortunately, it was impractical to perform too many experiments as the data obtained required 180 computer simulations, each of which sometimes took more than a day to complete. Another possible explanation is that .1 may not be "asymptotic enough." The numbers .5 and .1 were chosen somewhat arbitrarily to represent non-asymptotic and asymptotic results. However, there is no way of predicting from the theory how close the error must be to its minimum before asymptotic results become relevant. The fact that 1 perturbation outperformed 7 perturbations in three out of four cases is not surprising. As explained in section 3, there is in general no reason to believe that multiple perturbations will help on a randomly chosen problem. 5 CONCLUSION Our results show that, under ideal computational conditions, where the learning rate can be adjusted to proper size, that a single parallel perturbation is, except for unusual problems, superior to multiple parallel perturbations. However, under the precision constraints imposed by analog VLSI implementation, where learning rates may not be adjustable and presenting a pattern takes longer than performing a perturbation, multiple parallel perturbations are likely to be the best choice. Acknowledgment We thank Gert Cauwenberghs and James Spall for valuable and insightful discussIons. References [Alspector et al., 1991] Alspector, J., Gannett, J. W ., Haber, S., Parker, M. B., and Chu, R. (1991). A VLSI-efficient technique for generating multiple uncorrelated noise sources and its application to stochastic neural networks. IEEE Transactions on Circuits and Systems, 38:109-123 . [Alspector et al., 1993] Alspector, J., Meir, R., Yuhas, B., Jayakumar, A., and Lippe, D. (1993). A parallel gradient descent method for learning in analog 810 D. Lippe, J. A/spector VLSI neural networks. In Hanson, S. J., Cowan, J. D., and Giles, C. L., editors, Advances in Neural Information Processing Systems 5, pages 836-844, San Mateo, California. Morgan Kaufmann Publishers. [Biggs, 1989] Biggs, N. L. (1989). Discrete Math. Oxford University Press. [Brown et al., 1992] Brown, T. X., Tran, M. D., Duong, T., and Thakoor, A. P. (1992). Cascaded VLSI neural network chips: Hardware learning for pattern recognition and classification. Simulation, 58(5):340-347. [Cauwenberghs, 1993] Cauwenberghs, G. (1993). A fast stochastic error-descent algorithm for supervised learning and optimization. In Hanson, S. J., Cowan, J. D., and Giles, C. L., editors, Advances in Neural Information Processing Systems 5, pages 244-251, San Mateo, California. Morgan Kaufmann Publishers. [Cauwenberghs, 1994] Cauwenberghs, G. (1994). Analog VLSI Autonomous Systems for Learning and Optimization. PhD thesis, California Institute of Technology. [Flower and Jabri, 1993J Flower, B. and Jabri, M. (1993). Summed weight neuron perturbation: An o(n) improvement over weight perturbation. In Hanson, S. J., Cowan, J. D., and Giles, C. L., editors, Advances in Neural Information Processing Systems 5, pages 212-219, San Mateo, California. Morgan Kaufmann Publishers. [Jabri and Flower, 1991] Jabri, M. and Flower, B. (1991). Weight perturbation: An optimal architecture and learning technique for analog VLSI feedforward and recurrent multilayer networks. In Neural Computation 3, pages 546-565. [Kirk et al., 1993] Kirk, D., Kerns, D., Fleischer, K., and Barr, A. (1993). Analog VLSI implementation of gradient descent. In Hanson, S. J., Cowan, J. D., and Giles, C. L., editors, Advances in Neural Information Processing Systems 5, pages 789-796, San Mateo, California. Morgan Kaufmann Publishers. [Kushner and Clark, 1978] Kushner, H. and Clark, D. (1978). Stochastic Approzimation Methods for Constrained and Unconstrained Systems. Springer-Verlag, New York. [Lippe, 1994] Lippe, D. A. (1994). Parallel, perturbative gradient descent methods for learning in analog VLSI neural networks. Master's thesis, Massachusetts Institute of Technology. [Rumelhart et al., 1986] Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986). Learning internal representations by error propogation. In Rumelhart, D. E. and McClelland, J. L., editors, Parallel Distributed Processing: Ezplorations in the Microstructure of Cognition, page 318. MIT Press, Cambridge, MA. [Spall, 1992] Spall, J. C. (1992). Multivariate stochastic approximation using a simultaneous perturbation gradient approximation. IEEE Transactions on A utomatic Control, 37(3):332-341.
911 |@word briefly:1 simulation:6 initial:5 contains:1 duong:1 surprising:1 perturbative:9 must:1 chu:1 realistic:1 j1:5 update:5 stationary:3 math:1 five:2 along:1 become:1 yuhas:1 theoretically:2 expected:1 alspector:13 frequently:1 little:2 becomes:1 provided:1 begin:1 circuit:1 developed:1 proposing:1 nj:1 impractical:1 guarantee:1 pseudo:1 act:1 morristown:1 concave:1 shed:1 preferable:4 exactly:1 control:1 fle:5 superiority:1 appear:1 positive:4 before:1 local:5 oxford:1 might:6 chose:1 studied:1 mateo:4 fastest:2 limited:4 averaged:1 acknowledgment:1 practice:1 definite:1 empirical:1 thought:1 word:3 kern:1 get:1 close:5 imposed:1 go:2 williams:1 independently:1 convex:1 rule:2 continued:1 dominate:1 his:3 gert:1 autonomous:1 hurt:1 rumelhart:4 recognition:1 labeled:1 quicker:1 calculate:1 wj:1 decrease:1 valuable:1 mentioned:1 environment:1 rewrite:1 efficiency:2 biggs:3 chip:1 various:1 distinct:1 fast:1 larger:2 spector:1 otherwise:1 noisy:1 itself:2 obviously:1 indication:1 took:1 tran:1 remainder:1 relevant:2 combining:1 convergence:2 generating:1 help:4 recurrent:1 nearest:1 indicate:2 implies:1 met:1 direction:4 closely:2 thakoor:1 stochastic:5 barr:1 fix:1 microstructure:1 preliminary:1 adjusted:2 strictly:2 correction:1 sufficiently:1 seed:2 cognition:1 estimation:1 outperformed:1 successfully:1 mit:2 clearly:1 rather:1 reaching:3 improvement:1 sense:2 helpful:1 i2z:1 entire:1 unlikely:1 vlsi:12 issue:2 classification:1 bellcore:1 retaining:1 constrained:1 lif:1 summed:1 field:1 equal:1 spall:6 few:1 randomly:3 simultaneously:2 recalling:1 truly:1 light:1 tj:12 held:1 implication:1 accurate:1 worker:1 necessary:7 experience:1 theoretical:1 column:1 earlier:1 giles:4 measuring:1 too:1 answer:1 propogation:1 quickly:1 iy:2 w1:2 thesis:2 possibly:1 jayakumar:1 derivative:1 summarized:1 performed:1 multiplicative:1 helped:1 apparently:1 cauwenberghs:8 parallel:23 square:1 variance:3 kaufmann:4 utomatic:1 basically:1 multiplying:1 drive:2 simultaneous:1 reach:4 mysterious:1 james:1 hamming:2 massachusetts:1 knowledge:1 back:1 focusing:1 higher:1 day:1 follow:2 supervised:1 furthermore:2 sketch:1 hand:2 propagation:1 indicated:2 behaved:2 believe:3 effect:1 brown:3 true:2 hence:1 i2:2 presenting:1 complete:1 recently:1 superior:4 dubious:1 physical:2 empirically:2 perturbing:2 lippe:12 bwi:1 analog:8 discussed:1 measurement:2 cambridge:2 unconstrained:1 had:1 longer:1 surface:1 curvature:1 multivariate:1 own:1 recent:1 showed:2 certain:1 verlag:1 binary:2 arbitrarily:1 morgan:4 minimum:9 somewhat:4 shortest:4 semi:2 multiple:26 desirable:1 full:1 faster:2 calculation:1 long:2 calculates:1 multilayer:1 expectation:3 sometimes:1 represent:1 background:1 source:1 publisher:4 strict:2 induced:1 cowan:4 seem:1 near:1 presence:1 ideal:1 feedforward:1 enough:3 zi:2 ezplorations:1 architecture:1 reduce:1 fleischer:1 bottleneck:1 whether:3 motivated:1 penalty:2 speaking:1 hessian:1 york:1 repeatedly:1 generally:1 detailed:1 repeating:1 ten:1 hardware:3 concentrated:1 mcclelland:1 reduced:1 generate:1 meir:1 exist:1 sign:3 per:3 discrete:1 waiting:1 four:3 prevent:1 kept:1 asymptotically:2 sum:1 run:2 master:1 summarizes:1 bit:2 entirely:1 bound:1 followed:1 quadratic:1 occur:2 constraint:2 speed:1 argument:4 performing:2 tv:1 according:1 smaller:1 wi:10 lp:1 making:1 explained:1 equation:1 turn:1 mind:2 know:2 unusual:1 available:2 away:1 slower:1 assumes:1 running:3 kushner:3 disappear:1 hypercube:1 noticed:1 occurs:1 italic:1 gradient:22 distance:1 thank:1 considers:1 trivial:1 reason:3 code:2 index:1 unfortunately:1 negative:3 implementation:9 proper:1 adjustable:1 perform:1 neuron:1 descent:10 situation:1 hinton:1 perturbation:71 reproducing:1 arbitrary:1 required:2 hanson:4 raising:1 california:5 learned:2 quadratically:1 address:1 flower:6 pattern:10 below:2 explanation:2 haber:1 difficulty:1 natural:1 predicting:1 cascaded:1 improve:1 technology:2 gannett:1 faced:1 relative:3 asymptotic:8 loss:1 expect:1 limitation:1 proportional:1 clark:3 generator:1 degree:1 sufficient:1 editor:5 uncorrelated:1 course:1 surprisingly:1 last:1 free:1 allow:1 institute:2 neighbor:1 taking:1 distributed:2 calculated:1 seemed:1 san:4 far:2 transaction:2 implicitly:1 keep:2 table:5 learn:3 expanding:1 ignoring:2 mse:1 investigated:4 jabri:6 significance:1 noise:7 parker:1 precision:5 stray:1 wish:1 kirk:3 ix:2 removing:1 down:1 insightful:1 pz:1 dominates:3 importance:1 phd:1 magnitude:3 likely:1 failed:1 tjx:1 springer:1 ma:2 presentation:5 considerable:1 change:5 infinite:1 except:2 uniformly:1 specifically:2 averaging:1 internal:1 relevance:1 dept:1 tested:3
7,149
912
Advantage Updating Applied to a Differential Game Mance E. Harmon Wright Laboratory WL/AAAT Bldg. 635 2185 Avionics Circle Wright-Patterson Air Force Base, OH 45433-7301 harmonme@aa.wpafb.mil Leemon C. Baird III? Wright Laboratory baird@cs.usafa.af.mil A. Harry Klopr Wright Laboratory klopfah@aa.wpafb.mil Category: Control, Navigation, and Planning Keywords: Reinforcement Learning, Advantage Updating, Dynamic Programming, Differential Games Abstract An application of reinforcement learning to a linear-quadratic, differential game is presented. The reinforcement learning system uses a recently developed algorithm, the residual gradient form of advantage updating. The game is a Markov Decision Process (MDP) with continuous time, states, and actions, linear dynamics, and a quadratic cost function. The game consists of two players, a missile and a plane; the missile pursues the plane and the plane evades the missile. The reinforcement learning algorithm for optimal control is modified for differential games in order to find the minimax point, rather than the maximum. Simulation results are compared to the optimal solution, demonstrating that the simulated reinforcement learning system converges to the optimal answer. The performance of both the residual gradient and non-residual gradient forms of advantage updating and Q-learning are compared. The results show that advantage updating converges faster than Q-learning in all simulations. The results also show advantage updating converges regardless of the time step duration; Q-learning is unable to converge as the time step duration ~rows small. U.S .A.F. Academy, 2354 Fairchild Dr. Suite 6K4l, USAFA, CO 80840-6234 Mance E. Hannon, Leemon C. Baird ll/, A. Harry Klopf 354 1 ADVANTAGE UPDATING The advantage updating algorithm (Baird, 1993) is a reinforcement learning algorithm in which two types of information are stored. For each state x, the value V(x) is stored, representing an estimate of the total discounted return expected when starting in state x and performing optimal actions. For each state x and action u, the advantage, A(x,u), is stored, representing an estimate of the degree to which the expected total discounted reinforcement is increased by performing action u rather than the action currently considered best. The optimal value function V* (x) represents the true value of each state. The optimal advantage function A *(x,u) will be zero if u is the optimal action (because u confers no advantage relative to itself) and A*(x,u) will be negative for any suboptimal u (because a suboptimal action has a negative advantage relative to the best action). The optimal advantage function A * can be defined in terms of the optimal value function v*: A*(x,u) = ~[RN(X,U)- V*(x)+ rNV*(x')] bat (1) The definition of an advantage includes a l/flt term to ensure that, for small time step duration flt, the advantages will not all go to zero. Both the value function and the advantage function are needed during learning, but after convergence to optimality, the policy can be extracted from the advantage function alone. The optimal policy for state x is any u that maximizes A *(x,u). The notation ~ax (x) = max A(x,u) (2) " defines Amax(x). If Amax converges to zero in every state, the advantage function is said to be normalized. Advantage updating has been shown to learn faster than Q-Iearning (Watkins, 1989), especially for continuous-time problems (Baird, 1993). If advantage updating (Baird, 1993) is used to control a deterministic system, there are two equations that are the equivalent of the Bellman equation in value iteration (Bertsekas, 1987). These are a pair of two simultaneous equations (Baird, 1993): A(x,u)-maxA(x,u') =(R+ y l1l V(x')- V(X?)_l (3) maxA(x,u)=O (4) w ~t " where a time step is of duration L1t, and performing action u in state x results in a reinforcement of R and a transition to state Xt+flt. The optimal advantage and value functions will satisfy these equations. For a given A and V function, the Bellman residual errors, E, as used in Williams and Baird (1993) and defined here as equations (5) and (6).are the degrees to which the two equations are not satisfied: E1 (xl,u) =(R(x"u)+ y l1l V(xt+l1I)- V(XI?)~- A(x"u)+ max A(x,,u' ) (5) E2 (x,u)=-maxA(x,u) (6) ~t " w Advantage Updating Applied to a Differential Game 2 355 RESIDUAL GRADIENT ALGORITHMS Dynamic programming algorithms can be guaranteed to converge to optimality when used with look-up tables, yet be completely unstable when combined with functionapproximation systems (Baird & Harmon, In preparation). It is possible to derive an algorithm that has guaranteed convergence for a quadratic function approximation system (Bradtke, 1993), but that algorithm is specific to quadratic systems. One solution to this problem is to derive a learning algorithm to perfonn gradient descent on the mean squared Bellman residuals given in (5) and (6). This is called the residual gradient form of an algorithm. There are two Bellman residuals, (5) and (6), so the residual gradient algorithm must perfonn gradient descent on the sum of the two squared Bellman residuals. It has been found to be useful to combine reinforcement learning algorithms with function approximation systems (Tesauro, 1990 & 1992). If function approximation systems are used for the advantage and value functions, and if the function approximation systems are parameterized by a set of adjustable weights, and if the system being controlled is deterministic, then, for incremental learning, a given weight W in the functionapproximation system could be changed according to equation (7) on each time step: dW = _ a a[E;(x"u,) + E;(x"u,)] 2 aw __ E( ) aE1 (x, ,u,) _ E ( ) aE2 (x" u, ) - a 1 x"u, aw a 2 x"u, aw = _a(_l (R + yMV(X'+M) - V (x) ) - A(x"u,) + max A(X"U)) dt u _(_I (yfJJ aV(x,+fJJ) _ av(x,))_ aA(x"u,) + am~XA(xt'U)J dt aw -amaxA(x"u) U aw aw (7) aw amaxA(x"u) U a W As a simple, gradient-descent algorithm, equation (7) is guaranteed to converge to the correct answer for a deterministic system, in the same sense that backpropagation (Rumelhart, Hinton, Williams, 1986) is guaranteed to converge. However, if the system is nondetenninistic, then it is necessary to independently generate two different possible "next states" Xt+L1t for a given action Ut perfonned in a given state Xt. One Xt+L1t must be used to evaluate V(Xt+L1t), and the other must be used to evaluate %w V(xt+fJJ)' This ensures that the weight change is an unbiased estimator of the true Bellman-residual gradient, but requires a system such as in Dyna (Sutton, 1990) to generate the second Xt+L1t. The differential game in this paper was detenninistic, so this was not needed here. 356 Mance E. Harmon, Leemon C. Baird /11, A. Harry KLopf 3 THE SIMULATION 3.1 GAME DEFINITION We employed a linear-quadratic, differential game (Isaacs, 1965) for comparing Q-learning to advantage updating, and for comparing the algorithms in their residual gradient forms. The game has two players, a missile and a plane, as in games described by Rajan, Prasad, and Rao (1980) and Millington (1991). The state x is a vector (xm,xp) composed of the state of the missile and the state of the plane, each of which are composed of the poSition and velocity of the player in two-dimensional space. The action u is a vector (um,up) composed of the action performed by the missile and the action performed by the plane, each of which are the acceleration of the player in two-dimensional space. The dynamics of the system are linear; the next state xt+ 1 is a linear function of the current state Xl and action Ul. The reinforcement function R is a quadratic function of the accelerations and the distance between the players. R(x,u)= [distance2 + (missile acceleration)2 - 2(plane acceleration)2]6t (8) R(X,U)=[(X m -X p)2 +U~-2U!]llt (9) In equation (9), squaring a vector is equivalent to taking the dot product of the vector with itself. The missile seeks to minimize the reinforcement, and the plane seeks to maximize reinforcement. The plane receives twice as much punishment for acceleration as does the missile, thus allowing the missile to accelerate twice as easily as the plane. The value function V is a quadratic function of the state. In equation (10), Dm and Dp are weight matrices that change during learning. (10) The advantage function A is a quadratic function of the state X and action u. The actions are accelerations of the missile and plane in two dimensions. A(x,u)=x~Amxm +x~BmCmum +u~Cmum + (11) x~Apxp +x~BpCpup +u~Cpup The matrices A, B, and C are the adjustable weights that change during learning. Equation (11) is the sum of two general quadratic functions. This would still be true if the second and fifth terms were xBu instead of xBCu. The latter form was used to simplify the calculation of the policy. Using the xBu form, the gradient is zero when u=-C-lBx!2. Using the xBCu form, the gradient of A(x,u) with respect to u is zero when u=-Bx!2, which avoids the need to invert a matrix while calculating the policy. 3.2 THE BELLMAN RESIDUAL AND UPDATE EQUATIONS Equations (5) and (6) define the Bellman residuals when maximizing the total discounted reinforcement for an optimal control problem; equations (12) and (13) modify the algorithm to solve differential games rather than optimal control problems. Advantage Updating Applied to a Differential Game 357 E1(x"u,) = (R(x"u,)+ r6tV(xl+6t)- V(X,?)..!...- A(x"u,)+ minimax A(x,) tl.t E 2 (x"u,)=-minimax A(x,) (12) (13) The resulting weight update equation is: tl.W = -aU .((rt:., R+ r 6tV (X'M')- V(x,?) 1t - A(x"u,)+minimax A(X,?) aV~6t) aV(X,?)_1 _ aA(x"u,) + aminimax A(X,?) aw aw aW tl.t (14) " A() aminimax A(x,) -amzmmax x, aw For Q-leaming, the residual-gradient form of the weight update equation is: tl.W =-a( R+ r 6t minimax Q(Xl+dt)-Q(x"u,?) (15) .( r 6t -kminimax Q(x,+6t)--kQ(x"u,?) 4 RESULTS 4.1 RESIDUAL GRADIENT ADVANTAGE UPDATING RESULTS The optimal weight matrices A *, B *, C *, and D* were calculated numerically with Mathematica for comparison. The residual gradient form of advantage updating learned the correct policy weights, B, to three significant digits after extensive training. Very interesting behavior was exhibited by the plane under certain initial conditions. The plane learned that in some cases it is better to turn toward the missile in the short term to increase the distance between the two in the long term. A tactic sometimes used by pilots. Figure 1 gives an example. 1 0 r - - - - - - -...... ....................... ~/ ............................~ ......... ....... .' .'...... i ...?? I \. \. GO V C ....til? 0.01 .001 .0001 0 0.04 0.08 0.12 Time Figure 1: Simulation of a missile (dotted line) pursuing a plane (solid line), each having learned optimal behavior. The graph of distance vs. time show the effects of the plane's maneuver in turning toward the missile. Mance E. Harmon. Leemon C. Baird III. A. Harry Klopf 358 4.2 COMPARATIVE RESULTS . The error in the policy of a learning system was defined to be the sum of the squared errors in the B matrix weights. The optimal policy weights in this problem are the same for both advantage updating and Q-learning, so this metric can be used to compare results for both algorithms. Four different learning algorithms were compared: advantage updating, Q-Iearning, Residual Gradient advantage updating, and Residual Gradient Qlearning. Advantage updating in the non-residual-gradient form was unstable to the point that no meaningful results could be obtained, so simulation results cannot be given for it. 4.2.1 Experiment Set 1 The learning rates for both forms of Q-Iearning were optimized to one significant digit for each simulation. A single learning rate was used for residual-gradient advantage updating in all four simulations. It is possible that advantage updating would have performed better with different learning rates. For each algorithm, the error was calculated after learning for 40,000 iterations. The process was repeated 10 times using different random number seeds and the results were averaged. This experiment was performed for four different time step durations, 0.05, 0.005, 0.0005, and 0.00005. The non-residualgradient form of Q-Iearning appeared to work better when the weights were initialized to small numbers. Therefore, the initial weights were chosen randomly between 0 and 1 for the residual-gradient forms of the algorithms, and between 0 and 10-8 for the non-residualgradient form of Q-learning. For small time steps, nonresidual-gradient Q-Iearning performed so poorly that the error was lower for a learning rate of zero (no learning) than it was for a learning rate of 10- 8 . Table 1 gives the learning rates used for each simulation, and figure 2 shows the resulting error after learning. ? 8 6 Final Error ---0 0-- [J -D--FQ 4 2 ? ? -?-RAU 0 0.05 0.005 0.0005 0.00005 TIme Step Duration Figure 2: Error vs. time step size comparison for Q-Learning (Q), residual-gradient Q-Learning(RQ), and residual-gradient advantage updating(RAU) using rates optimal to one significant figure for both forms of Q-Iearning, and not optimized for advantage updating. The final error is the sum of squared errors in the B matrix weights after 40,000 time steps of learning. The final error for advantage updating was lower than both forms of Q-learning in every case. The errors increased for Qlearning as the time step size decreased. 359 Advantage Updating Applied to a Differential Game Time step duration, III 5.10-2 5.10- 3 5.10-4 5.10-5 Q 0.02 0.06 0.2 0.4 RQ 0.08 0.09 0 0 RAU 0.005 0.005 0.005 0.005 Table 1: Learning rates used for each simulation. Learning rates are optimal to one significant figure for both forms of Q-learning, but are not necessarily optimal for advantage updating. 4.2.2 Experiment Set 2 Figure 3 shows a comparison of the three algorithms' ability to converge to the correct policy. The figure shows the total squared error in each algorithms' policy weights as a function of learning time. This simulation ran for a much longer period than the simulations in table 1 and figure 2. The learning rates used for this simulation were identical to the rates that were found to be optimal for the shorter run. The weights for the non-Residual gradient form of Q-Iearning grew without bound in all of the long experiments, even after the learning rate was reduced by an order of magnitude. Residual gradient advantage updating was able to learn the correct policy, while Q-learning was unable to learn a policy that was better than the initial, random weights. Leorning Ability Comporison 10~------------------, Error ---RAU .1 - - - - -, RO, ,01 ,001 0 2 3 4 5 Time Steps in millions Figure 3 5 Conclusion The experimental data shows residual-gradient advantage updating to be superior to the three other algorithms in all cases. As the time step grows small, Q-learning is unable to learn the correct policy. Future research will include the use of more general networks and implementation of the wire fitting algorithm proposed by Baird and Klopf (1994) to calculate the policy from a continuous choice of actions in more general networks. 360 Mance E. Hannon. Leemon C. Baird Ill. A. Harry Klopf Acknowledgments This research was supported under Task 2312Rl by the Life and Environmental Sciences Directorate of the United States Air Force Office of Scientific Research. References Baird, L.C. (1993). Advantage updating Wright-Patterson Air Force Base, OH. (Wright Laboratory Technical Report WL-TR-93-1146, available from the Defense Technical information Center, Cameron Station, Alexandria, VA 22304-6145). Baird, L.C., & Harmon, M. E. (In preparation). Residual gradient algorithms WrightPatterson Air Force Base, OH. (Wright Laboratory Technical report). Baird, L.C., & Klopf, A. H. (1993). Reinforcement learning with high-dimensional. continuous actions Wright-Patterson Air Force Base, OH. (Wright Laboratory technical report WL-TR-93-1147, available from the Defense Technical information Center, Cameron Station, Alexandria, VA 22304-6145). Bertsekas, D. P. (1987). Dynamic programming : Deterministic and stochastic models. Englewood Cliffs, NJ: Prentice-Hall. Bradtke, S. J. (1993). Reinforcement Learning Applied to Linear Quadratic Regulation. Proceedings of the 5th annual Conference on Neural Information Processing Systems. Isaacs, Rufus (1965). Differential games. New York: John Wiley and Sons, Inc. Millington, P. J. (1991). Associative reinforcement learning for optimal control. Unpublished master's thesis, Massachusetts Institute of Technology, Cambridge, MA. Rajan, N., Prasad, U. R., and Rao, N. J. (1980). Pursuit-evasion of two aircraft in a horizontal plane. Journal of Guidance and Control. 3(3). May-June, 261-267. Rumelhart, D., Hinton, G., & Williams, R. (1986). Learning representations by backpropagating errors. Nature. 323.. 9 October, 533-536. Sutton, R. S. (1990). Integrated architectures for learning, planning, and reacting based on approximating dynamic programming. Proceedings of the Seventh International Conference on Machine Learning. Tesauro, G. (1990). Neurogammon: A neural-network backgammon program. Proceedings of the International Joint Conference on Neural Networks . 3.. (pp. 33-40). San Diego, CA. Tesauro, G. (1992). Practical issues in temporal difference learning. Machine Learning, 8(3/4), 279-292. Watkins, C. J. C. H. (1989). Learningfrom delayed rewards. Doctoral thesis, University, Cambridge, England. Cambr~dge
912 |@word aircraft:1 seek:2 prasad:2 simulation:12 tr:2 solid:1 initial:3 united:1 current:1 comparing:2 yet:1 must:3 john:1 update:3 v:2 alone:1 plane:16 short:1 differential:11 consists:1 combine:1 fitting:1 expected:2 behavior:2 planning:2 bellman:8 discounted:3 notation:1 maximizes:1 maxa:3 developed:1 nj:1 suite:1 perfonn:2 temporal:1 every:2 iearning:7 um:1 ro:1 control:7 l1l:2 bertsekas:2 maneuver:1 modify:1 sutton:2 cliff:1 reacting:1 twice:2 au:1 doctoral:1 co:1 averaged:1 bat:1 acknowledgment:1 practical:1 backpropagation:1 digit:2 cannot:1 prentice:1 confers:1 deterministic:4 equivalent:2 center:2 pursues:1 maximizing:1 go:2 regardless:1 starting:1 duration:7 williams:3 independently:1 estimator:1 amax:2 oh:4 dw:1 diego:1 programming:4 us:1 velocity:1 rumelhart:2 updating:29 calculate:1 ensures:1 evades:1 ran:1 rq:2 reward:1 dynamic:6 patterson:3 ae1:1 completely:1 accelerate:1 easily:1 joint:1 leemon:5 solve:1 ability:2 itself:2 final:3 associative:1 advantage:42 product:1 poorly:1 academy:1 convergence:2 comparative:1 incremental:1 converges:4 derive:2 keywords:1 functionapproximation:2 c:1 correct:5 stochastic:1 considered:1 wright:9 hall:1 ae2:1 seed:1 currently:1 wl:3 modified:1 rather:3 mil:3 office:1 ax:1 june:1 backgammon:1 fq:1 am:1 sense:1 squaring:1 integrated:1 issue:1 ill:1 having:1 identical:1 represents:1 look:1 future:1 report:3 simplify:1 residualgradient:2 randomly:1 composed:3 delayed:1 englewood:1 navigation:1 detenninistic:1 necessary:1 shorter:1 harmon:5 initialized:1 circle:1 guidance:1 increased:2 rao:2 cost:1 kq:1 alexandria:2 seventh:1 stored:3 answer:2 aw:11 combined:1 punishment:1 international:2 squared:5 thesis:2 satisfied:1 dr:1 til:1 return:1 bx:1 harry:5 includes:1 baird:16 inc:1 satisfy:1 performed:5 minimize:1 air:5 rufus:1 simultaneous:1 llt:1 definition:2 mathematica:1 pp:1 isaac:2 e2:1 wpafb:2 dm:1 pilot:1 massachusetts:1 ut:1 dt:3 xa:1 receives:1 horizontal:1 hannon:2 defines:1 scientific:1 dge:1 grows:1 mdp:1 effect:1 normalized:1 true:3 unbiased:1 laboratory:6 ll:1 game:16 during:3 backpropagating:1 tactic:1 bradtke:2 recently:1 superior:1 rl:1 avionics:1 million:1 numerically:1 significant:4 rau:4 cambridge:2 dot:1 longer:1 base:4 tesauro:3 certain:1 life:1 employed:1 converge:5 maximize:1 period:1 technical:5 faster:2 england:1 af:1 calculation:1 long:2 e1:2 cameron:2 controlled:1 va:2 fjj:2 metric:1 iteration:2 sometimes:1 invert:1 decreased:1 exhibited:1 neurogammon:1 iii:3 l1t:5 architecture:1 suboptimal:2 defense:2 ul:1 distance2:1 york:1 action:18 useful:1 category:1 reduced:1 generate:2 dotted:1 rajan:2 four:3 demonstrating:1 evasion:1 graph:1 sum:4 run:1 bldg:1 parameterized:1 master:1 pursuing:1 decision:1 bound:1 guaranteed:4 quadratic:10 annual:1 optimality:2 performing:3 missile:14 tv:1 according:1 son:1 equation:16 turn:1 dyna:1 needed:2 available:2 mance:5 pursuit:1 rnv:1 ensure:1 include:1 learningfrom:1 calculating:1 especially:1 approximating:1 rt:1 said:1 gradient:28 dp:1 distance:3 unable:3 simulated:1 unstable:2 toward:2 regulation:1 october:1 negative:2 implementation:1 policy:13 adjustable:2 allowing:1 av:4 wire:1 markov:1 descent:3 hinton:2 grew:1 rn:1 station:2 pair:1 unpublished:1 extensive:1 optimized:2 learned:3 able:1 xm:1 appeared:1 program:1 max:3 perfonned:1 force:5 turning:1 residual:28 minimax:5 representing:2 technology:1 apxp:1 relative:2 interesting:1 degree:2 xp:1 row:1 changed:1 supported:1 institute:1 taking:1 fifth:1 fairchild:1 dimension:1 calculated:2 transition:1 avoids:1 reinforcement:16 san:1 qlearning:2 xi:1 continuous:4 table:4 learn:4 nature:1 ca:1 necessarily:1 directorate:1 repeated:1 tl:4 wiley:1 position:1 lbx:1 xl:4 watkins:2 xt:10 specific:1 flt:3 magnitude:1 aa:4 environmental:1 extracted:1 ma:1 acceleration:6 leaming:1 change:3 total:4 called:1 experimental:1 player:5 meaningful:1 klopf:6 latter:1 l1i:1 preparation:2 evaluate:2
7,150
913
Generalisation in Feedforward Networks Adam Kowalczyk and Herman Ferra Telecom Australia, Research Laboratories 770 Blackburn Road, Clayton, Vic. 3168, Australia (a.kowalczyk@trl.oz.au, h.ferra@trl.oz.au) Abstract We discuss a model of consistent learning with an additional restriction on the probability distribution of training samples, the target concept and hypothesis class. We show that the model provides a significant improvement on the upper bounds of sample complexity, i.e. the minimal number of random training samples allowing a selection of the hypothesis with a predefined accuracy and confidence. Further, we show that the model has the potential for providing a finite sample complexity even in the case of infinite VC-dimension as well as for a sample complexity below VC-dimension. This is achieved by linking sample complexity to an "average" number of implement able dichotomies of a training sample rather than the maximal size of a shattered sample, i.e. VC-dimension. 1 Introduction A number offundamental results in computational learning theory [1, 2, 11] links the generalisation error achievable by a set of hypotheses with its Vapnik-Chervonenkis dimension (VC-dimension, for short) which is a sort of capacity measure. They provide in particular some theoretical bounds on the sample complexity, i.e. a minimal number of training samples assuring the desired accuracy with the desired confidence. However there are a few obvious deficiencies in these results: (i) the sample complexity bounds are unrealistically high (c.f. Section 4.), and (ii) for some networks they do not hold at all since VC-dimension is infinite, e.g. some radial basis networks [7]. 216 Adam Kowalczyk, Hemzan Ferra One may expect that there are at least three main reasons for this state of affairs: (a) that the VC-dimension is too crude a measure of capacity, (b) since the bounds are universal they may be forced too high by some malicious distributions, (c) that particular estimates themselves are too crude, and so might be improved with time. In this paper we will attack the problem along the lines of (a) and (b) since this is most promising. Indeed, even a rough analysis of some proofs of lower bound (e.g. [I)) shows that some of these estimates were determined by clever constructions of discrete, malicious distributions on sets of "shattered samples" (of the size of VC-dimension). Thus this does not necessarily imply that such bounds on the sample complexity are really tight in more realistic cases, e.g. continuous distributions and "non-malicious" target concepts, the point eagerly made by critics of the formalism. The problem is to find such restrictions on target concepts and probability distributions which will produce a significant improvement. The current paper discusses such a proposition which significantly improves the upper bounds on sample complexity. 2 A Restricted Model of Consistent Learning First we introduce a few necessary concepts and some basic notation. We assume we are given a space of samples X with a probability measure J.L, a set H of binary functions X t-+ {O, I} called the hypothesis space and a target concept t E H. For an n-sample i = (Z1, ... , zn) E xn and h : H -+ {O, I} the vector (h(Z1), ... , h(zn)) E {O, l}n will be denoted by h(i). We define two projections 7r:e and 7rt,:e of H onto {0,1 } as follows 7r:e(h) = h z) = (h(Z1), ... , h(zn)) and 7rt,:e(h) = 7r:e(lt - hi) = It - hl(i) for every h E H . Below we shall use the notation 151 for the cardinality of the set 5 . The average density of the sets of projections 7r:e( H) or 7rt ,:e( H) in {O, l}n is defined as n de! (_ de! de! PrH(i) d~ 17r:e(H)I/2n = l7rt,:e(H)I/2n (equivalently, this is the probability of a random vector in {O, l}n belonging to the set 7r:e(H)). Now we define two associated quantities: d.:J PrH'Il(n) J PrH(i)J.Ln(di) de! = = 2- n J 17r:e(H) lJ.Ln (di), (1) = We recall that dH max{n ; 3:eEX .. I7r:e(H)1 2n} is called the VapnikChervonenkis dimension (VC-dimension) of H [1, 11]. If d H ~ 00 then Sauer's lemma implies the estimates (c.f. [1, 2, 10)) PrH'Il(n) ~ PrH,max(n) ~ 2- n cf?(dH, n) ~ 2- n (en/dH )dH where cf?(d,n) d;J 2::1=0 (~) (we assume , (2) C) d;J ?ifi > n). Now we are ready to formulate our main assumption in the model. We say that the space of hypotheses His (J.Ln , C)-uniform around t E 2x iffor every set 5 C {O, l}n Jl 7r t,:e(H) n 51J.Ln(dx) ~ CI 5 IPrH'Il(n). (3) Generalisation in Feedforward Networks 217 The meaning of this condition is obvious: we postulate that on average the number of different projections 7rt,ar(h) of hypothesis h E H falling into S has a bound proportional to the probability PrH,~(n) of random vector in {O, l}n belonging to the set 7rt,ar(H). Another heuristic interpretation of (3) is as follows. Imagine that elements of 7rt,ar(H) are almost uniformly distributed in {O, 1}n, i.e. with average density Par ~ CI7rt,ar(H)1/2n. Thus the "mass" of the volume lSI is l7rt,ar(H) n SI ~ PariSI and so its average f l7rt,ar(H) n SIJ.tn(di) has the estimate ~ lSI f parJ.tn(di) ~ CISiPrH,~(n). Of special interest is the particular case of con8istent learning [1], i.e. when the target concept and the hypothesis fully agree on the training sample. In this case, for any E > 0 we introduce the notation QE(m) d;J de! m {i E Xm ; 3hEH ert,ar(h) = 0 & ert,~(h) ~ E}, de! f where ert,ar(h) = Ei=llt-hl(zi)/mand ert,~(h) = It-hl(z)J.t(dz) denote error rates on the training sample i = (ZlJ ... , zm) and X, respectively. Thus QE(m) is the set of all m-samples for which there exists a hypothesis in H with no error on the sample and the error at least E on X. Theorem 1 If the hypothesis space H is (J.t 2m , C)-uniform around t E H then for anyE > 8/m (4) (5) Proof of the theorem is given in the Appendix. Given E,6 > O. The integer mL(6, E) d;J minim > 0 j J.tm(QE) ~ 6} will be called the 8ample complezity following the terminology of computational learning theory (c.f. [1]). Note that in our case the sample complexity depends also (implicitly) on the target concept t, the hypothesis space H and the probability measure J.t. Corollary 2 If the hypothesis space H is (J.t n , C)-uniform around t E H for any n> 0, then mL(6, E) < maz{8/E, minim j 2CPrH,~(2m)(3/2r < 6}} (6) C < maz{8/E, 6.9 dH + 2.4log2 6}' 0 (7) The estimate (7) of Corollary 2 reduces to the estimate mL(6, E) ~ 6.9 dH independent of 6 and E. This predicts that under the assumption of the corollary a transition to perfect generalisation occurs for training samples of size ~ 6.9 d H , which is in agreement with some statistical physics predictions showing such transition occurring below ~ 1.5dH for some simple neural networks (c.f. [5]). Proof outline. Estimate (6) follows from the first estimate (5). Estimate (7) can be derived from the second bound in (5) (virtually by repeating the proof of [1, Theorem 8.4.1] with substitution of 4log2(4/3) for E and 6/C for 6). Q.E.D. 2I8 Adam Kowalczyk, Herman Ferra 10-1 fie 10-2 10-300 '---'-..J....J............L.U.L_.L-J'--'--<u..LJw......L...L!-O.!.-'-'-u.uJ 10 102 m 103 104 :; . 10-3 10 m Figure 1: Plots of estimates on PrH,J.I(m) and f(m, ?)/C ~ f(m,O)/C = (3/2) 3m PrH,J.I (2m) (Fig_ a) for the analytic threshold neuron (PrH'J.I(m) = <P(dH' 2m)(3/4)3~ and (Fig_ b) for the abstract perceptron according to the estimate (11) on PrH,J.I(m) for m1 = 50, m3 = 1000 and p = 0.015_ The upper bound on the sample complexity, mL(?, 6), corresponds to the abscissa value of the intersection of the curve f(m,O) with the level 6/C (c.f. point A in Fig. b). In this manner for 6/C 0.01 we obtain estimates mL ~ 602 and mL < 1795 in the case of Fig. a for d H = 100 and d H = 300 and mL ~ 697 in the case of Fig. b (d H = ~ = 1000), respectively. = 3 An application to feedforward networks In this section we shall discuss the problem of estimation of PrH'J.I(m) which is crucial for application of the above formalism. First we discuss an example of an analytic threshold neuron on R n [6] when H is the family of all functions R n -+ {O, 1}, x 1-+ 9(~:=1 Wiai(X)), where ai : R n -+ R are fixed real analytic functions and 9 the ordinary hard threshold. In this case d H equals the number of linearly independent functions among ab a3, ... , all. For any continuous probability distribution J.L on R n we have: PrH(i) = <p(dH , m)/2m (Vm and 'Vi E (Rnr with probability 1), (8) and consequently PrH'J.I(m) = PrH,max(m) = <P(dH' m)/2m for every m. (9) Note that this class of neural networks includes as particular cases, the linear threshold neuron (if Ie = n+ 1 and a1, ... , a n +1 are chosen as 1, Xl, ... , xn) and higher order networks (if ai(x) are polynomials); in the former case (8) follows also from the classical result of T. Cover [3]. Now we discuss the more complex case of a linear threshold multilayer perceptron on X = R n , with H defined as the family of all functions R n -+ {0,1} that such an architecture may implement and J.L is any continuous probability measure Generalisation in Feedforward Networks 219 on Rn. In this case there exist two functions, Blow (m) and Bup(m), such that Blow(m) ~ PrH(x) ~ Bup(m)foranyxE (Rn)mwithprobabilityl. In other words, PrH(x) takes values within a "bifurcation region" similar to the shaded region in Fig. l.b. Further, it is known that Blow(m) = 1 for m ~ nh l + 1, Bup(m) = 1 for m ~ dH and, in general, Blow (m) ~ cI.i(nh l + 1, m) and Bup(m) ~ cI.i(dH , m), where hl is the number of neurons in the first hidden layer. Given this we can say that PrH'I'(m) takes values somewhere within the "bifurcation region". It is worth noting that the width of the "bifurcation region" , which approximately equals 2( dH - nh l ) (since it is known that values on the boundaries have a positive probability of being randomly attained) increases with increasing hl since [9] O(nhllog:z(h l )) = maxp(n - p)(hd2 - 2P ) ~ d H ? (10) p Estimate (6) is better than (7) in general, and in particular, if PrH,I'(m) "drops" to 0 much quicker than Bup = cI.i(dH , m), we may expect that it will provide an estimate of sample complexity mL even below dHi if PrH'I'(m) is close to Bup , then the difference between both estimates will be negligible. In order to clarify this issue we shall consider now a third, abstract example. We introduce the abstract perceptron defined as the set of hypotheses H on a probabilistic space (X, J.?) with the following property for a random m + 1-tuple (x,z) E xm x X: dvc(x, z) = dvc(x) + 1 with probability = 1 if dvc(x) < ml and with probability pifml ~ dvc(x) < m:z, and dvc(x,z) = dvc(X), otherwise. Here 0 ~ ml ~ m2 ~ 00 are two (integer) constants, 0 ~ p ~ 1 is another constant and dvc(z1J ... , zm) is the maximal n such that 17r(:Z:'l, ...,:z:,,,)(H)1 = 2n for some 1 ~ i l < ... < in ~ m. It can easily be seen that d H = m:z in this case and that the threshold analytic neuron is a particular example of abstract percept ron (with de! = 0 and ml = m:z = dH, Blow(m) = Bup = cI.i(mlJ m)J2m ) . Note further, that if 0 < p < 1, then for any m ~ ml, dvc(x) = ml with probability> 0, and for any m ~ m:z, dvc(x) = dH = m:z with probability> o. In this regard the abstract percept ron resembles the linear threshold multilayer percept ron (with ml and m:z corresponding to nhl + 1 and d H , respectively). However, the main advantage of this model is that we can derive the following estimate: p PrH,.( m) <:; 2- m ~~' (m ~ m, ) pm-m, -'( 1 - p)'<J> (min( m - i, m,), m) (11) Using this estimate we find that for sufficiently low p (and sufficiently large m:z) the sample complexity upper bound (6) is determined by ml and can even be lower than m:z = dH (c.f. Figure l.b). In particular, the sample complexity determined by Eqns. (6) and (11) can be finite even if dH = m:z = 00 (c.f. the curve E(m, 0) for p = .05 in Fig. l.b which is the same for m:z = 1000 and m:z = (0). 4 Discussion The paper strongly depends on the postulate (3) of (J.?n, C)-uniformity. We admit that this is an ad hoc assumption here as we do not give examples when it is 220 Adam Kowalczyk, Herman Ferra satisfied nor a method to determine the constant C. From this point of view our results at the current stage have no predictive power, perchaps only explanatory one. The paper should be viewed as an attempt in the direction to explain within VC-formalism some known generalisation properties of neural networks which are out of the reach of the formalism to date, such as the empirically observed peak generalisation for backpropagation network trained with samples of the size well below VC-dimension [8] or the phase transitions to perfect generalisation below 1.5. xVC-dimension [5]. We see the formalism in this paper as one of a number of possible approaches in this direction. There are other possibilities here as well (e.g. [5, 12]) and in particular other, weaker versions of (p.fI., C)-uniformity can be used leading to similar results. For instance in Theorem 1 and Corollary 2 it was c.f. the enough to assume (p.fI. , C)-uniformity for a special class of sets S (S = Appendix); we intend to discuss other options in this regard on another occasion. s;;:;m, Now we relate this research to some previous results (e.g. [2, 4]) which imply the following estimates on sample complexity (c.f. [1, Theorems 8.6.1-2]): - 1 ,-In(<<5)/?)~mL(<<5,?)~ .. max ( dH32? r-; 4 ( 12 2)1 ' dH log:l-;-+log:l6" (12) where the lower bound is proved for all ? ~ 1/8 and ?5 ~ 1/100; here mL(<<5, ?) is the "universal" sample complexity, i.e. for all target concepts t and all probability distributions p.. For ? = ?5 = 0.01 and dH ? 1 this estimate yields 3dH < mi.(.Ol, .01) < 4000dH' These bounds should be compared against estimates of Corollary 2 of which (7) provides a much tighter upper bound, mL(.Ol, .01) ~ 6.9dH , if the assumption on (p.m, C)-uniformity of the hypothesis space around the target concept t is satisfied. 5 Conclusions We have shown that und~r appropriate restriction on the probability distribution and target concept, the upper bound on sample complexity (and "perfect generalisation") can be lowered to ~ 6.9x VC-dimension, and in some cases even below VC-dimension (with a strong possibility that multilayer perceptron could be such). We showed that there are other parameters than VC-dimension potentially impacting on generalisation capabilities of neural networks. In particular we showed by example (abstract perceptron) that a system may have finite sample complexity and infinite VC dimension at the same time. The formalism of this paper predicts transition to perfect generalisation at relatively low training sample sizes but it is too crude to predict scaling laws for learning curves (c.f. [5, 12] and references in there). Acknowledgement. The permission of Managing Director, Research and Information Technology, Telecom Australia, to publish this paper is gratefully acknowledged. References [1] M. Anthony and N. Biggs. Computational Learning Theory. Cambridge Uni- 221 Generalisation in Feedfom'ard Networks versity Press, 1992. [2] A. Blumer, A. Ehrenfeucht, D. Haussler, and M.K. Warmuth. Learnability and the Vapnik-Chervonenkis dimensions. Journal of the ACM, 36:929-965, (Oct. 1989). [3] T.M. Cover. Geometrical and statistical properties of linear inequalities with applications to pattern recognition. IEEE Trans. Elec. Comp., EC-14:326334, 1965. [4] A. Ehrenfeucht, D. Haussler, M. Kearns, and L. Valiant. A general lower bound on the number of examples needed for learning. Information and Computation, 82:247-261, 1989. [5] D. Hausler, M. Kearns, H.S. Seung, and N. Tishby. Rigorous learning curve bounds from statistical mechanics. Technical report, 1994. [6] A. Kowalczyk. Separating capacity of analytic neuron. In Proc. ICNN'94 , Orlando, 1994. [7] A. Macintyre and E. Sontag. Finiteness results for sigmoidal "neural" networks. In Proc. of the 25th Annual ACM Symp. Theory of Comp., pages 325-334, 1993. [8] G.L. Martin and J .A. Pitman. Recognizing handprinted letters and digits using backpropagation learning. Neural Comput., 3:258-267, 1991. [9] A. Sakurai. Tighter bounds of the VC-dimension of three-layer networks. In Proceedings of the 1999 World Congress on Neural Networks, 1993. [10] N. Sauer. On the density of family of sets. Journal of Combinatorial Theory (Series A), 13:145-147, 1972). [11] V. Vapnik. Estimation of Dependences Based on Empirical Data. SpringerVerlag, 1982. [12] V. Vapnik, E. Levin, and Y. Le Cun. Measuring the vc-dimension of a learning machine. Neural Computation, 6 (5):851-876, 1994). 6 Appendix: Sketch of the proof of Theorem 1 The proof is a modification of the proof of [1, Theorem 8.3.1]. We divide it into three stages. Stage 1. Let . de! 'RJ = {(x,y) E Xm X Xm ~ X 2 m ; 3hEHer:ch = 0 & eryh = jim} (13) for j E {O, 1, ... , m}. Using a Chernoff bound on the "tail" of binomial distribution it can be shown [1, Lemma 8.3.2] that for m ~ 8/? m Jr(Qf(m)) ::; 2 L j.L2m(nj) (14) j~rmf/21 Stage 2. Now we use a combinatorial argument to estimate j.L2m(1~j). We consider the 2m -element commutative group G m of transformations of xm x xm ~ X 2m generated by all "co-ordinate swaps" of the form (Xb ... , Xm, Y1I ... , Ym) 1-+ (Xl, ... , Xi-1, Yi, Xi+1I ... , Xm, Yb ... , Yi-1, Xi, Yi+b ... , Ym), 222 Adam Kowalczyk, Herman Ferra for 1 ~ i ~ m. We assume also that Gm transforms {O, l}m x {O, l}m ~ {O, 1}2m in a similar fashion. Note that (15) As transformation u E Gm preserve the measure p.2m on xm 2mp.2m(ni) = IGmlp.2m(ni) = I: X xm we obtain J p.2m(d?dY)X'RAu(?, Y? uEG ... J p.2m(didY) I: X'R.i(u(i, Y?. (16) uEG ... Let s;;:;m = de! {h- = (hI, h2) E {O,l}m x {O, l}m = ?& - - hI IIh211 = j} and de! = {h- E {O,l} 2m ; Ilhll = j}, where Ilhll = hI + ... + hm for any h = (h1, ... , hm) E {O, l}m. Then 181ml = (2j), u(s;;:;m) C 81m for any u E Gm and 8j2m de! ni = {(i, YJ E xm x xm ; 3h E 7rt,(z,y)(H) n s;;:;m}, (17) Thus from Eqn. (16) we obtain 2mp.2m(ni) ~ J J J J p.2m(didY) I: xs;:-;-(uh) uEG ... hEWt,(.,f)(H)nS;- I: p.2m(didY) I: xs;:-;",(uh) hEwt,{.,f)(H)ns;'" uEG ... p.2m(didY) _ = I: I: _... I{u E Gm j uh E 81m}1 hEw t,(.,f)(H)nS o,; p.2m (didY) I{7rt,(:ii,y) (H) n 8;m}1 2m- j . Applying now the condition of (p.2m, C)-uniformity (Eqn. 3), Eqn. 17 and dividing by 2m we get Stage 3. On substitution of the above estimate into (14) we obtain estimate (4). To derive (5) let us observe that 2:7=rmE / 21 (2j) 2- j ~ (1 + 1/2)2m. Q.E.D.
913 |@word eex:1 version:1 maz:2 polynomial:1 achievable:1 substitution:2 series:1 chervonenkis:2 current:2 si:1 dx:1 realistic:1 analytic:5 plot:1 drop:1 warmuth:1 affair:1 short:1 bup:7 provides:2 ron:3 attack:1 sigmoidal:1 along:1 director:1 symp:1 manner:1 introduce:3 indeed:1 themselves:1 abscissa:1 nor:1 mechanic:1 ol:2 versity:1 cardinality:1 increasing:1 notation:3 mass:1 transformation:2 nj:1 every:3 positive:1 negligible:1 congress:1 approximately:1 might:1 au:2 resembles:1 shaded:1 co:1 yj:1 implement:2 backpropagation:2 digit:1 empirical:1 universal:2 significantly:1 projection:3 confidence:2 road:1 radial:1 word:1 get:1 onto:1 clever:1 selection:1 close:1 applying:1 restriction:3 dz:1 formulate:1 m2:1 haussler:2 his:1 ert:4 imagine:1 target:9 construction:1 assuring:1 gm:4 hypothesis:13 agreement:1 element:2 recognition:1 predicts:2 observed:1 quicker:1 region:4 und:1 complexity:17 seung:1 trained:1 uniformity:5 tight:1 predictive:1 basis:1 biggs:1 swap:1 uh:3 easily:1 elec:1 forced:1 dichotomy:1 heuristic:1 say:2 otherwise:1 maxp:1 prh:20 hoc:1 advantage:1 parisi:1 maximal:2 zm:2 date:1 oz:2 produce:1 adam:5 perfect:4 derive:2 ard:1 strong:1 dividing:1 implies:1 direction:2 vc:16 dvc:9 australia:3 orlando:1 really:1 icnn:1 proposition:1 tighter:2 clarify:1 hold:1 around:4 sufficiently:2 predict:1 estimation:2 proc:2 combinatorial:2 rough:1 rather:1 corollary:5 derived:1 zlj:1 improvement:2 vapnikchervonenkis:1 rigorous:1 rmf:1 shattered:2 lj:1 explanatory:1 hidden:1 issue:1 among:1 denoted:1 impacting:1 special:2 bifurcation:3 equal:2 chernoff:1 blackburn:1 report:1 few:2 randomly:1 preserve:1 phase:1 ab:1 attempt:1 interest:1 possibility:2 xb:1 predefined:1 tuple:1 necessary:1 sauer:2 divide:1 desired:2 theoretical:1 minimal:2 ferra:6 formalism:6 instance:1 ar:8 cover:2 sakurai:1 measuring:1 zn:3 ordinary:1 uniform:3 recognizing:1 levin:1 too:4 learnability:1 tishby:1 ilhll:2 density:3 peak:1 ie:1 probabilistic:1 physic:1 vm:1 j2m:2 ym:2 dhi:1 postulate:2 satisfied:2 admit:1 leading:1 potential:1 de:11 blow:5 includes:1 mp:2 depends:2 vi:1 ad:1 view:1 h1:1 sort:1 option:1 capability:1 il:3 ni:4 accuracy:2 percept:3 yield:1 complezity:1 worth:1 comp:2 explain:1 llt:1 reach:1 against:1 obvious:2 proof:7 associated:1 di:4 mi:1 proved:1 recall:1 improves:1 mlj:1 higher:1 attained:1 improved:1 yb:1 strongly:1 stage:5 sketch:1 eqn:3 ei:1 concept:10 former:1 laboratory:1 fig_:2 ehrenfeucht:2 width:1 eqns:1 qe:3 occasion:1 outline:1 tn:2 geometrical:1 meaning:1 fi:2 empirically:1 volume:1 jl:1 linking:1 interpretation:1 m1:1 tail:1 nh:3 significant:2 cambridge:1 rau:1 ai:2 pm:1 gratefully:1 lowered:1 showed:2 inequality:1 binary:1 iih211:1 yi:3 seen:1 additional:1 managing:1 determine:1 ii:2 rj:1 reduces:1 technical:1 fie:1 heh:1 a1:1 prediction:1 basic:1 multilayer:3 publish:1 achieved:1 unrealistically:1 malicious:3 finiteness:1 crucial:1 virtually:1 ample:1 integer:2 noting:1 feedforward:4 enough:1 zi:1 architecture:1 ifi:1 tm:1 iffor:1 rnr:1 sontag:1 transforms:1 repeating:1 macintyre:1 exist:1 lsi:2 discrete:1 l2m:2 shall:3 group:1 terminology:1 threshold:7 nhl:1 falling:1 acknowledged:1 i7r:1 letter:1 almost:1 family:3 z1j:1 appendix:3 scaling:1 dy:1 bound:19 hi:4 layer:2 annual:1 deficiency:1 hew:1 y1i:1 rme:1 argument:1 min:1 relatively:1 martin:1 according:1 belonging:2 jr:1 cun:1 modification:1 hl:5 restricted:1 sij:1 ln:4 agree:1 discus:6 needed:1 hd2:1 observe:1 kowalczyk:7 appropriate:1 permission:1 binomial:1 cf:2 log2:2 l6:1 somewhere:1 uj:1 classical:1 intend:1 quantity:1 occurs:1 rt:8 dependence:1 link:1 separating:1 capacity:3 reason:1 providing:1 handprinted:1 equivalently:1 trl:2 potentially:1 relate:1 minim:2 allowing:1 upper:6 neuron:6 finite:3 jim:1 rn:2 ordinate:1 clayton:1 z1:3 trans:1 able:1 below:7 pattern:1 xm:12 herman:4 max:4 power:1 vic:1 technology:1 imply:2 ready:1 hm:2 acknowledgement:1 law:1 fully:1 expect:2 par:1 proportional:1 h2:1 consistent:2 i8:1 critic:1 qf:1 l_:1 weaker:1 perceptron:5 pitman:1 distributed:1 regard:2 curve:4 dimension:19 l7rt:3 xn:2 transition:4 boundary:1 world:1 wiai:1 made:1 ec:1 uni:1 implicitly:1 ml:19 xi:3 continuous:3 promising:1 necessarily:1 complex:1 anthony:1 main:3 linearly:1 fig:5 telecom:2 en:1 fashion:1 n:3 xl:2 comput:1 crude:3 third:1 theorem:7 showing:1 x:2 a3:1 exists:1 vapnik:4 valiant:1 ci:5 occurring:1 commutative:1 intersection:1 lt:1 ch:1 corresponds:1 dh:23 acm:2 oct:1 viewed:1 blumer:1 consequently:1 hard:1 springerverlag:1 generalisation:12 infinite:3 determined:3 uniformly:1 lemma:2 kearns:2 called:3 m3:1
7,151
914
A Rigorous Analysis Of Linsker-type Hebbian Learning J. Feng Mathematical Department University of Rome "La Sapienza? P. Ie A. Moro, 00185 Rome, Italy H. Pan V. P. Roychowdhury School of Electrical Engineering Purdue University West Lafayette, IN 47907 feng~at.uniroma1.it hpan~ecn.purdue.edu vwani~drum.ecn.purdue.edu Abstract We propose a novel rigorous approach for the analysis of Linsker's unsupervised Hebbian learning network. The behavior of this model is determined by the underlying nonlinear dynamics which are parameterized by a set of parameters originating from the Hebbian rule and the arbor density of the synapses. These parameters determine the presence or absence of a specific receptive field (also referred to as a 'connection pattern') as a saturated fixed point attractor of the model. In this paper, we perform a qualitative analysis of the underlying nonlinear dynamics over the parameter space, determine the effects of the system parameters on the emergence of various receptive fields, and predict precisely within which parameter regime the network will have the potential to develop a specially designated connection pattern. In particular, this approach exposes, for the first time, the crucial role played by the synaptic density functions, and provides a complete precise picture of the parameter space that defines the relationships among the different receptive fields. Our theoretical predictions are confirmed by numerical simulations. lian/eng Feng, H. Pan, V. P. Roychowdhury 320 1 Introduction For the purpose of understanding the self-organization mechanism of primary visual system, Linsker has proposed a multilayered unsupervised Hebbian learning network with random un correlated inputs and localized arborization of synapses between adjacent layers (Linsker, 1986 & 1988). His simulations have shown that for appropriate parameter regimes, several structured connection patterns (e.g., centre-surround and oriented afferent receptive fields (aRFs)) occur progressively as the Hebbian evolution of the weights is carried out layer by layer. The behavior of Linsker's model is determined by the underlying nonlinear dynamics which are parameterized by a set of parameters originating from the Hebbian rule and the arbor density of the synapses. For a nonlinear system, usually, there coexist several attractors for the same set of system parameters. That is, for a given set of the parameters, the state space comprises several attractive basins, each corresponding to a steady state respectively. The initial condition determines which attractor will be eventually reached. At the same time, a nonlinear system could have a different group of coexisting attractors for a different set of system parameters. That is, one could make the presence or absence of a specific state as a fixed point attractor by varying the set of the parameters. For a development model like Linsker's network, what is expected to be observed is that the different aRFs could emerge under different sets of parameters but should be relatively not sensitive to the initial conditions. In other words, the dynamics should avoid the coexistence of several attractors in an appropriate way. The purpose of this paper is to gain more insights into the dynamical mechanism of this self-organization model by performing a rigorous analysis on its parameter space without any approximation. That is, our goal is to reveal the effects of the system parameters on the stability of aRFs, and to predict precisely within which parameter regime the network will have the potential to develop a specially designated aRF. The novel rigorous approach presented here applies not only to the Linsker-type Hebbian learning but also to other related self-organization models about neural development. In Linsker's network, each cell in the present layer M receives synaptic inputs from a number of cells in the preceding layer C. The density of these synaptic connections decreases monotonically with distance rC, from the point underlying the M-cell's position. Since the synaptic weights change on a long time scale compared to the variation of random inputs, by averaging the Hebb rule over the ensemble of inputs in layer C, the dynamical equation for the development of the synaptic strength wT(i) between a M-cell and i-th C-cell at time Tis Nc. WT+1 (i) = f{wT(i) + k1 + l:)Qj + k2]r(j)wT(j)} (1) j=l where k1' k2 are system parameters which are particular combinations of the constants of the Hebb rule, r(?) is a non-negative normalized synaptic density function (SDF) 1, and L:iEC, rei) = 1, and 10 is a limiter function defined by I(x) = W max , if x > W max ; = x, if Ix I ~ W max ; and = -W max , if x < -w max ? The covariance IThe SDF is explicitly incorporated into the dynamics (1) which is equivalent to Linsker's formulation. A rigorous explanation for this equivalence is given in MacKay & Miller, 1990. A Rigorous Analysis of Linsker-type Hebbian Learning 321 matrix {Qij} of the layer C describes the correlation of activities of the i-th and the j-th C-cells. Actually, the covariance matrix of each layer is determined by SDFs r(?) of all layers preceding the layer under consideration. The idea of this paper is the following. It is well known that in general it is intractable to characterize the behavior of a nonlinear dynamics, since the nonlinearity is the cause of the coexistence of many attractors. And one has the difficulty in obtaining the complete characteristics of attractive basins in the state space. But usually for some cases, it is relatively easy to derive a necessary and sufficient condition to check whether a given state is a fixed point of the dynamics. In terms of this condition, the whole parameter regime for the emergence of a fixed point of the dynamics may be obtained in the parameter space. If we are further able to prove the stability of the fixed point, which implies that this fixed point is a steady state if the initial condition is in a non empty vicinity in the state space, we can assert the occurrence of this fixed point attractor in that parameter regime. For Linsker's network, fortunately, the above idea can be carried out because of the specific form of the nonlinear function 1(?). Due to space limitations, the rigorous proofs are in (Feng, Pan, & Roychowdhury, 1995). 2 The Set Of Saturated Fixed Point Attractors And The Criterion For The Division Of Parameter Regimes In fact, Linsker's model is a system of first-order nonlinear difference equations, taking the form = j[wr(i) + hi(Wr' kl' k2)], Wr = {wr(j),j = 1, .. . , Nc}, (2) where hi(W r , kl' k2) = kl + 2:f~dQ~ + k2]r(j)wr (j). And the aRFs observed in Linsker's simulation are the saturated fixed point attractors of this nonlinear system (2). Since the limiter function 1(?) is defined on a hypercube n = [-w max , wmax]N wr+l(i) C in weight state space within which the dynamics is dominated by the linear system wr+l(i) = wr(i) + hi(Wr, kl' k 2 ) , the short-time behaviors of evolution dynamics of connection patterns can be fully characterized in terms of the properties of eigenvectors and their eigenvalues. But this method of stability analysis will not be suitable for the long-time evolution of equation (1) or (2), provided the hypercube constraint is reached as the first largest component of W reaches saturation. However, it is well-known that a fixed point or an equilibrium state of dynamics (2) satisfies (3) Because of the special form of the nonlinear function 1(?), the fixed point equation (3) implies that 3T, such that for T > T, I wr(i) + hi(Wr, kl' k 2) I ~ Wmax , if hi (w, kl , k2) i= o. So a saturated fixed point Wr (i) must have the same sign as hi(wr , kl' k 2 ), i.e. wr(i)hi(Wr, kl' k 2 ) > o. By using the above idea, our Theorems 1 & 2 (proven in Feng, Pan, & Roychowdhury, 1995) state that the set of saturated fixed point attractors of the dynamics in 322 lian/eng Feng, H. Pan, V. P. Roychowdhury equation (1) is given by nFP = {w I w(i)hi(W r ,kl,k2) > 0,1::; i::; N.d, and w E nFP is stable, where the weight vector w belongs to the set of all extreme points of the hypercube n (we assume W max = 1 without loss of generality) . We next derive an explicit necessary and sufficient condition for the emergence of structured aRFs, i.e., we derive conditions to determine whether a given w belongs to nFP. Define J+(w) = {i I wei) = 1} as the index set of cells at the preceding layer C with excitatory weight for a connection pattern w, and J-(w) = {i I wei) = -1} as the index set of C-cells with inhibitory weight for w. Note from the property of fixed point attractors that a connection pattern w is an attractor of the dynamics (1) if and only if for i E J+(w), we have w(i){k 1 + l)Q~ + k2]r(j)w(j)} = j w(i){k 1 + EjEJ+(w)[Qt + k2]r(j)w(j) + EjEJ-(w)[Qt + k2]r(j)w(j)} > O. By the definition of J+(w) and J-(w), we deduce from the above inequality that kl 2: + [Q~ 2: + k2]r(j) - jEJ+(w) [Q~ + k2]r(j) > 0 jEJ-(w) namely kl + k2[ 2: 2: r(j) - jEJ+(w) r(j)] > jEJ-(w) 2: Q~r(j) - jEJ-(w) 2: Q~r(j). jEJ+(w) Inequality above is satisfied for all i in J+(w), and the left hand is independent of i. Hence, kl + k2[ 2: r(j) - jEJ+(w) 2: r(j)] > . max [ \EJ+(w) jEJ-(w) I: I: Q~r(j) - jEJ-(w) Q~r(j)]. jEJ+(w) On the other hand, for i E J-(w), we can similarly deduce that kl + k2[ 2: r(j) - jEJ+(w) 2: r(j)] < . min [ jEJ-(w) IEJ-(w) 2: Q~r(j) - jEJ-(w) 2: Q~r(j)] . jEJ+(w) We introduce the slope function: c(w) ~f 2: jEJ+(w) r(j) - 2: r(j) jEJ-(w) which is the difference of sums of the SDF r( ?) over J+(w) and J-(w), and two kl -intercept functions: d1(w) ~ { maxtEJ+(w)(EjEJ-(w) Qf;r(j) - EjEJ+(w) Qtr(j)), -00, if J+(w) # 0 if J+(w) = 0 A Rigorous Analysis of Linsker-type Hehbian Learning 323 A ~---k , (a) (b) o o E Figure 1: The parameter subspace of (kl, k 2 ). (a) Parameter regime of (kl' k 2 ) to ensure the emergence of all-excitatory (regime A) and all-inhibitory (regime B) connection patterns. The dark grey regime C is the coexistence regime for both all-excitatory and all-inhibitory connection patterns. And the regime D without texture are the regime that Linsker's simulation results are based on , in which both all-excitatory and all-inhibitory connection patterns are no longer an attractor. (b) The principal parameter regimes. Now from our Theorem 3 in Feng, Pan, & Roychowdhury, 1995, for every layer of Linsker's network, the new rigorous criterion for the division of stable parameter regimes to ensure the development of various structured connection patterns is d2 (w) > ki + c(w)k 2 > dl(w). That is, for a given SDF r(-), the parameter regime of (kl' k2) to ensure that w is a stable attractor of dynamics (1) is a band between two parallel lines ki + c(w)k 2 > dl(w) and ki +c(w)k 2 < d2 (w) (See regimes E and F in Fig.1(b)). It is noticed that as dl(w) > d2 (w), there is no regime of (kl' k2) for the occurrence of that aRF w as an attractor of equation (1). Therefore, the existence of such a structured aRF w as an attractor of equation (1) is determined by k1-intercept functions dIe) and d2 (?), and therefore by the covariance matrix Q.c or SDFs r(?) of all preceding layers. 3 Parameter Regimes For aRFs Between Layers BAnd C Based on our general theorems applicable to all layers, we mainly focus on describing the stabilization process of synaptic development from the 2nd (B) to the 3rd layer (C) by considering the effect of the system parameters on the weight development. For the sake of convenience, we assume that the input at 1st layer (A) is independent normal distribution with mean 0 and variance 1, and the connection strengths from layer A to B are all-excitatory same as in Linsker's simulations. The emergence of various aRFs between layer Band C have been previously studied in the literature, and in this paper we mention only the following new results made possible by our approach: (1) For the cell in layer C, the all-excitatory and the all-inhibitory connection patterns still have the largest stable regimes. Denote both SDFs from layer A to B and from B to C as r AB ( .,. ) and r BC (-) respectively. The parameter plane of (kl' k2) 324 lianfeng Feng, H. Pan, V. P. Roychowdhury Table 1: The Principal Parameter Regimes TYPE Regime A ATTRACTOR All-excitatory aRF Regime B All-inhibitory aRF Regime All-excitatory and all-inhibitory aRFs coexist The structured aRFs may have separate parameter regimes Any connection pattern in which the excitatory connections constitute the majority Any connection pattern in which the inhibitory connections constitute the majority A small coexistence regime of many connection patterns around the origin point of the parameter plane of ( kl , k2) =AnB Regime F d2{w) > kl where Regime G d2(W 1) > + C{W)k2 > d1(w) c(w) < 0 kl + c(w 1)k2 > d 1 (wI) =EnFnAnB is divided into four regimes by for all-excitatory pattern and for all-inhibitory pattern (See Fig.I(a?. (2) The parameter with large and negative k2 and approximately -1 < -kdk2 < 1 is favorable for the emergence of various structured connection patterns (e.g., ONcenter cells, OFF-center cells, bi-Iobed cells, and oriented cells) . This is because this regime (See regime D in Fig.I) is removed from the parameter regime where both all-excitatory and all-inhibitory aRFs are dominant, including the coexistence regime of many kind of at tractors around the origin point (See regime G in Fig.I(b The above results provide a precise picture about the principal parameter regimes summarized in Table 1. ?. (3) The relative size of the radiuses of two SDFs r AS (-,.) and r Sc (-) plays a key role in the evolution of various structured aRFs from B to C. A given SDF r.cM (i, j), i E M, j E e will be said to have a range r M if r.cM (i, j) is 'sufficient small' for lIi- jll ~ rM. For a Gaussian SDF r.cM(j,k) '" exp(-lIj-kll/r~), j E e,k EM, the range r M is its standard deviation. We give the analytic prediction about the influence of the SDF's ranges rs, rc on the dynamics by changing rs from the smallest extreme to the largest one with respect to rc. For the smallest extreme of rs (i.e. the A Rigorous Analysis of Linsker-type Hebbian Learning 325 synaptic connections from A to B are concentrated enough, and those from layer B to C are fully feedforward connected), we proved that any kind of connection pattern has a stable parameter regime and emerge under certain parameters, because each synaptic connection within an aRF is developed independently. As rB is changed from the smallest to the largest extreme, the development of synaptic connections between layer Band C will depend on each other stronger and stronger in the sense that most of connections have the same sign as their neighbors in an aRF. So for the largest extreme of rB (i.e. the weights from layer A to B are fully feedforward but there is no constraint on the SDF r BC (.)), any structured aRFs except for the all-excitatory and the all-inhibitory connection patterns will never arise at all, although there exist correlation in input activities (for a proof see Feng, Pan, & Roychowdhury, 1995). Th_erefore, without localized SDF, there would be no structured covariance matrix Q = {[Qij + k 2 ]r(j)} which embodies localized correlation in afferent activities. And without structured covariance matrix Q, no structured aRFs would emerge. (4) As another application of our analyses, we present several numerical results on the parameter regimes of (kl' k2' rB, rc) for the formation of various structured aRFs (Feng & Pan, 1993; Feng, Pan, & Roychowdhury, 1995) (where we assume that rAB(i,j) "" exp(-lli-jll/r~), i E B,j E A, and rBC(i) "" exp(-llill/r2), i E B as in (Linsker, 1986 & 1988)). For example, we show that various aRFs as at tractors have different relative stability. For a fixed rc, the SDF's range rB of the preceding layer as the third system parameter has various critical values for different attractors. That is, an attractor will no longer be stable if rB exceeds its corresponding critical value (See Fig. 2). For circularly symmetric ON-center cells, those aRFs with large ON-center core (which have positive or small negative slope value c(w) ~ -kt/k 2 ) always have a stable parameter regime. But for those ON-center cells with large negative slope value c(w), their stable parameter regimes decrease in size with c(w). Similarly, circularly symmetric OFF-center cells with large OFF-center core (which have negative or small positive slope value c(w)) will be more stable than those with large positive average of weights. But for non-circularly-symmetric patterns (e.g., bi-Iobed cells and oriented cells), only those at tractors with zero average synaptic strength might always have a stable parameter regime (See regime H in Fig.1(b)). If the third parameter rB is large enough to exceed its critical values for other aRFs and k2 is large and negative, then ON-center aRFs with positive c(w) and OFFcenter aRFs with negative c(w) will be almost only at tractors in regime DnE and regime DnF respectively. This conclusion makes it clear why we usually obtain ON-center aRFs in regime DnE and OFF-center aRFs in regime DnF much more easily than other patterns. 4 Concluding Remarks One advantage of our rigorous approach to this kind of unsupervised Hebbian learning network is that, without approximation, it unifies the treatment of many diverse problems about dynamical mechanisms. It is important to notice that there is no assumption on the second item hi(w T ) on the right hand side of equation (1), and there is no restriction on the matrix Q. Our Theorems 1 and 2 provide the general framework for the description of the fixed point attractors for any difference equation of the type stated in (2) that uses a limiter function. Depending on the 326 Jianfeng Feng, H. Pan. V. P. Roychowdhury ~ __-O-N-C-e-"-te-r-a--,R_FS~~____.(r c= 10) ~~ __-O-rie-"-ted---,a.R.-Fs------._,(r c= 10) . I~~~~~~~~~?~' . I~~~~~~~~~?~ 000@@@@ ??? 000~~~~ ? ? ? ~ __.O.... FF.-c ....8n ___ ter___a-R-Fs---__(r c= 10) ~ __-B-I-I-obed--a-R-F.. s ..--_ _ _(rc= 10) .I~~~~~~~~~?~' ooooo~a ??? Figure 2: The critical values of the SDF's range ra for different connection patterns. structure of the second item, hi(w T ) , it is not difficult to adapt our Theorem 3 to obtain the precise relationship among system parameters in other kind of models as long as 10 is a limiter function. Since the functions in the necessary and sufficient condition are computable (like our slope and k1-intercept functions), one is always able to check whether a designated fixed point is stable for a specific set of parameters. Acknowledgements The work ofV. P. Roychowdhury and H. Pan was supported in part by the General Motors Faculty Fellowship and by the NSF Grant No. ECS-9308814. J. Feng was partially supported by Chinese National Key Project of Fundamental Research "Climbing Program" and CNR of Italy. References R. Linsker. (1986) From basic network principle to neural architecture (series). Proc. Natl. Acad. Sci. USA 83: 7508-7512,8390-8394,8779-8783. R. Linsker. (1988) Self-organization in a perceptual network. Computer 21(3): 105-117. D. MacKay, & K. Miller. (1990) Analysis of Linsker's application of Hebbian rules to linear networks. Network 1: 257-297. J. Feng, & H. Pan. (1993) Analysis of Linsker-type Hebbian learning: Rigorous results. Proc. 1993 IEEE Int. Con! on Neural Networks - San Francisco Vol. III, 1516-1521. Piscataway, NJ: IEEE. J. Feng, H. Pan, & V. P. Roy chowdhury. (1995) Linsker-type Hebbian learning: A qualitative analysis on the parameter space. (submitted).
914 |@word faculty:1 stronger:2 oncenter:1 nd:1 grey:1 d2:6 simulation:5 r:3 eng:2 covariance:5 mention:1 initial:3 series:1 bc:2 must:1 numerical:2 analytic:1 motor:1 progressively:1 item:2 plane:2 short:1 core:2 provides:1 mathematical:1 rc:6 qualitative:2 qij:2 prove:1 uniroma1:1 kdk2:1 introduce:1 ra:1 expected:1 behavior:4 considering:1 provided:1 project:1 underlying:4 what:1 kind:4 cm:3 developed:1 nj:1 assert:1 every:1 ti:1 k2:24 rm:1 grant:1 positive:4 engineering:1 acad:1 approximately:1 might:1 studied:1 equivalence:1 bi:2 range:5 lafayette:1 word:1 convenience:1 coexist:2 influence:1 intercept:3 restriction:1 equivalent:1 center:9 independently:1 rule:5 insight:1 his:1 stability:4 variation:1 play:1 us:1 origin:2 roy:1 observed:2 role:2 electrical:1 connected:1 decrease:2 removed:1 qtr:1 iobed:2 dynamic:15 depend:1 ithe:1 division:2 ofv:1 easily:1 various:8 dnf:2 sc:1 formation:1 jianfeng:1 emergence:6 advantage:1 eigenvalue:1 propose:1 description:1 empty:1 rbc:1 derive:3 develop:2 depending:1 qt:2 school:1 implies:2 rei:1 radius:1 stabilization:1 dne:2 around:2 normal:1 exp:3 equilibrium:1 predict:2 smallest:3 purpose:2 favorable:1 proc:2 applicable:1 limiter:4 expose:1 sensitive:1 largest:5 gaussian:1 always:3 avoid:1 ej:1 varying:1 vwani:1 arf:7 focus:1 check:2 mainly:1 rigorous:12 sense:1 originating:2 among:2 development:7 special:1 mackay:2 field:4 never:1 ted:1 unsupervised:3 linsker:24 oriented:3 national:1 kll:1 attractor:21 ab:1 arborization:1 organization:4 chowdhury:1 saturated:5 extreme:5 natl:1 kt:1 coexisting:1 necessary:3 theoretical:1 deviation:1 characterize:1 st:1 density:5 fundamental:1 ie:1 off:4 satisfied:1 lii:1 potential:2 summarized:1 int:1 explicitly:1 afferent:2 reached:2 parallel:1 slope:5 variance:1 characteristic:1 ensemble:1 miller:2 climbing:1 unifies:1 lli:1 confirmed:1 submitted:1 synapsis:3 reach:1 synaptic:11 definition:1 proof:2 con:1 gain:1 coexistence:5 proved:1 treatment:1 actually:1 wei:2 rie:1 formulation:1 generality:1 correlation:3 hand:3 receives:1 wmax:2 nonlinear:10 defines:1 jll:2 rab:1 reveal:1 usa:1 effect:3 normalized:1 evolution:4 vicinity:1 hence:1 symmetric:3 attractive:2 adjacent:1 self:4 steady:2 die:1 criterion:2 complete:2 consideration:1 novel:2 sdf:11 ecn:2 surround:1 rd:1 similarly:2 centre:1 nonlinearity:1 stable:11 longer:2 deduce:2 dominant:1 italy:2 belongs:2 certain:1 inequality:2 fortunately:1 preceding:5 determine:3 monotonically:1 hebbian:13 exceeds:1 characterized:1 adapt:1 long:3 divided:1 prediction:2 basic:1 cell:18 fellowship:1 crucial:1 specially:2 cnr:1 presence:2 feedforward:2 exceed:1 easy:1 enough:2 iii:1 architecture:1 drum:1 idea:3 computable:1 qj:1 whether:3 f:2 cause:1 constitute:2 remark:1 clear:1 eigenvectors:1 ooooo:1 dark:1 band:4 concentrated:1 exist:1 nsf:1 inhibitory:11 notice:1 roychowdhury:11 sign:2 wr:15 rb:6 diverse:1 vol:1 group:1 key:2 four:1 changing:1 sum:1 parameterized:2 almost:1 sapienza:1 layer:25 hi:10 ki:3 played:1 activity:3 strength:3 occur:1 precisely:2 constraint:2 sake:1 dominated:1 min:1 concluding:1 performing:1 relatively:2 department:1 designated:3 structured:12 piscataway:1 combination:1 describes:1 pan:14 em:1 wi:1 equation:9 previously:1 describing:1 eventually:1 mechanism:3 appropriate:2 occurrence:2 existence:1 ensure:3 embodies:1 k1:4 chinese:1 hypercube:3 feng:15 noticed:1 receptive:4 primary:1 said:1 subspace:1 distance:1 separate:1 sci:1 majority:2 index:2 relationship:2 nc:2 difficult:1 negative:7 stated:1 perform:1 purdue:3 incorporated:1 precise:3 anb:1 rome:2 moro:1 namely:1 kl:23 connection:26 offcenter:1 able:2 usually:3 pattern:22 dynamical:3 regime:45 saturation:1 program:1 max:8 including:1 explanation:1 suitable:1 critical:4 difficulty:1 picture:2 carried:2 lij:1 understanding:1 literature:1 acknowledgement:1 tractor:4 relative:2 fully:3 loss:1 limitation:1 proven:1 localized:3 basin:2 sufficient:4 dq:1 principle:1 qf:1 excitatory:12 changed:1 supported:2 iej:1 side:1 nfp:3 neighbor:1 taking:1 emerge:3 made:1 san:1 ec:1 sdfs:4 ejej:4 francisco:1 un:1 why:1 table:2 iec:1 obtaining:1 multilayered:1 whole:1 arise:1 fig:6 referred:1 west:1 ff:1 hebb:2 position:1 comprises:1 explicit:1 perceptual:1 third:2 ix:1 theorem:5 specific:4 r2:1 dl:3 intractable:1 circularly:3 texture:1 te:1 jej:16 visual:1 partially:1 applies:1 determines:1 satisfies:1 goal:1 absence:2 change:1 determined:4 except:1 averaging:1 wt:4 principal:3 arbor:2 la:1 lian:2 d1:2 correlated:1
7,152
915
Template-Based Algorithms for Connectionist Rule Extraction Jay A. Alexander and Michael C. Mozer Department of Computer Science and Institute for Cognitive Science University of Colorado Boulder, CO 80309--0430 Abstract Casting neural network weights in symbolic terms is crucial for interpreting and explaining the behavior of a network. Additionally, in some domains, a symbolic description may lead to more robust generalization. We present a principled approach to symbolic rule extraction based on the notion of weight templates, parameterized regions of weight space corresponding to specific symbolic expressions. With an appropriate choice of representation, we show how template parameters may be efficiently identified and instantiated to yield the optimal match to a unit's actual weights. Depending on the requirements of the application domain, our method can accommodate arbitrary disjunctions and conjunctions with O(k) complexity, simple n-of-m expressions with O( k!) complexity, or a more general class of recursive n-of-m expressions with O(k!) complexity, where k is the number of inputs to a unit. Our method of rule extraction offers several benefits over alternative approaches in the literature, and simulation results on a variety of problems demonstrate its effectiveness. 1 INTRODUCTION The problem of understanding why a trained neural network makes a given decision has a long history in the field of connectionist modeling. One promising approach to this problem is to convert each unit's weights and/or activities from continuous numerical quantities into discrete, symbolic descriptions [2, 4, 8]. This type of reformulation, or rule extraction, can both explain network behavior and facilitate transfer of learning. Additionally, in intrinsically symbolic domains, there is evidence that a symbolic description can lead to more robust generalization [4]. 610 Jay A. Alexander, Michael C. Mozer We are interested in extracting symbolic rules on a unit-by-unit basis from connectionist nets that employ the conventional inner product activation and sigmoidal output functions. The basic language of description for our rules is that of n-of-m expressions. An n-of-m expression consists of a list of m subexpressions and a value n such that 1 ~ n ~ m. The overall expression is true when at least n of the m subexpressions are true. An example of an n-of-m expression stated using logical variables is the majority voter function X = 2 of (A, B, C). N-of-m expressions are interesting because they are able to model behaviors intermediate to standard Boolean OR (n = 1) and AND (n = m) functions. These intermediate behaviors reflect a limited form of two-level Boolean logic. (To see why this is true, note that the expression for X above is equivalent to AB + BC + AC.) In a later section we describe even more general behaviors that can be represented using recursive forms of these expressions. N-of-m expressions fit well with the activation behavior of sigmoidal units, and they are quite amenable to human comprehension. To extract an n-of-m rule from a unit's weights, we follow a three-step process. First we generate a minimal set of candidate templates, where each template is parameterized to represent a given n-of-m expression. Next we instantiate each template's parameters with optimal values. Finally we choose the symbolic expression whose instantiated template is nearest to the actual weights. Details on each of these steps are given below. 2 TEMPLATE-BASED RULE EXTRACTION 2.1 Background Following McMillan [4], we define a weight template as a parameterized region of weight space corresponding to a specific symbolic function. To see how weight templates can be used to represent symbolic functions, consider the weight vector for a sigmoidal unit with four inputs and a bias: w = WI w2 w3 w4 b Now consider the following two template vectors: t1 t2 = = -p P 0 -p P -p P P J.5p -D.5p These templates are parameterized by the variable p . Given a large positive value of p (say 5.0) and an input vector 1 (whose components are approximately 0 and 1), t1 describes the symbolic expression J of('lz, 12, 14), while t2 describes the symbolic expression 2 of(Ib 12? h. 14), A general description for n-of-m templates of this form is the following: 1. M of the weight values are set to ?p , P > 0; all others are set to O. (+p is used for normal subexpressions, -p for negated sUbexpressions) 2. The bias value is set to (0.5 + m neg - n)p, where m neg represents the number of negated subexpressions. When the inputs are Boolean with values -1 and +1, the form of the templates is the same, except the template bias takes the value (J + m - 2n)p. This seemingly trivial difference turns out to have a significant effect on the efficiency of the extraction process. Template-Based Algorithms for Connectionist Rule Extraction 611 2.2 Basic extraction algorithm Generating candidate templates Given a sigmoidal unit with k inputs plus a bias, the total number of n-of-m expressions that unit may compute is an exponential function of k: /c T /c m = ?.J ~ ~ ?.J m=ln=1 2m(k) m = ?.J ~ m=1 2 mk! (k-m)!(m-l)! = 2k3/C-1 For example, Tk=1O is 393,660, while Tk=20 is over 46 billion. Fortunately we can apply knowledge of the unit's actual weights to explore this search space without generating a template for each possible n-of-m expression. Alexander [1] proves that when the -11+1 input representation is used, we need consider at most one template for each possible choice of n and m. For a given choice of nand m, a template is indicated when sign(1 + m - 2n) =sign(b). A required template is formed by setting the template weights corresponding to the m highest absolute value actual weights to sp, where s represents the sign of the corresponding actual weight. The template bias is set to (1 + m - 2n)p. This reduces the number of templates required to a polynomial function of k: Values for Tk=1O and Tk=20 are now 30 and 110, respectively, making for a very efficient pruning of the search space. When 011 inputs are used, this simple procedure does not suffice and many more templates must be generated. For this reason, in the remainder of this paper we focus on the -11+ 1 case and assume the use of symmetric sigmoid functions. Instantiating template parameters Instantiating a weight template t requires finding a value for p such that the Euclidean distance d = lit - wl1 2 is minimized. Letting Uj = 1 if template weight tj is nonzero, Uj =0 otherwise, the value of p that minimizes this distance for any -11+ 1 template is given by: /c p* = L IWjIUj + (1 +m-2n)b ==-I!...-_ _ _ _ _~- I:....: ? m + (1 + m - 2n) 2 Finding the nearest template and checking extraction validity Once each template is instantiated with its value of p*, the distance between the template and the actual weight vector is calculated, and the minimal distance template is selected as the basis for rule extraction. Having found the nearest template t*, we can use its values as part of a rudimentary check on extraction validity. For example, we can define the extraction error as 100% x Ilt*-wI1 2/1IwI1 2 to measure how well the nearest symbolic rule fits the actual weights. We can also examine the value of p* used in t*. Small values of p* translate into activation levels in the linear regime of the sigmoid functions, compromising the assumption of Boolean outputs propagating to subsequent inputs. 612 Jay A. Alexander, Michael C. Mozer 2.3 Extending expressiveness While the n-of-m expressions treated thus far are fairly powerful, there is an interesting class of symbolic behaviors that cannot be captured by simple n-of-m expressions. The simplest example of this type of behavior may be seen in the single hidden unit version of xor described in [6]. In this 2-1-1 network the hidden unit H learns the expression AND(h 12), while the output unit (which connects to the two inputs as well as to the hidden unit) learns the expression AND[OR(h 12 ),Hj. This latter expression may be viewed as a nested or recursive form of n-of-m expression, one where some of the m subexpressions may themselves be n-of-m expressions. The following two forms of recursive n-of-m expressions are linearly separable and are thus computable by a single sigmoidal unit: OR [Cn-of-m ' COR j AND [C n-of-m , CAND j where is a nested n-of-m expression (1 S; n S; m) COR is a nested OR expression (n = 1) CAND is a nested AND expression (n =m) Cn-of-m These expressions may be seen to generalize simple n-of-m expressions in the same way that simple n-of-m expressions generalize basic disjunctions and conjunctions.1 We term the above forms augmented n-of-m expressions because they extend simple n-of-m expressions with additional disjuncts or conjuncts. Templates for these expressions (under the -1/+ 1 input representation) may be efficiently generated and instantiated using a procedure similar to that described for simple n-of-m expressions. When augmented expressions are included in the search, the total number of templates required becomes: This figure is O(k) worse than for simple n-of-m expressions, but it is still polynomial in k and is quite manageable for many problems. (Values for Tk=1O and Tk=20 are 150 and 1250, respectively.) A more detailed treatment of augmented n-of-m expressions is given in [1]. 3 RELATED WORK Here we briefly consider two alternative systems for connectionist rule extraction. Many other methods have been developed; a recent summary and categorization appears in [2]. 3.1 McMillan McMillan described the projection of actual weights to simple weight templates in [4]. McMillan's parameter selection and instantiation procedures are inefficient compared to those described here, though they yield equivalent results for the classes of templates he used. McMillan treated only expressions with m S; 2 and no negated sUbexpressions. 1 In fact the nesting may continue beyond one level. Thus sigmoidal units can compute expressions like OR[AND(Cn _of_m ' CAND), COR]. We have not yet experimented with extensions of this sort. Template-Based Algorithms for Connectionist Rule Extraction 613 3.2 Towell and Shavlik Towell and Shavlik [8] use a domain theory to initialize a connectionist network, train the network on a set of labeled examples, and then extract rules that describe the network's behavior. To perform rule extraction, Towell and Shavlik first group weights using an iterative clustering algorithm. After applying additional training, they typically check each training pattern against each weight group and eliminate groups that do not affect the classification of any pattern. Finally, they scan remaining groups and attempt to express a rule in purely symbolic n-of-m form. However, in many cases the extracted rules take the form of a linear inequality involving multiple numeric quantities. For example, the following rule was extracted from part of a network trained on the promoter recognition task [5] from molecular biology: Minus35 " -10 < + 5.0 * + 3.1 * + 1.9 * + 1.5 * - 1.5 * - 1.9 * - 3.1 * nt(@-37 nt(@-37 nt(@-37 nt(@-37 nt(@-37 nt(@-37 nt(@-37 '--T-G--A' ) '---GT---') '----C-CT' ) '---C--A-' ) '------GC' ) '--CAW---' ) '--A----C' ) where nt() returns the number of true subexpressions, @-37 locates the subexpressions on the DNA strand, and "_N indicates a don't-care subexpression. Towell and Shavlik's method can be expected to give more accurate results than our approach, but at a cost. Their method is very compute intensive and relies substantially on access to a fixed set of training patterns. Additionally, it is not clear that their rules are completely symbolic. While numeric expressions were convenient for the domains they studied, in applications where one is interested in more abstract descriptions, such expressions may be viewed as providing too much detail, and may be difficult for people to interpret and reason about. Sometimes one wants to determine the nearest symbolic interpretation of unit behavior rather than a precise mathematical description. Our method offers a simpler paradigm for doing this. Given these differences, we conclude that both methods have their place in rule extraction tool kits. 4 SIMULATIONS 4.1 Simple logic problems We used a group of simple logic problems to verify that our extraction algorithms could produce a correct set of rules for networks trained on the complete pattern space of each function. Table 1 summarizes the results. 2 The rule-plus-exception problem is defined as /= AB + 1\B CD; xor-l is the 2-1-1 version of xordescribed in Section 2.3; and xor-2 is a strictly layered (2-2-1) version of xor [6]. The negation problem is also described in [6]; in this problem one of the four inputs controls whether the other inputs appear normally or negated at the outputs. (As with xor-l, the network for negation makes use of direct inputJ output connections.) In addition to the perfect classification performance of the rules, the large values of p* and small values of extraction error (as defined in Section 2.2) provide evidence that the extraction process is very accurate. 614 Jay A. Alexander, MichaeL C. Mozer Averagep? Extraction Error Network Topology Hidden Unit Penalty Term Hidden Unit(s) Output Unit(s) Hidden Unit(s) Output Unit(s) Patterns Correctly Classified by Rules 4-2-1 - 2.72 6.15 0.8% 1.3% 100.0% xor-l 2-1-1 - 5.68 4.40 0.1 % 0.1 % 100.0 % xor-2 2-2-1 - 4.34 5.68 0.4% 1.0% 100.0 % negation 4-3-4 activation 5.40 5.17 0.2% 2.2% 100.0 % Problem rule-plus-exception Table 1: Simulation summary for simple logic problems Symbolic solutions for these problems often come in fonns different from the canonical fonn of the function. For example, the following rules for the rule-pLus-exception problem show a level of negation within the network: H1 H2 o OR (A, = B, c, D) AND (A, B) OR (H1' H2 ) Example results on xor- J show the expected use of an augmented n-of-m expression: 12) H OR (1 1 , o OR [AND(I 1 , 1 2 ), H! 4.2 The MONK's problems We tested generalization perfonnance using the MONK's problems [5,7], a set of three problems used to compare a variety of symbolic and connectionist learning algorithms. A summary of these tests appears in Table 2. Our perfonnance was equal to or better than all of the systems tested in [7] for the monks- J and monks-2 problems. Moreover, the rules extracted by our algorithm were very concise and easy to understand, in contrast to those produced by several of the symbolic systems. (The two connectionist systems reported in [7] were opaque, Le., no rules were extracted.) As an example, consider the following output for the monks-2 problem: HI 2 of (head_shape round, body_shape round, is_smiling yes, H2 3 of (head_shape round, body_shape round, is_smiling yes, holding sword, jacket_color red, has_tie not no) o AND holding sword, jacket_color red, has_tie yes) (HI' H2 ) The target concept for this problem is exactly 2 of the attributes have their first value. These rules demonstrate an elegant use of n-of-m expressions to describe the idea of "exactly 2" as "at least 2 but not 3". The monks-3 problem is difficult due to (intentional) training set noise, but our results are comparable to the other systems tested in [7]. 2 All results in this paper are for networks trained using batch-mode back propagation on the cross-entropy error function. Training was stopped when outputs were within 0.05 of their target values for each pattern or a fixed number of epochs (typically 10(0) was reached. Where indicated, a penalty term for nonBoolean hidden activations or hidden weight decay was added to the main error function. For the breast cancer problem shown in Table 4.3, hidden rules were extracted first and the output units were retrained briefly before extracting their rules. Results for the problems in Table 4.3 used leave-one-out testing or 100fold cross-validation (with 10 different initial orderings) as indicated. All results are averages over 10 replications with different initial weights. Template-Based Algorithms for Connectionist Rule Extraction 615 Problem Network Topology Hidden Unit Penalty Term monies-I 17-3--1 decay 124 100.0% 100.0% 432 100.0% 100.0% monles-2 17-2-1 decay 169 100.0% 100.0% 432 100.0% 100.0% monks-3 l7"'{)""l - 122 93.4% 93.4% 432 97.2% 97.2% Training Set Test Set #of Patterns Perf. of Network Perf. of Rules #of Patterns Perf. of Network Perf. of Rules Table 2: Simulation summary for the MONK's problems 4.3 VCI repository problems The final set of simulations addresses extraction performance on three real-world databases from the UCI repository [5]. Table 3 shows that good results were achieved. For the promoters task, we achieved generalization performance of nearly 88%, compared to 93-96% reported by Towell and Shavlik [8]. However, our results are impressive when viewed in light of the simplicity and comprehensibility of the extracted output. While Towell and Shavlik's results for this task included 5 rules like the one shown in Section 3.2, our single rule is quite simple: promoter = 5 of (@-45 'AA-------TTGA-A-----T------T-----AAA----C') Results for the house-votes-84 and breast-eaneer-wise problems are especially noteworthy since the generalization performance of the rules is virtually identical to that of the raw networks. This indicates that the rules are capturing a significant portion of the computation being performed by the networks. The following rule was the one most frequently extracted for the house-votes-84 problem, where the task is to predict party affiliation: V7 , V9 , V1O ' Vl l , V12 ) v4 Democrat OR [ 5 of (V 3 , where V3 voted for adoption-of-the-budget-resolution bill , 1 V4 voted for physician-fee-freeze bill V7 voted for anti-satellite-test-ban bill V9 = voted for rnx-missile bill V1 0 = voted for immigration bill Vll voted for synfuels-corporation-cutback bill V12 = voted for education-spending bill Shown below is a typical rule set extracted for the breast-eaneer-wise problem. Here the goal is to diagnose a tumor as benign or malignant based on nine clinical attributes. Malignant = AND (H 1 , H2 ) 4 of (thickness> 3, size> 1, adhesion> 1, epithelial> 5, nuclei> 3, chromatin> 1, normal> 2, mi toses > 1) = 30f(thickness>6,size>1,shape>1,epithelial>1, nuclei> 8, normal> 9) = not used Hl = H2 H3 As suggested by the rules, we used a thermometer (cumulative) coding of the nominally valued attributes so that less-than or greater-than subexpressions could be efficiently represented in the hidden weights. Such a representation is often useful in diagnosis tasks. We also limited the hidden weights to positive values due to the nature of the attributes. 616 Jay A. Alexander, Michael C. Mozer Training Set Test Set Problem Network Topology #I of Patterns Perf. of Network Perf. of Rules #I of Patterns Perf. of Network Perf. of Rules promoters 2284-1 105 100.0% 95.9% I 94.2% 87.6% house-votes-84 164-1 387 97.3 % 96.2% 43 95.7% 95.9% breast-cancer-wisc 81-3-1 630 98.5 % 96.3 % 70 95.8% 95.2% Table 3: Simulation summary for uel repository problems Taken as a whole our simulation results are encouraging, and we are conducting further research on rule extraction for more complex tasks. 5 CONCLUSION We have described a general approach for extracting various types of n-of-m symbolic rules from trained networks of sigmoidal units, assuming approximately Boolean activation behavior. While other methods for interpretation of this sort exist, ours represents a valuable price/performance point, offering easily-understood rules and good extraction performance with computational complexity that scales well with the expressiveness desired. The basic principles behind our approach may be flexibly applied to a wide variety of problems. References [1] Alexander, J. A. (1994). Template-based procedures for neural network interpretation. MS Thesis. Department of Computer Science, University of Colorado, Boulder, CO. [2] Andrews, R., Diederich, l, and Tickle, A. B. (1995). A survey and critique of techniques for extracting rules from trained artificial neural networks. To appear in Fu, L. M. (Ed.), Knowledge-Based Systems, Special Issue on Knowledge-Based Neural Networks. [3] Mangasarian, O. L. and Wolberg, W. H. (1990). Cancer diagnosis via linear programming. SIAM News 23:5, pages 1 & 18. [4] McMillan, C. (1992). Rule induction in a neural network through integrated symbolic and subsymbolic processing. PhD Thesis. Department of Computer Science, University of Colorado, Boulder, CO. [5] Murphy, P. M. and Aha, D. W. (1994). UCI repository of machine learning databases. [Machine-readable data repository]. Irvine, CA: University of California, Department of Information and Computer Science. Monks data courtesy of Sebastian Thrun, promoters data courtesy of M. Noordewier and J. Shavlik, congressional voting data courtesy of Jeff Schlimmer, breast cancer data courtesy of Dr. William H. Wolberg (see also [3] above). [6] Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986). Learning internal representations by error propagation. In Rumelhart, D. E., McClelland, l L., and the PDP Research Group, Parallel Distributed Processing: Explorations in the Microstructure of Cognition. Volume 1: Foundations, pages 318-362. Cambridge, MA: MIT Press. [7] Thrun, S. B., and 23 other authors (1991). The MONK's problems - A performance comparison of different learning algorithms. Technical Report CS-CMU-91-197. Carnegie Mellon University, Pittsburgh, PA. [8] Towell, G. and Shavlik, J. W. (1992). Interpretation of artificial neural networks: Mapping knowledge-based neural networks into rules. In Moody, J. E., Hanson, S. J., and Lippmann, R. P. (Eds.), Advances in Neurallnfonnation Processing Systems, 4:977-984. San Mateo, CA: Morgan Kaufmann.
915 |@word repository:5 briefly:2 version:3 manageable:1 polynomial:2 simulation:7 fonn:1 concise:1 accommodate:1 initial:2 offering:1 bc:1 ours:1 nt:8 activation:6 yet:1 must:1 numerical:1 subsequent:1 benign:1 shape:1 instantiate:1 selected:1 monk:10 inputj:1 sigmoidal:7 simpler:1 thermometer:1 mathematical:1 direct:1 replication:1 consists:1 expected:2 wl1:1 themselves:1 examine:1 cand:3 frequently:1 behavior:11 actual:8 encouraging:1 becomes:1 moreover:1 suffice:1 minimizes:1 substantially:1 developed:1 finding:2 corporation:1 voting:1 exactly:2 control:1 unit:26 normally:1 appear:2 t1:2 positive:2 before:1 understood:1 critique:1 approximately:2 noteworthy:1 plus:4 voter:1 studied:1 mateo:1 rnx:1 co:3 limited:2 adoption:1 testing:1 recursive:4 procedure:4 w4:1 projection:1 convenient:1 symbolic:23 cannot:1 selection:1 layered:1 applying:1 conventional:1 equivalent:2 bill:7 courtesy:4 williams:1 flexibly:1 survey:1 resolution:1 simplicity:1 rule:50 nesting:1 notion:1 vll:1 target:2 colorado:3 programming:1 pa:1 rumelhart:2 recognition:1 labeled:1 database:2 region:2 news:1 ordering:1 highest:1 valuable:1 principled:1 mozer:5 tickle:1 complexity:4 trained:6 purely:1 efficiency:1 basis:2 completely:1 easily:1 represented:2 various:1 train:1 instantiated:4 describe:3 artificial:2 disjunction:2 quite:3 whose:2 valued:1 say:1 otherwise:1 final:1 seemingly:1 net:1 product:1 remainder:1 uci:2 translate:1 description:7 billion:1 requirement:1 extending:1 satellite:1 produce:1 generating:2 categorization:1 perfect:1 leave:1 tk:6 depending:1 andrew:1 ac:1 propagating:1 nearest:5 h3:1 c:1 come:1 correct:1 compromising:1 attribute:4 exploration:1 human:1 education:1 microstructure:1 generalization:5 comprehension:1 extension:1 strictly:1 intentional:1 normal:3 k3:1 cognition:1 predict:1 mapping:1 wi1:1 epithelial:2 tool:1 mit:1 rather:1 hj:1 casting:1 conjunction:2 focus:1 check:2 indicates:2 contrast:1 vl:1 typically:2 eliminate:1 nand:1 integrated:1 hidden:12 interested:2 overall:1 classification:2 l7:1 issue:1 special:1 fairly:1 initialize:1 field:1 once:1 equal:1 extraction:24 having:1 biology:1 represents:3 lit:1 identical:1 nearly:1 minimized:1 connectionist:10 t2:2 others:1 report:1 employ:1 murphy:1 connects:1 william:1 negation:4 ab:2 attempt:1 immigration:1 light:1 behind:1 tj:1 schlimmer:1 amenable:1 accurate:2 fu:1 perfonnance:2 euclidean:1 aha:1 desired:1 v1o:1 minimal:2 mk:1 stopped:1 modeling:1 boolean:5 cost:1 noordewier:1 too:1 reported:2 thickness:2 siam:1 v4:2 physician:1 michael:5 moody:1 thesis:2 reflect:1 choose:1 dr:1 worse:1 cognitive:1 v7:2 inefficient:1 return:1 coding:1 later:1 h1:2 performed:1 diagnose:1 doing:1 red:2 reached:1 sort:2 portion:1 parallel:1 subexpression:1 formed:1 voted:7 xor:8 kaufmann:1 v9:2 efficiently:3 conducting:1 yield:2 yes:3 generalize:2 raw:1 produced:1 history:1 classified:1 explain:1 sebastian:1 ed:2 diederich:1 against:1 mi:1 irvine:1 treatment:1 intrinsically:1 logical:1 knowledge:4 back:1 appears:2 follow:1 though:1 vci:1 propagation:2 mode:1 indicated:3 facilitate:1 effect:1 validity:2 verify:1 true:4 concept:1 averagep:1 symmetric:1 nonzero:1 round:4 m:1 complete:1 demonstrate:2 interpreting:1 rudimentary:1 spending:1 wise:2 mangasarian:1 sigmoid:2 volume:1 extend:1 he:1 interpretation:4 interpret:1 significant:2 mellon:1 caw:1 freeze:1 cambridge:1 language:1 access:1 impressive:1 gt:1 neurallnfonnation:1 recent:1 inequality:1 affiliation:1 continue:1 neg:2 captured:1 seen:2 fortunately:1 additional:2 care:1 kit:1 greater:1 morgan:1 determine:1 paradigm:1 v3:1 multiple:1 reduces:1 technical:1 match:1 offer:2 long:1 cross:2 clinical:1 molecular:1 locates:1 instantiating:2 involving:1 basic:4 fonns:1 breast:5 cmu:1 represent:2 sometimes:1 achieved:2 background:1 want:1 addition:1 adhesion:1 crucial:1 w2:1 comprehensibility:1 elegant:1 virtually:1 effectiveness:1 extracting:4 intermediate:2 easy:1 congressional:1 variety:3 affect:1 fit:2 w3:1 identified:1 topology:3 inner:1 idea:1 cn:3 computable:1 intensive:1 whether:1 expression:44 penalty:3 nine:1 wolberg:2 useful:1 detailed:1 clear:1 mcclelland:1 simplest:1 dna:1 generate:1 exist:1 canonical:1 sign:3 towell:7 correctly:1 diagnosis:2 discrete:1 carnegie:1 express:1 group:6 four:2 reformulation:1 wisc:1 v1:1 convert:1 parameterized:4 powerful:1 ilt:1 opaque:1 place:1 v12:2 decision:1 summarizes:1 fee:1 comparable:1 capturing:1 ct:1 hi:2 fold:1 activity:1 separable:1 missile:1 department:4 describes:2 wi:1 making:1 aaa:1 hl:1 boulder:3 taken:1 ln:1 turn:1 malignant:2 letting:1 cor:3 apply:1 appropriate:1 alternative:2 batch:1 clustering:1 remaining:1 readable:1 prof:1 uj:2 especially:1 added:1 quantity:2 distance:4 thrun:2 majority:1 trivial:1 reason:2 induction:1 assuming:1 providing:1 difficult:2 holding:2 stated:1 negated:4 perform:1 anti:1 hinton:1 precise:1 pdp:1 gc:1 arbitrary:1 retrained:1 mcmillan:6 expressiveness:2 required:3 connection:1 hanson:1 california:1 address:1 able:1 beyond:1 suggested:1 below:2 pattern:10 regime:1 treated:2 perf:8 extract:2 epoch:1 literature:1 understanding:1 checking:1 subexpressions:10 interesting:2 validation:1 h2:6 nucleus:2 foundation:1 sword:2 principle:1 cd:1 cancer:4 summary:5 ban:1 bias:5 understand:1 institute:1 explaining:1 template:41 shavlik:8 wide:1 absolute:1 benefit:1 distributed:1 calculated:1 numeric:2 world:1 cumulative:1 author:1 san:1 lz:1 far:1 party:1 pruning:1 lippmann:1 logic:4 instantiation:1 pittsburgh:1 conclude:1 don:1 continuous:1 search:3 iterative:1 why:2 table:8 additionally:3 promising:1 nature:1 transfer:1 robust:2 ca:2 conjuncts:1 complex:1 domain:5 sp:1 main:1 promoter:5 linearly:1 whole:1 noise:1 augmented:4 exponential:1 candidate:2 house:3 ib:1 jay:5 learns:2 subsymbolic:1 specific:2 list:1 experimented:1 decay:3 evidence:2 phd:1 budget:1 entropy:1 democrat:1 explore:1 strand:1 nominally:1 aa:1 nested:4 relies:1 extracted:8 ma:1 viewed:3 goal:1 jeff:1 price:1 included:2 typical:1 except:1 tumor:1 total:2 vote:3 exception:3 internal:1 people:1 latter:1 scan:1 alexander:7 tested:3 chromatin:1
7,153
916
An Actor/Critic Algorithm that Equivalent to Q-Learning ? IS Robert H. Crites Computer Science Department University of Massachusetts Amherst, MA 01003 Andrew G. Barto Computer Science Department University of Massachusetts Amherst, MA 01003 crites~cs.umass.edu barto~cs.umass.edu Abstract We prove the convergence of an actor/critic algorithm that is equivalent to Q-Iearning by construction. Its equivalence is achieved by encoding Q-values within the policy and value function of the actor and critic. The resultant actor/critic algorithm is novel in two ways: it updates the critic only when the most probable action is executed from any given state, and it rewards the actor using criteria that depend on the relative probability of the action that was executed. 1 INTRODUCTION In actor/critic learning systems, the actor implements a stochastic policy that maps states to action probability vectors, and the critic attempts to estimate the value of each state in order to provide more useful reinforcement feedback to the actor. The result is two interacting adaptive processes: the actor adapts to the critic, while the critic adapts to the actor. The foundations of actor/critic learning systems date back at least to Samuel's checker program in the late 1950s (Samuel,1963). Examples of actor/critic systems include Barto, Sutton, & Anderson's (1983) ASE/ ACE architecture and Sutton's (1990) Dyna-PI architecture. Sutton (1988) notes that the critic in these systems performs temporal credit assignment using what he calls temporal difference (TD) methods. Barto, Sutton, & Watkins (1990) note a relationship between actor/critic 402 Robert Crites, Andrew G. Barto architectures and a dynamic programming (DP) algorithm known as policy iteration. Although DP is a collection of general methods for solving Markov decision processes (MDPs), these algorithms are computationally infeasible for problems with very large state sets. Indeed, classical DP algorithms require multiple complete sweeps of the entire state set. However, progress has been made recently in developing asynchronous, incremental versions of DP that can be run online concurrently with control (Watkins, 1989; Barto et ai, 1993). Most of the theoretical results for incremental DP have been for algorithms based on a DP algorithm known as value iteration. Examples include Watkins' (1989) Q-Iearning algorithm (motivated by a desire for on-line learning), and Bertsekas & Tsitsiklis' (1989) results on asynchronous DP (motivated by a desire for parallel implementations). Convergence proofs for incremental algorithms based on policy iteration (such as actor/critic algorithms) have been slower in coming. Williams & Baird (1993) provide a valuable analysis of the convergence of certain actor/critic learning systems that use deterministic policies. They assume that a model of the MDP (including all the transition probabilities and expected rewards) is available, allowing the use of operations that look ahead to all possible next states. When a model is not available for the evaluation of alternative actions, one must resort to other methods for exploration, such as the use of stochastic policies. We prove convergence for an actor/critic algorithm that uses stochastic policies and does not require a model of the MDP. The key idea behind our proof is to construct an actor/critic algorithm that is equivalent to Q-Iearning. It achieves this equivalence by encoding Q-values within the policy and value function of the actor and critic. By illustrating the way Qlearning appears as an actor/critic algorithm, the construction sheds light on two significant differences between Q-Iearning and traditional actor/critic algorithms. Traditionally, the critic attempts to provide feedback to the actor by estimating V1I', the value function corresponding to the current policy'll". In our construction, instead of estimating ~, the critic directly estimates the optimal value function V"'. In practice, this means that the value function estimate V is updated only when the most probable action is executed from any given state. In addition, our actor is provided with more discriminating feedback, based not only on the TD error, but also on the relative probability of the action that was executed. By adding these modifications, we can show that this algorithm behaves exactly like Q-Iearning constrained by a particular exploration strategy. Since a number of proofs of the convergence of Q-Iearning already exist (Tsitsiklis, 1994; Jaakkola et ai, 1993; Watkins & Dayan, 1992), the fact that this algorithm behaves exactly like Q-Iearning implies that it too converges to the optimal value function with probability one. 2 MARKOV DECISION PROCESSES Actor/critic and Q-Iearning algorithms are usually studied within the Markov decision process framework. In a finite MDP, at each discrete time step, an agent observes the state :z: from a finite set X, and selects an action a from a finite set Ax by using a stochastic policy'll" that assigns a probability to each action in Ax. The agent receives a reward with expected value R(:z:, a), and the state at the next An Actor/Critic Algorithm That Is Equivalent to Q-Learning 403 time step is y with probability pll(:z:,y). For any policy 7?' and :z: E X, let V""(:Z:) denote the ezpected infinite-horizon discounted return from :z: given that the agent uses policy 7?'. Letting rt denote the reward at time t, this is defined as: V""(:z:) = E7r [L:~o,trtl:z:o = :z:], (1) where :Z:o is the initial state, 0 :::; , < 1 is a factor used to discount future rewards, and E7r is the expectation assuming the agent always uses policy 7?'. It is usual to call V 7r (:z:) the value of:z: under 7?'. The function V"" is the value function corresponding to 7?'. The objective is to find an optimal policy, i.e., a policy,7?'*, that maximizes the value of each state :z: defined by (1). The unique optimal value function, V*, is the value function corresponding to any optimal policy. Additional details on this and other types of MOPs can be found in many references. 3 ACTOR/CRITIC ALGORITHMS A generic actor/critic algorithm is as follows: 1. Initialize the stochastic policy and the value function estimate. 2. From the current state :z:, execute action a randomly according to the current policy. Note the next state y, the reward r, and the TO error e = [r + ,V(y)] - V(:z:), where 0 :::; , < 1 is the discount factor. 3. Update the actor by adjusting the action probabilities for state :z: using the TO error. If e > 0, action a performed relatively well and its probability should be increased. If e < 0, action a performed relatively poorly and its probability should be decreased. 4. Update the critic by adjusting the estimated value of state :z: using the TO error: V(:z:) -- V(:z:) +a e where a is the learning rate. 5. :z: -- y. Go to step 2. There are a variety of implementations of this generic algorithm in the literature. They differ in the exact details of how the policy is stored and updated. Barto et al (1990) and Lin (1993) store the action probabilities indirectly using parameters w(:z:, a) that need not be positive, and need not sum to one. Increasing (or decreasing) the probability of action a in state :z: is accomplished by increasing (or decreasing) the value of the parameter w(:z:, a). Sutton (1990) modifies the generic algorithm so that these parameters can be interpreted as action value estimates. He redefines e in step 2 as follows: e = [r + ,V(y)] - w(:z:, a). For this reason, the Oyna-PI architecture (Sutton, 1990) and the modified actor/critic algorithm we present below both reward less probable actions more readily because of their lower estimated values. 404 Robert Crites, Andrew G. Barto Barto et al (1990) select actions by adding exponentially distributed random numbers to each parameter w(:z:, a) for the current state, and then executing the action with the maximum sum. Sutton (1990) and Lin (1993) convert the parameters w(:z:, a) into action probabilities using the Boltzmann distribution, where given a temperature T, the probability of selecting action i in state :z: is ew(x,i)/T '" L.JaEA .. ew(x,a)/T' In spite of the empirical success of these algorithms, their convergence has never been proven. 4 Q-LEARNING Rather than learning the values of states, the Q-Iearning algorithm learns the values of state/action pairs. Q(:z:, a) is the expected discounted return obtained by performing action a in state :z: and performing optimally thereafter. Once the Q function has been learned, an optimal action in state :z: is any action that maximizes Q(:z:, .). Whenever an action a is executed from state :z:, the Q-value estimate for that state/action pair is updated as follows: Q(:z:, a) +- Q(:z:, a) + O!xa(n) [r + "y maxbEAlI Q(y, b) - Q(:z:, a)), where O!xa (n) is the non-negative learning rate used the nth time action a is executed from state :z:. Q-Learning does not specify an exploration mechanism, but requires that all actions be tried infinitely often from all states. In actor/critic learning systems, exploration is fully determined by the action probabilities of the actor. 5 A MODIFIED ACTOR/CRITIC ALGORITHM For each value v E !R, the modified actor/critic algorithm presented below uses an invertible function, H.", that assigns a real number to each action probability ratio: H1J : (0,00) -+ !R. Each H." must be a continuous, strictly increasing function such that H.,,(l) and HH.. (Z2)(i;) = H",(Zl) for all Zl,Z2 = v, > o. = One example of such a class of functions is H.,,(z) T In(z) + v, v E !R, for some positive T. This class of functions corresponds to Boltzmann exploration in Qlearning. Thus, a kind of simulated annealing can be accomplished in the modified actor/critic algorithm (as is often done in Q-Iearning) by gradually lowering the "temperature" T and appropriately renormalizing the action probabilities. It is also possible to restrict the range of H." if bounds on the possible values for a given MDP are known a priori. For a state :z:, let Pa be the probability of action a, let Pmax be the probability of the most probable action, a max , and let Za = ~. An Actor/Critic Algorithm That Is Equivalent to Q-Leaming 405 The modified actor/critic algorithm is as follows: 1. Initialize the stochastic poliCj and the value function estimate. 2. From the current state :z:, execute an action randomly according to the current policy. Call it action i. Note the next state y and the immediate reward r, and let e = [r + -yV(y)] - Hy(X) (Zi). 3. Increase the probability of action i if e > 0, and decrease its probability if e < O. The precise probability update is as follows. First calculate zt = H~tX)[HY(x)(Zi) + aXi(n) e]. Then determine the new action probabilities by dividing by normalization factor N = zt + E#i Zj, as follows: a:~ +-:W, Pi and Pj +- a:' =jt, j =P i. 4. Update V(:z:) only if i = UomIlX' Since the action probabilities are updated after every action, the most probable action may be different before and after the update. If i = amllx both before and after step 3 above, then update the value function estimate as follows: V(:z:) Otherwise, if i = UomIlX +- V(:z:) + aXi(n) e before or after step 3: V(:z:) +- HY(x)(Npk), where action Ie is the most probable action after step 3. 5. :z: 6 +- y. Go to step 2. CONVERGENCE OF THE MODIFIED ALGORITHM The modified actor/critic algorithm given above converge6 to the optimal value function V? with probability one if: Theorem: 1. The 6tate and action 6et6 are finite. 2. E:=o axlI(n) =00 and E:=o a!lI(n) < 00. Space does not permit us to supply the complete proof, which follows this outline: 1. The modified actor/critic algorithm behaves exactly the same as a Q- learning algorithm constrained by a particular exploration strategy. 2. Q-Iearning converges to V? with probability one, given the conditions above (Tsitsiklis, 1993; Jaakkola et aI, 1993; Watkins & Dayan, 1992). 3. Therefore, the modified actor/critic algorithm also converges to V? with probability one. Robert Crites, Andrew G. Barto 406 The commutative diagram below illustrates how the modified actor/critic algorithm behaves exactly like Q-Iearning constrained by a particular exploration strategy. The function H recovers Q-values from the policy ?r and value function V. H- 1 recovers (?r, V) from the Q-values, thus determining an exploration strategy. Given the ability to move back and forth between (?r, V) and Q, we can determine how to change (?r, V) by converting to Q, determining updated Q-values, and then converting back to obtain an updated (?r, V). The modified actor/critic algorithm simply collapses this process into one step, bypassing the explicit use of Q-values. (7r , VA) t Modified Actor/Critic -------.. ( 7r , VA) t+1 H-l H A A (Jt ------------Q---L-ea-r-ru-?n-g----------~~ (Jt+1 Following the diagram above, (?r, V) can be converted to Q-values as follows: Going the other direction, Q-values can be converted to (?r, V) as follows: and The only Q-value that should change at time t is the one corresponding to the state/action pair that was visited at time tj call it Q(:z:, i). In order to prove the convergence theorem, we must verify that after an iteration ofthe modified actor/critic algorithm, its encoded Q-values match the values produced by Q-Iearning: Qt+1(:Z:, a) = Qt(:Z:, i) + Qx.(n) [r + "y max Qt(Y, b) - Qt(:Z:, i)], a = i. bEAli (2) (3) In verifying this, it is necessary to consider the four cases where Q(:z:, i) is, or is not, the maximum Q-value for state :z: at times t and t + 1. Only enough space exists to present a detailed verification of one case. Case 1: Qt(:Z:, i) = ma:z: Qt(:Z:,?) and Qt+l(:Z:, i) = ma:z: Qt+l(:Z:, .). In this case, Jli(t) = Pmax(t) and P.(t + 1) = Pmax(t + 1), since Hyt(x) and H yt +1 (x) are strictly increasing. Therefore Zi (t) = 1 and Zi (t + 1) = 1. Therefore, Vi (:z:) = HYt (x)[1] = HYt(x)[Zi(t)] = Qt(:Z:, i), and An Actor/Critic Algorithm That Is Equivalent to Q-Learning 407 H yt+1 (x) [Zi(t + 1)] Hyt+ 1 (x) [1] Vt+l(:C) Vt(:c) + O!xi(n) e Qt(:Z:, i) + O!xi(n) [r + "y max Qt(Y, b) - Qt(:Z:, i)]. bEAJI This establishes (2). To show that (3) holds, we have that Vt+l(:Z:) Vt(:c) + O!xi(n) e Qt(:z:, i) + O!xi(n) e HYt(x)[Zi(t)] + O!xi(n) e HYt(x)[H~t~x)[HYt(x)[Zi(t)] + O!xi(n) e]] (4) Hyt(x) [zt(t)] and H yt + 1 (x) [za(t + 1)] H [ Pa(t + 1) ] Yt+ 1 (x) Pmax(t + 1) H [Za(t)/lV] Yt+ 1 (x) zt(t)/lV H [Za(t)] Yt+ 1 (x) zt(t) i f a i= i za(t) HHfrt(.. )[zt(t)][zt(t) 1 by (4) Hyt(x) [za(t)] by a property of H Qt(:Z:, a). The other cases can be shown similarly. 7 CONCLUSIONS We have presented an actor/critic algorithm that is equivalent to Q-Iearning constrained by a particular exploration strategy. Like Q-Iearning, it estimates V" directly without a model of the underlying decision process. It uses exactly the same amount of storage as Q-Iearning: one location for every state/action pair. (For each state, IAI- 1 locations are needed to store the action probabilities, since they must sum to one. The remaining location can be used to store the value of that state.) One advantage of Q-Iearning is that its exploration is uncoupled from its value function estimates. In the modified actor/critic algorithm, the exploration strategy is more constrained. 408 Robert Crites, Andrew G. Barto It is still an open question whether other actor/critic algorithms are guaranteed to converge. One way to approach this question would be to investigate further the relationship between the modified actor/critic algorithm described here and the actor/critic algorithms that have been employed by others. Acknowledgements We thank Vijay GullapaUi and Rich Sutton for helpful discussions. This research was supported by Air Force Office of Scientific Research grant F49620-93-1-0269. References A. G. Barto, S. J. Bradtke &. S. P. Singh. (1993) Learning to act using real-time dynamic programming. Artificial Intelligence, Accepted. A. G. Barto, R. S. Sutton &. C. W. Anderson. (1983) Neuronlike adaptive elements that can solve difficult learning control problems. IEEE Transactions on Systems, Man, and Cybernetics 13:835-846. A. G. Barto, R. S. Sutton &. C. J. C. H. Watkins. (1990) Learning and sequential decision making. In M. Gabriel &. J. Moore, editors, Learning and Computational Neuroscience: Foundations of Adaptive Networks. MIT Press, Cambridge, MA. D. P. Bertsekas &. J. N. Tsitsiklis. (1989) Parallel and Distributed Computation: Numerical Metkods. Prentice-Hall, Englewood Cliffs, NJ. T. Jaakkola, M. 1. Jordan &. S. P. Singh. (1993) On the convergence of stochastic iterative dynamic programming algorithms. MIT Computational Cognitive Science Technical Report 9307. L. Lin. (1993) Reinforcement Learning for Robots Using Neural Networks. PhD Thesis, Carnegie Mellon University, Pittsburgh, PA. A. L. Samuel. (1963) Some studies in machine learning using the game of checkers. In E. Feigenbaum &. J. Feldman, editors, Computers and Tkougkt. McGraw-Hill, New York, NY. R. S. Sutton. (1988) Learning to predict by the methods of temporal differences. Mackine Learning 3:9-44. R. S. Sutton. (1990) Integrated architectures for learning, planning, and reacting based on approximating dynamic programming. In Proceedings of tke Seventk International Conference on Mackine Learning. J. N. Tsitsiklis. (1994) Asynchronous stochastic approximation and Q-Iearning. Mackine Learning 16:185-202. C. J. C. H. Watkins. (1989) Learning from Delayed Rewards. PhD thesis, Cambridge University. C. J. C. H. Watkins &. P. Dayan. (1992) Q-Iearning. Mackine Learning 8:279-292. R. J. Williams &. L. C. Baird. (1993) Analysis ofsome incremental variants of policy iteration: first steps toward understanding actor-critic learning systems. Technical Report NU-CCS-93-11. Northeastern University College of Computer Science. PART V ALGORITHMS AND ARCIDTECTURES
916 |@word illustrating:1 version:1 open:1 tried:1 initial:1 uma:2 selecting:1 current:6 z2:2 must:4 readily:1 numerical:1 update:7 intelligence:1 location:3 supply:1 prove:3 indeed:1 expected:3 planning:1 discounted:2 decreasing:2 td:2 increasing:4 provided:1 estimating:2 underlying:1 maximizes:2 what:1 kind:1 interpreted:1 nj:1 temporal:3 every:2 act:1 iearning:19 shed:1 exactly:5 control:2 zl:2 grant:1 bertsekas:2 positive:2 before:3 sutton:12 encoding:2 cliff:1 reacting:1 studied:1 equivalence:2 collapse:1 range:1 unique:1 practice:1 implement:1 empirical:1 spite:1 storage:1 prentice:1 equivalent:7 map:1 deterministic:1 yt:6 modifies:1 williams:2 go:2 assigns:2 jli:1 traditionally:1 updated:6 construction:3 exact:1 programming:4 us:5 pa:3 element:1 verifying:1 calculate:1 decrease:1 valuable:1 observes:1 reward:9 dynamic:4 depend:1 solving:1 singh:2 tx:1 artificial:1 ace:1 encoded:1 solve:1 otherwise:1 ability:1 online:1 advantage:1 coming:1 date:1 poorly:1 adapts:2 forth:1 pll:1 convergence:9 renormalizing:1 incremental:4 converges:3 executing:1 andrew:5 qt:14 progress:1 dividing:1 c:2 implies:1 differ:1 direction:1 stochastic:8 exploration:11 require:2 probable:6 strictly:2 bypassing:1 hold:1 credit:1 hall:1 predict:1 mop:1 achieves:1 visited:1 v1i:1 establishes:1 mit:2 concurrently:1 always:1 feigenbaum:1 modified:15 rather:1 barto:14 jaakkola:3 office:1 ax:2 e7r:2 helpful:1 tate:1 dayan:3 entire:1 integrated:1 going:1 selects:1 priori:1 constrained:5 initialize:2 construct:1 never:1 once:1 look:1 future:1 others:1 report:2 randomly:2 delayed:1 tke:1 attempt:2 neuronlike:1 englewood:1 investigate:1 evaluation:1 light:1 behind:1 tj:1 hyt:9 necessary:1 theoretical:1 increased:1 assignment:1 too:1 optimally:1 stored:1 international:1 amherst:2 discriminating:1 ie:1 invertible:1 thesis:2 cognitive:1 resort:1 return:2 li:1 converted:2 baird:2 vi:1 performed:2 yv:1 parallel:2 air:1 ofthe:1 produced:1 cc:1 cybernetics:1 za:6 whenever:1 resultant:1 proof:4 recovers:2 adjusting:2 massachusetts:2 ea:1 back:3 appears:1 specify:1 iai:1 execute:2 done:1 anderson:2 xa:2 receives:1 scientific:1 mdp:4 verify:1 moore:1 ll:2 game:1 samuel:3 criterion:1 hill:1 outline:1 complete:2 performs:1 bradtke:1 temperature:2 novel:1 recently:1 behaves:4 exponentially:1 he:2 significant:1 mellon:1 cambridge:2 feldman:1 ai:3 similarly:1 robot:1 actor:49 store:3 certain:1 success:1 vt:4 accomplished:2 additional:1 employed:1 converting:2 determine:2 converge:1 multiple:1 technical:2 match:1 lin:3 ase:1 va:2 variant:1 expectation:1 iteration:5 normalization:1 achieved:1 addition:1 decreased:1 annealing:1 diagram:2 appropriately:1 checker:2 jordan:1 call:4 enough:1 variety:1 zi:8 architecture:5 restrict:1 idea:1 whether:1 motivated:2 york:1 action:46 gabriel:1 useful:1 detailed:1 redefines:1 amount:1 discount:2 exist:1 zj:1 estimated:2 neuroscience:1 discrete:1 carnegie:1 key:1 thereafter:1 four:1 pj:1 lowering:1 sum:3 convert:1 run:1 decision:5 bound:1 guaranteed:1 ahead:1 hy:3 performing:2 relatively:2 department:2 developing:1 according:2 modification:1 making:1 gradually:1 computationally:1 mechanism:1 dyna:1 hh:1 needed:1 letting:1 available:2 operation:1 permit:1 indirectly:1 generic:3 alternative:1 slower:1 remaining:1 include:2 npk:1 approximating:1 classical:1 sweep:1 objective:1 move:1 already:1 question:2 strategy:6 rt:1 usual:1 traditional:1 dp:7 thank:1 simulated:1 reason:1 toward:1 assuming:1 ru:1 relationship:2 ratio:1 difficult:1 executed:6 robert:5 negative:1 pmax:4 implementation:2 zt:7 policy:22 boltzmann:2 allowing:1 markov:3 finite:4 immediate:1 precise:1 interacting:1 pair:4 learned:1 uncoupled:1 nu:1 usually:1 below:3 program:1 including:1 max:3 force:1 nth:1 mdps:1 literature:1 acknowledgement:1 understanding:1 determining:2 relative:2 fully:1 proven:1 lv:2 foundation:2 agent:4 verification:1 editor:2 critic:48 pi:3 supported:1 asynchronous:3 infeasible:1 tsitsiklis:5 distributed:2 f49620:1 feedback:3 axi:2 transition:1 rich:1 collection:1 reinforcement:2 adaptive:3 made:1 qx:1 transaction:1 qlearning:2 mcgraw:1 pittsburgh:1 xi:6 continuous:1 iterative:1 crites:6 ny:1 explicit:1 watkins:8 late:1 learns:1 northeastern:1 theorem:2 jt:3 h1j:1 exists:1 adding:2 sequential:1 phd:2 illustrates:1 commutative:1 horizon:1 vijay:1 simply:1 infinitely:1 desire:2 corresponds:1 ma:5 leaming:1 man:1 change:2 infinite:1 determined:1 accepted:1 ew:2 select:1 college:1
7,154
917
Temporal Dynamics of Generalization Neural Networks Changfeng Wang Department of Systems Engineering University Of Pennsylvania Philadelphia, PA 19104 fwang~ender.ee.upenn.edu ? In Santosh S. Venkatesh Department of Electrical Engineering University Of Pennsylvania Philadelphia, PA 19104 venkateshGee.upenn.edu Abstract This paper presents a rigorous characterization of how a general nonlinear learning machine generalizes during the training process when it is trained on a random sample using a gradient descent algorithm based on reduction of training error . It is shown, in particular, that best generalization performance occurs, in general, before the global minimum of the training error is achieved. The different roles played by the complexity of the machine class and the complexity of the specific machine in the class during learning are also precisely demarcated . 1 INTRODUCTION In learning machines such as neural networks, two major factors that affect the 'goodness of fit' of the examples are network size (complexity) and training time. These are also the major factors that affect the generalization performance of the network. Many theoretical studies exploring the relation between generalization performance and machine complexity support the parsimony heuristics suggested by Occam's razor, to wit that amongst machines with similar training performance one should opt for the machine of least complexity. Multitudinous numerical experiments (cf. [5]) suggest, however , that machines of larger size than strictly necessary to explain the data can yield generalization performance similar to that of smaller machines (with 264 Changieng Wang. Santosh S. Venkatesh similar empirical error) if learning is optimally stopped at a critical point before the global minimum of the training error is achieved . These results seem to fly in contradiction with a learning theoretic interpretation of Occam's razor. In this paper, we ask the following question: How does the gradual reduction of training error affect the generalization error when a machine of given complexity is trained on a finite number of examples? Namely, we study the simultaneous effects of machine size and training time on the generalization error when a finite sample of examples is available. Our major result is a rigorous characterization of how a given learning machine generalizes during the training process when it is trained using a learning algorithm based on minimization of the empirical error (or a modification of the empirical error). In particular, we are enabled to analytically determine conditions for the existence of a finite optimal stopping time in learning for achieving optimal generalization. We interpret the results in terms of a time-dependent, effective machine size which forms the link between generalization error and machine complexity during learning viewed as an evolving process in time. Our major results are obtained by introducing new theoretical tools which allow us to obtain finer results than would otherwise be possible by direct applications of the uniform strong laws pioneered by Vapnik and Cervonenkis (henceforth refered to as VC-theory). The different roles played by the complexity of the machine class and the complexity of the specific machine in the class during learning are also precisely demarcated in our results. Since the generalization error is defined in terms of an abstract loss function, the results find wide applicability including but not limited to regression (square-error loss function) and density estimation (log-likelihood loss) problems. 2 THE LEARNING PROBLEM We consider the problem of learning from examples a relation between two vectors x and y determined by a fixed but unknown probability distribution P(x, y). This model includes, in particular, the input-output relation described by y = g(x, e) , (1) e, where g is some unknown function of x and which are random vectors on the same probability space. The vector x can be viewed as the input to an unknown system, a random noise term (possibly dependent on x) , and y the system's output. e The hypothesis class from which the learning procedure selects a candidate function (hypothesis) approximating g is a parametric family of functions 1{d = {!(x, fJ) : fJ E 8d ~ ?R d } indexed by a subset 8d of d-dimensional Euclidean space. For example, if x E ?R m and y is a scalar, 1{d can be the class of functions computed by a feedforward neural network with one hidden layer comprised of h neurons and activation function 1/;, viz ., Temporal Dynamics of Generalization in Neural Networks In the above, d 265 = (m + 2)h denotes the number of adjustable parameters. The goal of learning within the hypothesis class 1ld is to find the best approximation of the relation between x and y in 1ld from a finite set of n examples 'Dn = {(Xl , yd, ? .. , (xn , Yn)} drawn by independent sampling from the distribution P(x , y) . A learning algorithm is simply a map which, for every sample 'Dn (n ~ 1) , produces a hypothesis in 1ld. In practical learning situations one first selects a network of fixed structure (a fixed hypothesis class Jid), and then determines the "best" weight vector e* (or equivalently, the best function I(x, e*) in this class) using some training algorithm. The proximity of an approximation I( x, e) to the target function g( x, e) at each point X is measured by a loss function q : (J(x,e),g(x,e) I--t ~+. For a given hypothesis class, the function 1(?, e) is completely determined by the parameter vector e. With g fixed, the loss function may be written, with a slight abuse of notation, as a map q( x, y, e) from ~m x ~ x 8 into ~+ . Examples of the forms of loss (g(x,e) - f(x,e))2 functions are the familiar square-law loss function q(x,y,e) commonly used in regression and learning in neural networks, and the KulbackLeibler distance (or relative entropy) q( x, y, e) = In PP y f9 x, for density estimation, where p(Y I i(x , e?) denotes the conditional density function of y given i(x, e). = t t: ) The closeness of 1(-, e) to g(.) is measured by the expected (ensemble) loss or error ?(e, d) ~ J q(x, y, e)p(dx, dy) . Ie, = The optimal approximation e*) is such that ?(e*, d) min9EE>d ?(e, d). In similar fashion, we define the corresponding empirical loss (or training error) by 1 n ?n(e, d) = J q(x, y, e)Pn(dx, dy) = ~ ?= q(Xi, Yi , e) . ? =1 where Pn denotes the joint empirical distribution of input-output pairs (x, y). The global minimum of the empirical error over 8d is denoted by 0, namely, arg min9EE> ?n(e, d). An iterative algorithm for minimizing ?n(e, d) (or a modification of it) over 8 d generates at each epoch t a random vector et :'D n -+ 8 d . The quantity ?(e t , d) = E f q(x, et)p(dx, de) is referred to as the generalization error of et . We are interested in the properties of the process { et : t 1, 2, ... }, and the time-evolution of the sequence {?(et, d) : t 1,2, .. . }. e= = = Note that each et is a functional of Pn . When P = Pn , learning reduces to an optimization problem. Deviations from optimality arise intrinsically as a consequence of the discrepancy between Pn and P. The central idea of this work is to analyze the consequence of the deviation .6. n ~ Pn - P on the generalization error. To simplify notation , we henceforth suppress d and write simply 8 , ?(e) and ?n(e) instead of 8 d , ?(e, d), and ?n(e, d), respectively. 2.1 RegUlarity Conditions We will be interested in the local behavior of learning algorithms. Consequently, we assume that 8 is a compact set, and e* is the unique global minimum of ?(e) on 8. 266 Changfeng Wang, Santosh S. Venkatesh It can be argued that these assumptions are an idealization of one of the following situations: ? A global algorithm is used which is able to find the global minimum of En (9), and we are interested in the stage of training when 9t has entered a region 8 where 9* is the only global minimum of E(9); ? A local algorithm is used, and the algorithm has entered a region 8 which contains 9* as the unique global minimum of E(9) or as a unique local minimum with which we are content. In the sequel, we write () / {)9 to denote the gradient operator with respect to the vector 9, and likewise write ()2/{)02 to denote the matrix of operators L:l{/~;{/J:,j=l. In the rest of the development we assume the following regularity conditions: Al. The loss function q(x, y,.) is twice continuously differential for all 0 E 8 and for almost all (x, Y)j A2. P(x, y) has compact support; A3. The optimal network 9* is an interior point of 8; A4. The matrix ~(9*) = ~E(O*) is nonsingular. These assumptions are typically satisfied in neural network applications. We will also assume that the learning algorithm converges to the global minimum of En (0) over 8 (note that 8 may not be a true global minimum, so the assumption applies to gradient descent algorithms which converge locally). It is easy to demonstrate that for each such algorithm, there exists an algorithm which decreases the empirical error monotonically at each step of iteration. Thus, without loss of generality, we also assume that all the algorithms we consider have this monotonicity property. 3 3.1 GENERALIZATION DYNAMICS First Phase of Learning The quality of learning based on the minimization of the empirical error depends on the value of the quantity sUPe IEn(9) - E(9)1. Under the above assumptions, it is shown in [3] that E(O) = En((}) + Op ( Inn) ..;n and n) E(OA) = E(9*) + 0 (In -;;- . Therefore, for any iterative algorithm for minimizing En(O), in the initial phase of learning the reduction of training error is essentially equivalent to the reduction of generalization error. It can be further shown that this situation persists until the estimates ()t enter an n- 6,. neighborhood of 8, where bn - t 1/2. The basic tool we have used in arriving at this conclusion is the VC-method. The characterization of the precise generalization properties of the machine after Ot enters an n- 6 neighborhood of the limiting solution needs a more precise language than can be provided by the VC-method, and is the main content of the rest of this work. Temporal Dynamics of Generalization in Neural Networks 3.2 267 Learning by Gradient Descent In the following, we focus on generalization properties when the machine is trained using the gradient descent algorithm (Backpropagation is a Gauss-Seidel implementation of this algorithm); in particular, the adaptation is governed by the recurrence (t ~ 0), (2) where the positive quantity ( governs the rate of learning. Learning and generalization properties for other algorithms can be studied using similar techniques. Replace fn by f in (2) and let { (); , t ~ O} denote the generated sequence of vectors. We can show (though we will not do so here) that the weight vector ()t is asymptotically normally distributed with expectation (); and covariance matrix with all It is precisely the deviation of ()t from (); caused by the entries of order 0 ( perturbation of amount ~n = Pn - P to the true distribution P which results in interesting artifacts such as a finite optimal stopping time when the number of examples is finite. k). 3.3 The Main Equation of Generalization Dynamics Under the regularity conditions mentioned in the last section, we can find the generalization error at each epoch of learning as an explicit function of the number of iterations, machine parameters, and the initial error. Denote by Al 2: A2 2: ... ~ Ad the eigenvalues of the matrix <1>( ()*) and suppose T is the orthogonal diagonalizing matrix for <1>(()*), viz., T'?(()*)T = diag(Al, ... ,Ad). Set 8 = (8 l , ... ,8d)' ~ T( ()o - ()*) and for each i let Vi denote the ith diagonal element of the d x d matrix T'E { (:0 q( x, ()*?) (:0 q( x, ()*?) '} T. Also let S( (), p) denote the open ball of radius p at (). Under Assumptions Al-A4, the generalization error of the machine trained according to (2) is governed by the following equation for all starting points ()o E S(()*, n- r ) (0 < r ~ t), and uniformly for all t 2: 0: MAIN THEOREM d f(()t) = f(()*) + -.!.. ,,{ Vi [1 - (1- (Ai)t] 2 + 8; Ai(l- (Ai)2t} + O(n- 3r ) . 2n L...J A' i=l (3) ' If ()o (/:. S(()*, n-t), then the generalization dynamics is governed by the following equation valid for all r > 0: where tl is the smallest t such that E 1[1 - (<1>]t11 = An- r for some A > 0, and C(tl)' Ci(tl) ~ 0 are constants depending on network parameters and tlIn the special case when the data is generated by the following additive noise model y=g(x)+e, (5) 268 Changfeng Wang. Santosh S. Venkatesh with E [~Ix] = 0, and E [elx] = 0- 2 = constant, if g(x) = f(x,O*) and the loss function q(x, 0) is given by the square-error loss function, the above equation reduces to the following form: 2 ?(Ot) = ?(0*) + ;n d ?= {[1 - (1- f.~i)t] 2 + 61.Ai(1- f.Ai)2t} + O(n- 3r ). ,=1 In particular, if f(x , 0) is linear in 0, we obtain our previous result [4] for linear machines. The result (3) is hence a substantive extension of the earlier result to very general settings. It is noted that the extension goes beyond nonlinearity and the original additive noise data generating model-we no longer require that the 'true' model be contained in the hypothesis class. 3.4 Effective Complexity Write Ci ~ to be Vi /.Ai . The effective complexity of the nonlinear machine Ot at t is defined 2 d C(O*, d, t) ~ 2: ci (1 - (1- f.Ad) . i=1 Analysis shows that the term Ci indicates the level of sensitivity of output of the machine to the ith component of the normalized weight vector, 0; C(O*, d, t) denotes the degree to which the approximation power of the machine is invoked by the learning process at epoch t. Indeed, as t ---+ 00, C(O*, d, t) ---+ Cd = '2:.1=1 Ci, which is the complexity of the limiting machine {} which represents the maximal fitting of examples to the machine (i.e., minimized training error). For the additive noise data generating model (5) and square-error loss function, the effective complexity becomes, d C(O*, d, t) = 2:(1- (1- cAif)2 . i=1 The sum can be interpreted as the effective number of parameters used at epoch t. At the end of training, it becomes exactly the number of parameters of the machine . Now write 0; ~ 0* + (0 0 - 0* )(1- f.Ad. With these definitions, (3) can be rewritten to give the following approximation error and complexity error decomposition of generalization error in the learning process: (t ~ 0). (6) The first term on the right-hand-side, ?(0;), denotes the approximation error at epoch t and is the error incurred in using as an approximation of the 'truth.' Note that the approximation error depends on time t and the initial value 00 , but not the examples. Clearly, it is the error one would obtain at epoch t in minimizing the function ?(0) (as opposed to ?n(O? using the same learning algorithm and starting with the same step length f and initial value 00 . The second term on the righthand-side is the complexity error at epoch t. This is the part of the generalization error at t due to the substitution of ?n(O) for ?(0). 0; Temporal Dynamics of Generalization in Neural Networks 269 The overfitting phenomena in learning is often intuitively attributed to the 'fitting of noise.' We see that is only partly correct: it is in fact due to the increasing use of the capacity of the machine, that the complexity penalty becomes increasingly large, this being true even when the data is clean, i.e., when == O! Therefore, we see that (6) gives an exact trade-off of the approximation error and complexity error in the learning process. e For the case of large initial error, we see from the main theorem that the complexity error is essentially the same as that at the end of training , when the initial error is reduced to about the same order as before. The reduction of the training error leads to monotone decrease in generalization error in this case. 3.5 Optimal Stopping Time We can phrase the following succinct open problem in learning in neural networks: When should learning be ideally stopped? The question was answered for linear machines which is a special form of neural networks in [4] . This section extends the result to general nonlinear machines (including neural networks) in regular cases. For this purpose, we write the generalization error in the following form: where <p(t) ~ : 2 d ?= {li(l- dd 2t - dj (1- fAdt} , s=l and dj and Ii are machine parameters. The time-evolution of generalization error during the learning process is completely determined by the function <p(t). Define tmin E { T ~ 0 : ?(OT) ~ ?(Ot) for all t ~ O}, that is tmin denotes an epoch at which the generalization error is minimized. The smallest such number will be referred to as the optimal stopping time of learning. In general we have Cj > a for all i . In this case, it is possible to determine that there is a finite optimal stopping time. More specifically, there exists two constants t/ and tu which depend on the machine parameters such that t/ ~ tmin ~ tu. Furthermore, it can be shown that the function <p(t) decreases monotonically for t ~ t/ and increases monotonically for all t ~ tu. Finally, we can relate the generalization performance when learning is optimally stopped to the best achievable performance by means of the following inequality: ?(0 . ) < ?(0.) tmm - + (1 - IC)Cd 2n ' where IC = O(nO) is a constant depending on Ii and dj's, and is in the interval (O,~], and Cd denotes, as before, the limiting value of the effective machine complexity C(O?, d, t) as t -+ 00. = In the pathological case where there exists i such that Cj 0, there may not exist a finite optimal stopping time. However, even in such cases, it can be shown that if In(l - fAdl In(l - fAd) < 2, a finite optimal stopping time still exists. 270 4 Changfeng Wang, Santosh S. Venkatesh CONCLUDING REMARKS This paper describes some major results of our recent work on a rigorous characterization of the generalization process in neural network types of learning machines. In particular, we have shown that reduction of training error may not lead to improved generalization performance. Two major techniques involved are the uniform weak law (VC-theory) and differentiable statistical functionals, with the former delivering an initial estimate, and the latter giving finer results. The results shows that the complexity (e.g. VC-dimension) of a machine class does not suffice to describe the role of machine complexity in generalization during the learning process; the appropriate complexity notion required is a time-varying and algorithm-dependent concept of effective machine complexity. Since results in this work contain parameters which are typically unknown, they cannot be used directly in practical situations. However, it is possible to frame criteria overcoming such difficulties. More details of the work described here and its extensions and applications can be found in [3]. The methodology adopted here is also readily adapted to study the dynamical effect of regularization on the learning process [3]. Acknowledgements This research was supported in part by the Air Force Office of Scientific Research under grant F49620-93-1-0120. References [1] Kolmogorov, A. and V. Tihomirov (1961). f-entropy and f-capacity of sets in functional spaces. Amer. Math. Soc. Trans. (Ser. 2), 17:277-364. [2] Vapnik, V. (1982). Estimation of Dependences Based on Empirical Data. Springer-Verlag, New York. [3] Wang, C. (1994). A Theory of Generalization in Learning Machines. Ph. D. Thesis, University of Pennsylvania. [4] Wang, C., S. S. Venkatesh, and J. S. Judd (1993) . Optimal stopping and effective machine size in learning. Proceedings of NIPS'93. [5] Weigend, A. (1993). On overtraining and the effective number of hidden units. Proceedings of the 1993 Connectionist Models Summer School. 335-342. Ed. Mozer, M. C. et al. Hillsdale , NJ: Erlbaum Associates.
917 |@word achievable:1 open:2 gradual:1 bn:1 covariance:1 decomposition:1 ld:3 reduction:6 initial:7 substitution:1 contains:1 demarcated:2 activation:1 dx:3 written:1 readily:1 fn:1 additive:3 numerical:1 ith:2 characterization:4 math:1 dn:2 direct:1 differential:1 fitting:2 upenn:2 indeed:1 expected:1 behavior:1 increasing:1 becomes:3 provided:1 notation:2 suffice:1 interpreted:1 parsimony:1 nj:1 temporal:4 every:1 exactly:1 ser:1 normally:1 grant:1 unit:1 yn:1 before:4 positive:1 engineering:2 local:3 persists:1 consequence:2 yd:1 abuse:1 twice:1 studied:1 limited:1 practical:2 unique:3 backpropagation:1 procedure:1 empirical:9 evolving:1 regular:1 suggest:1 cannot:1 interior:1 operator:2 equivalent:1 map:2 go:1 starting:2 wit:1 contradiction:1 enabled:1 notion:1 limiting:3 target:1 suppose:1 pioneered:1 exact:1 hypothesis:7 pa:2 element:1 associate:1 ien:1 role:3 fly:1 wang:7 electrical:1 enters:1 region:2 decrease:3 trade:1 mentioned:1 mozer:1 complexity:23 ideally:1 dynamic:7 trained:5 depend:1 completely:2 joint:1 kolmogorov:1 effective:9 describe:1 neighborhood:2 heuristic:1 larger:1 otherwise:1 supe:1 sequence:2 eigenvalue:1 differentiable:1 inn:1 maximal:1 adaptation:1 tu:3 entered:2 regularity:3 produce:1 generating:2 converges:1 depending:2 measured:2 school:1 op:1 strong:1 soc:1 radius:1 correct:1 vc:5 hillsdale:1 argued:1 require:1 generalization:35 opt:1 exploring:1 strictly:1 extension:3 tmm:1 proximity:1 ic:2 major:6 a2:2 smallest:2 purpose:1 estimation:3 tool:2 minimization:2 clearly:1 pn:7 varying:1 office:1 focus:1 viz:2 elx:1 likelihood:1 indicates:1 rigorous:3 dependent:3 stopping:8 typically:2 hidden:2 relation:4 selects:2 interested:3 arg:1 denoted:1 development:1 special:2 santosh:5 sampling:1 represents:1 discrepancy:1 minimized:2 connectionist:1 simplify:1 pathological:1 familiar:1 phase:2 righthand:1 necessary:1 orthogonal:1 indexed:1 euclidean:1 theoretical:2 stopped:3 earlier:1 goodness:1 phrase:1 applicability:1 introducing:1 deviation:3 subset:1 entry:1 uniform:2 comprised:1 erlbaum:1 optimally:2 tihomirov:1 density:3 sensitivity:1 ie:1 sequel:1 off:1 continuously:1 thesis:1 central:1 satisfied:1 opposed:1 possibly:1 henceforth:2 li:1 de:1 includes:1 caused:1 depends:2 ad:4 vi:3 analyze:1 square:4 air:1 likewise:1 ensemble:1 yield:1 nonsingular:1 weak:1 finer:2 overtraining:1 explain:1 simultaneous:1 ed:1 definition:1 pp:1 involved:1 attributed:1 intrinsically:1 ask:1 cj:2 ender:1 methodology:1 improved:1 amer:1 though:1 generality:1 furthermore:1 stage:1 until:1 hand:1 nonlinear:3 artifact:1 quality:1 scientific:1 effect:2 normalized:1 true:4 concept:1 contain:1 evolution:2 analytically:1 hence:1 former:1 regularization:1 during:7 recurrence:1 razor:2 noted:1 criterion:1 theoretic:1 demonstrate:1 fj:2 invoked:1 functional:2 interpretation:1 slight:1 interpret:1 enter:1 ai:6 nonlinearity:1 language:1 dj:3 longer:1 recent:1 verlag:1 inequality:1 yi:1 minimum:10 determine:2 converge:1 monotonically:3 ii:2 reduces:2 seidel:1 regression:2 changfeng:4 basic:1 essentially:2 expectation:1 jid:1 iteration:2 achieved:2 interval:1 ot:5 rest:2 seem:1 ee:1 feedforward:1 easy:1 affect:3 fit:1 pennsylvania:3 idea:1 penalty:1 york:1 remark:1 governs:1 delivering:1 amount:1 locally:1 ph:1 reduced:1 exist:1 write:6 achieving:1 drawn:1 clean:1 asymptotically:1 monotone:1 idealization:1 sum:1 weigend:1 extends:1 family:1 almost:1 dy:2 layer:1 summer:1 played:2 adapted:1 precisely:3 generates:1 answered:1 optimality:1 concluding:1 department:2 according:1 ball:1 smaller:1 describes:1 increasingly:1 modification:2 refered:1 intuitively:1 equation:4 end:2 adopted:1 generalizes:2 available:1 rewritten:1 appropriate:1 fad:1 existence:1 original:1 denotes:7 cf:1 t11:1 a4:2 giving:1 approximating:1 question:2 quantity:3 occurs:1 parametric:1 dependence:1 diagonal:1 gradient:5 amongst:1 distance:1 link:1 oa:1 capacity:2 substantive:1 length:1 minimizing:3 equivalently:1 relate:1 suppress:1 implementation:1 unknown:4 adjustable:1 neuron:1 finite:9 descent:4 situation:4 tmin:3 precise:2 frame:1 perturbation:1 overcoming:1 venkatesh:6 namely:2 pair:1 required:1 nip:1 trans:1 able:1 suggested:1 beyond:1 dynamical:1 including:2 power:1 critical:1 difficulty:1 force:1 diagonalizing:1 philadelphia:2 epoch:8 acknowledgement:1 relative:1 law:3 loss:14 interesting:1 incurred:1 degree:1 dd:1 occam:2 cd:3 supported:1 last:1 arriving:1 side:2 allow:1 wide:1 distributed:1 f49620:1 dimension:1 xn:1 valid:1 judd:1 commonly:1 functionals:1 compact:2 monotonicity:1 global:10 overfitting:1 xi:1 iterative:2 diag:1 main:4 noise:5 arise:1 succinct:1 referred:2 f9:1 en:4 tl:3 fashion:1 explicit:1 xl:1 candidate:1 governed:3 ix:1 theorem:2 specific:2 closeness:1 a3:1 exists:4 vapnik:2 ci:5 entropy:2 simply:2 contained:1 scalar:1 applies:1 springer:1 truth:1 determines:1 conditional:1 viewed:2 goal:1 consequently:1 replace:1 content:2 determined:3 specifically:1 uniformly:1 partly:1 gauss:1 support:2 latter:1 phenomenon:1
7,155
918
Limits on Learning Machine Accuracy Imposed by Data Quality Corinna Cortes, L. D. Jackel, and Wan-Ping Chiang AT&T Bell Laboratories Holmdel, NJ 07733 Abstract Random errors and insufficiencies in databases limit the performance of any classifier trained from and applied to the database. In this paper we propose a method to estimate the limiting performance of classifiers imposed by the database. We demonstrate this technique on the task of predicting failure in telecommunication paths. 1 Introduction Data collection for a classification or regression task is prone to random errors, e.g. inaccuracies in the measurements of the input or mis-labeling of the output. Missing or insufficient data are other sources that may complicate a learning task and hinder accurate performance of the trained machine. These insufficiencies of the data limit the performance of any learning machine or other statistical tool constructed from and applied to the data collection - no matter how complex the machine or how much data is used to train it. In this paper we propose a method for estimating the limiting performance of learning machines imposed by the quality of the database used for the task. The method involves a series of learning experiments. The extracted result is, however, independent of the choice of learning machine used for these experiments since the estimated limiting performance expresses a characteristic of the data. The only requirements on the learning machines are that their capacity (VC-dimension) can be varied and can be made large, and that the learning machines with increasing capacity become capable of implementing any function. 240 Corinna Cortes. L. D. Jackel. Wan-Ping Chiang We have applied the technique to data collected for the purpose of predicting failures in telecommunication channels of the AT&T network. We extracted information from one of AT&T's large databases that continuously logs performance parameters of the network. The character and amount of data comes to more material than humans can survey. The processing of the extracted information is therefore automated by learning machines. We conjecture that the quality of the data imposes a limiting error rate on any learning machine of,... 25%, so that even with an unlimited amount of data, and an arbitrarily complex learning machine, the performance for this task will not exceed ,... 75% correct. This conjecture is supported by experiments. The relatively high noise-level of the data, which carries over to a poor performance of the trained classifier, is typical for many applications: the data collection was not designed for the task at hand and proved inadequate for constructing high performance classifiers. 2 Basic Concepts of Machine Learning We can picture a learning machine as a device that takes an unknown input vector and produces an output value. More formally, it performs some mapping from an input space to an output space. The particular mapping it implements depends of the setting of the internal parameters of the learning machine. These parameters are adjusted during a learning phase so that the labels produced on the training set match, as well as possible, the labels provided. The number of patterns that the machine can match is loosely called the "capacity" of the machine. Generally, the capacity of a machine increases with the number of free parameters. After training is complete, the generalization ability of of the machine is estimated by its performance on a test set which the machine has never seen before. The test and training error depend on both the the number of training examples I, the capacity h of the machine, and, of course, how well suited the machine is to implement the task at hand. Let us first discuss the typical behavior of the test and training error for a noise corrupted task as we vary h but keep the amount I of training data fixed. This scenario can, e.g., be obtained by increasing the number of hidden units in a neural network or increasing the number of codebook vectors in a Learning Vector Quantization algorithm [6]. Figure la) shows typical training and test error as a function of the capacity of the learning machine. For h < I we have many fewer free parameters than training examples and the machine is over constrained. It does not have enough complexity to model the regularities of the training data, so both the training and test error are large (underfitting). As we increase h the machine can begin to fit the general trends in the data which carries over to the test set, so both error measures decline. Because the performance of the machine is optimized on only part of the full pattern space the test error will always be larger than the training error. As we continue to increase the capacity of the learning machine the error on the training set continues to decline, and eventually it reaches zero as we get enough free parameters to completely model the training set. The behavior of the error on the test set is different. Initially it decreases, but at some capacity, h*, it starts to rise. The rise occurs because the now ample resources of the training machine are applied to learning vagaries of the training Limits on Learnillg Machine Accuracy Imposed by Data Quality % error a) % error b) 241 c) % error ? . ?. ...,..- --- ~ Ie " capacity ( h) -- training .. training set size ( I) .. . intrinsic noise level capacity ( h) Figure 1: Errors as function of capacity and training set size. Figure la) shows characteristic plots of training and test error as a function of the learning machine capacity for fixed training set size. The test error reaches a minimum at h = h* while the training error decreases as h increases. Figure Ib) shows the training and test errors at fixed h for varying I. The dotted line marks the asymptotic error Eoo for infinite I. Figure lc) shows the asymptotic error as a function of h. This error is limited from below by the intrinsic noise in the data. set, which are not reproduced in the test set (overfitting). Notice how in Figure la) the optimal test error is achieved at a capacity h* that is smaller than the capacity for which zero error is achieved on the training set. The learning machine with capacity h* will typically commit errors on misclassified or outlying patterns of the training set. We can alternatively discuss the error on the test and training set as a function of the training set size I for fixed capacity h of the learning machine. Typical behavior is sketched in Figure Ib) . For small I we have enough free parameters to completely model the training set, so the training error is zero. Excess capacity is used by the learning machine to model details in the training set, leading to a large test error. As we increase the training set size I we train on more and more patterns so the test error declines. For some critical size of the training set, Ie, the machine can no longer model all the training patterns and the training error starts to rise. As we further increase I the irregularities of the individual training patterns smooth out and the parameters of the learning machine is more and more used to model the true underlying function . The test error declines, and asymptotically the training and test error reach the same error value Eoo . This error value is the limiting performance of the given learning machine to the task . In practice we never have the infinite amount of training data needed to achieve Eoo. However, recent theoretical calculations [8, 1, 2, 7, 5] and experimental results [3] have shown that we can estimate Eoo by averaging the training and test errors for I> Ie. This means we can predict the optimal performance of a given machine. For a given type of learning machine the value of the asymptotic error Eoo of the machine depends on the quality of the data and the set of functions it can implement. The set of available functions increases with the capacity of the machine: 242 Corinna Cortes, L. D. Jackel, Wall-Ping Chiang low capacity machines will typically exhibit a high asymptotic error due to a big difference between the true noise-free function of the patterns and the function implemented by the learning machine, but as we increase h this difference decreases. If the learning machine with increasing h becomes a universal machine capable of modeling any function the difference eventually reaches zero, so the asymptotic error Eoo only measures the intrinsic noise level of the data. Once a capacity of the machine has been reached that matches the complexity of the true function no further improvement in Eoo can be achieved. This is illustrated in Figure lc). The intrinsic noise level of the data or the limiting performance of any learning machine may hence be estimated as the asymptotic value of Eoo as obtained for asymptotically universal learning machines with increasing capacity applied to the task. This technique will be illustrated in the following section. 3 Experimental Results In this section we estimate the limiting performance imposed by the data of any learning machine applied to the particular prediction task. 3.1 Task Description To ensure the highest possible quality of service, the performance parameters of the AT&T network are constantly monitored. Due to the high complexity of the network this performance surveillance is mainly corrective: when certain measures exceed preset thresholds action is taken to maintain reliable, high quality service. These reorganizations can lead to short, minor impairments of the quality of the communication path. In contrast, the work reported here is preventive: our objective is to make use of the performance parameters to form predictions that are sufficiently accurate that preemptive repairs of the channels can be made during periods of low traffic. In our study we have examined the characteristics of long-distance, 45 Mbitsfs communication paths in the domestic AT&T network. The paths are specified from one city to another and may include different kinds of physical links to complete the paths. A path from New York City to Los Angeles might include both optical fiber and coaxial cable. To maintain high-quality service, particular links in a path may be switched out and replaced by other, redundant links. There are two primary ways in which performance degradation is manifested in the path. First is the simple bit-error rate, the fraction of transmitted bits that are not correctly received at the termination of the path. Barring catastrophic failure (like a cable being cut), this error rate can be measured by examining the error-checking bits that are transmitted along with the data. The second instance of degradation, ''framing error" , is the failure of synchronization between the transmitter and receiver in a path. A framing error implies a high count of errored bits. In order to better characterize the distribution of bit errors, several measures are historically used to quantify the path performance in a 15 minutes interval. These measures are: Low-Rate The number of seconds with exactly 1 error. Limits Oil 243 Learning Machine Accllracy Imposed by Data Quality "No-Trouble" patterns: Frame-Error Rate-High I ? ? .. Rate-Medlum IJ Rate-Low I , o "Trouble" pattems: In time (days) 21 II I ,,' , I l"e Figure 2: Errors as function of time. The 3 top patterns are members of the "No-Trouble" class. The 3 bottom ones are members of the "Trouble" class. Errors are here plotted as mean values over hours. Medium-Rate The number of seconds with more than one but less than 45 errors. High-Rate The number of seconds with 45 or more errors, corresponding to a bit error rate of at least 10- 6 ? Frame-Error The number of seconds with a framing error. A second with a frameerror is always accompanied by a second of High-Rate error. Although the number of seconds with the errors described above in principle could be as high as 900, any value greater than 255 is automatically clipped back to 255 so that each error measure value can be stored in 8 bits. Daily data that include these measures are continuously logged in an AT&T database that we call Perf(ormance)Mon(itor). Since a channel is error free most of the time, an entry in the database is only made if its error measures for a 15 minute period exceed fixed low thresholds, e.g. 4 Low-Rate seconds, 1 Medium- or HighRate second, or 1 Frame-Error. In our research we "mined" PerfMon to formulate a prediction strategy. We extracted examples of path histories 28 days long where the path at day 21 had at least 1 entry in the PerfMon database. We labeled the examples according to the error-measures over the next 7 days. If the channel exhibited a 15-minute period with at least 5 High-Rate seconds we labeled it as belonging to the class "Trouble". Otherwise we labeled it as member of "No-Trouble" . The length of the history- and future-windows are set somewhat arbitrarily. The history has to be long enough to capture the state of the path but short enough that our learning machine will run in a reasonable time. Also the longer the history the more likely the physical implementation of the path was modified so the error measures correspond to different media. Such error histories could in principle be eliminated from the extracted examples using the record of the repairs and changes 244 Corinna Cortes, L. D. lackel, Wan-Ping Chiang of the network. The complexity of this database, however, hinders this filtering of examples. The future-window of7 days was set as a design criterion by the network system engineers. Examples of histories drawn from PerfMon are shown in Figure 2. Each group of traces in the figure includes plots of the 4 error measures previously described. The 3 groups at the top are examples that resulted in No-Trouble while the examples at the bottom resulted in Trouble. Notice how bursty and irregular the errors are, and how the overall level of Frame- and High-Rate errors for the Trouble class seems only slightly higher than for the No-Trouble class, indicating the difficulty of the classification task as defined from the database PerfMon. PerfMon constitutes, however, the only stored information about the state of a given channel in its entirety and thus all the knowledge on which one can base channel end-to-end predictions: it is impossible to install extra monitoring equipment to provide other than the 4 mentioned end-to-end error measures. The above criteria for constructing examples and labels for 3 months of PerfMon data resulted in 16325 examples from about 900 different paths with 33.2% of the examples in the class Trouble. This means, that always guessing the label of the largest class, No-Trouble, would produce an error rate of about 33%. 3.2 Estimating Limiting Performance The 16325 path examples were randomly divided into a training set of 14512 examples and a test set of 1813 examples. Care was taken to ensure that a path only contributes to one of the sets so the two sets were independent, and that the two sets had similar statistical properties. Our input data has a time-resolution of 15 minutes. For the results reported here the 4 error measures of the patterns were subsampled to mean values over days yielding an input dimensionality of 4 x 21. We performed two sets of independent experiments. In one experiment we used fully connected neural networks with one layer of hidden units. In the other we used LVQ learning machines with an increasing number of codebook vectors. Both choices of machine have two advantages: the capacity of the machine can easily be increased by adding more hidden units, and by increasing the number of hidden units or number of codebook vectors we can eventually model any mapping [4]. We first discuss the results with neural networks. Baseline performance was obtained from a threshold classifier by averaging all the input signals and thresholding the result. The training data was used to adjust the single threshold parameter. With this classifier we obtained 32% error on the training set and 33% error on the test set. The small difference between the two error measures indicate statistically induced differences in the difficulty of the training and test sets. An analysis of the errors committed revealed that the performance of this classifier is almost identical to always guessing the label of the largest class "No-Trouble": close to 100% of the errors are false negative. A linear classifier with about 200 weights (the network has two output units) obtained 28% error on the training set and 32% error on the test set. Limits on Learning Machine Accuracy Imposed by Data Quality 40 classification error. % 40 245 classification error. % test 30 30 20 20 10~ ____- .__________ 3 weights (log 10) training ~~ 4 10 ~-------------~ 2 3 codebook vectors (log 1(} Figure 3: a) Measured classification errors for neural networks with increasing number of weights (capacity). The mean value between the test and training error estimates the performance of the given classifier trained with unlimited data. b) Measured classification errors for LVQ classifiers with increasing number of codebook vectors. Further experiments exploited neural nets with one layer of respectively 3, 5, 7, 10, 15, 20, 30, and 40 hidden units. All our results are summarized in Figure 3a). This figure illustrates several points mentioned in the text above. As the complexity of the network increases, the training error decreases because the networks get more free parameters to memorize the data. Compare to Figure 1a). The test error also decreases at first, going through a minimum of 29% at the network with 5 hidden units. This network apparently has a capacity that best matches the amount and character of the available training data. For higher capacity the networks overfit the data at the expense of increased error on the test set. Figure 3a) should also be compared to Figure 1c). In Figure 3a) we plotted approximate values of Eoo for the various networks - the minimal error of the network to the given task. The values of Eoo are estimated as the mean of the training and test errors. The value of Eoo appears to flatten out around the network with 30 units, asymptotically reaching a value of 24% error. An asymptotic Eoo-value of 25% was obtained from LVQ-experiments with increasing number of codebook vectors. These results are summarized in Figure 3b). We therefore conjecture that the intrinsic noise level of the task is about 25%, and this number is the limiting error rate imposed by the quality of the data on any learning machine applied to the task. 246 4 Corinna Cortes, L. D. Jackel, Wan-Ping Chiang Conclusion In this paper we have proposed a method for estimating the limits on performance imposed by the quality of the database on which a task is defined. The method involves a series of learning experiments. The extracted result is, however, independent of the choice of learning machine used for these experiments since the estimated limiting performance expresses a characteristic of the data. The only requirements on the learning machines are that their capacity can be varied and be made large, and that the machines with increasing capacity become capable of implementing any function. In this paper we have demonstrated the robustness of our method to the choice of classifiers: the result obtained with neural networks is in statistical agreement with the result obtained for LVQ classifiers. Using the proposed method we have investigated how well prediction of upcoming trouble in a telecommunication path can be performed based on information extracted from a given database. The analysis has revealed a very high intrinsic noise level of the extracted information and demonstrated the inadequacy of the data to construct high performance classifiers. This study is typical for many applications where the data collection was not necessarily designed for the problem at hand. Acknowledgments We gratefully acknowledge Vladimir Vapnik who brought this application to the attention of the Holmdel authors. One of the authors (CC) would also like to thank Walter Dziama, Charlene Paul, Susan Blackwood, Eric Noel, and Harris Drucker for lengthy explanations and helpful discussions of the AT&T transport system. References [1] [2] [3] [4] [5] [6] [7] [8] s. Bos, W. Kinzel, and M. Opper. The generalization ability of perceptrons with continuous output. Physical Review E, 47:1384-1391, 1993. Corinna Cortes. Prediction of Generalization Ability in Learning Machines. PhD thesis, University of Rochester, NY, 1993. Corinna Cortes, L. D. Jackel, Sara A. So1la, V. Vapnik, and John S. Denker. Learning curves: Asymptotic value and rate of convergence. In Advances in Neural Information Processing Systems, volume 6. Morgan Kaufman, 1994. G. Cybenko, K. Hornik, M. Stinchomb, and H. White. Multilayer feedforward neural networks are universal approximators. Neural Networks, 2:359-366, 1989. T. L. Fine. Statistical generalization and learning. Technical Report EE577, Cornell University, 1993. Teuvo Kohonen, Gyorgy Barna, and Ronald Chrisley. Statistical pattern recognition with neural networks: Benchmarking studies. In Proc. IEEE Int. Con! on Neural Networks, IJCNN-88, volume 1, pages 1-61-1-68, 1988. N. Murata, S. Yoshizawa, and S. Amari. Learning curves, model selection, and complexity of neural networks. In Advances in Neural Information Processing Systems, volume 5, pages 607-614. Morgan Kaufman, 1992. H. S. Seung, H. Sompolinsky, and N. Tishby. Statistical mechanics of learning from examples. Physical Review A, 45:6056-6091, 1992.
918 |@word seems:1 termination:1 carry:2 series:2 john:1 ronald:1 designed:2 plot:2 fewer:1 device:1 short:2 record:1 chiang:5 codebook:6 along:1 constructed:1 install:1 become:2 underfitting:1 behavior:3 mechanic:1 automatically:1 window:2 increasing:11 becomes:1 provided:1 estimating:3 begin:1 underlying:1 domestic:1 medium:3 kind:1 kaufman:2 nj:1 exactly:1 classifier:13 unit:8 before:1 service:3 limit:7 insufficiency:2 path:19 might:1 examined:1 sara:1 limited:1 statistically:1 acknowledgment:1 practice:1 implement:3 irregularity:1 universal:3 bell:1 flatten:1 get:2 close:1 selection:1 impossible:1 imposed:9 demonstrated:2 missing:1 of7:1 attention:1 survey:1 formulate:1 resolution:1 limiting:10 agreement:1 trend:1 recognition:1 continues:1 cut:1 database:12 labeled:3 bottom:2 capture:1 susan:1 connected:1 hinders:1 sompolinsky:1 decrease:5 highest:1 mentioned:2 complexity:6 seung:1 hinder:1 trained:4 depend:1 eric:1 completely:2 easily:1 various:1 fiber:1 corrective:1 train:2 walter:1 labeling:1 mon:1 larger:1 otherwise:1 amari:1 ability:3 commit:1 reproduced:1 advantage:1 net:1 propose:2 kohonen:1 achieve:1 description:1 los:1 convergence:1 regularity:1 requirement:2 produce:2 measured:3 ij:1 minor:1 received:1 implemented:1 entirety:1 involves:2 come:1 implies:1 quantify:1 indicate:1 memorize:1 correct:1 vc:1 human:1 material:1 implementing:2 eoo:12 generalization:4 wall:1 cybenko:1 adjusted:1 sufficiently:1 around:1 bursty:1 mapping:3 predict:1 vary:1 purpose:1 proc:1 label:5 jackel:5 largest:2 city:2 tool:1 brought:1 always:4 modified:1 reaching:1 cornell:1 varying:1 surveillance:1 improvement:1 transmitter:1 mainly:1 contrast:1 equipment:1 baseline:1 bos:1 helpful:1 typically:2 initially:1 hidden:6 pattems:1 misclassified:1 going:1 sketched:1 overall:1 classification:6 constrained:1 once:1 never:2 barring:1 construct:1 eliminated:1 identical:1 constitutes:1 future:2 report:1 randomly:1 resulted:3 individual:1 subsampled:1 replaced:1 phase:1 maintain:2 adjust:1 yielding:1 accurate:2 capable:3 daily:1 loosely:1 plotted:2 theoretical:1 minimal:1 instance:1 increased:2 modeling:1 entry:2 examining:1 teuvo:1 inadequate:1 tishby:1 characterize:1 reported:2 stored:2 corrupted:1 ie:3 chrisley:1 continuously:2 thesis:1 wan:4 leading:1 accompanied:1 summarized:2 includes:1 int:1 matter:1 depends:2 performed:2 apparently:1 traffic:1 reached:1 start:2 rochester:1 accuracy:3 characteristic:4 who:1 murata:1 correspond:1 produced:1 monitoring:1 cc:1 history:6 ping:5 reach:4 complicate:1 lengthy:1 failure:4 yoshizawa:1 mi:1 monitored:1 con:1 proved:1 knowledge:1 dimensionality:1 back:1 appears:1 higher:2 day:6 overfit:1 hand:3 transport:1 quality:13 oil:1 concept:1 true:3 hence:1 laboratory:1 illustrated:2 white:1 vagary:1 during:2 criterion:2 complete:2 demonstrate:1 performs:1 kinzel:1 physical:4 volume:3 measurement:1 gratefully:1 had:2 longer:2 base:1 recent:1 scenario:1 certain:1 manifested:1 arbitrarily:2 continue:1 approximators:1 exploited:1 seen:1 minimum:2 transmitted:2 greater:1 care:1 somewhat:1 morgan:2 gyorgy:1 period:3 redundant:1 signal:1 ii:1 full:1 smooth:1 technical:1 match:4 calculation:1 long:3 divided:1 prediction:6 regression:1 basic:1 multilayer:1 achieved:3 irregular:1 fine:1 interval:1 source:1 extra:1 exhibited:1 errored:1 induced:1 ample:1 member:3 call:1 exceed:3 revealed:2 enough:5 feedforward:1 automated:1 fit:1 decline:4 drucker:1 angeles:1 inadequacy:1 york:1 action:1 impairment:1 generally:1 amount:5 notice:2 dotted:1 estimated:5 correctly:1 express:2 group:2 threshold:4 drawn:1 asymptotically:3 fraction:1 run:1 telecommunication:3 logged:1 clipped:1 almost:1 reasonable:1 preemptive:1 holmdel:2 bit:7 layer:2 mined:1 ijcnn:1 unlimited:2 optical:1 relatively:1 conjecture:3 according:1 poor:1 belonging:1 smaller:1 slightly:1 character:2 cable:2 repair:2 taken:2 resource:1 previously:1 discus:3 eventually:3 count:1 needed:1 end:4 available:2 denker:1 preventive:1 corinna:7 robustness:1 top:2 ensure:2 include:3 trouble:14 upcoming:1 objective:1 occurs:1 coaxial:1 primary:1 strategy:1 guessing:2 exhibit:1 distance:1 link:3 thank:1 capacity:27 collected:1 length:1 reorganization:1 insufficient:1 vladimir:1 expense:1 trace:1 negative:1 rise:3 implementation:1 design:1 unknown:1 acknowledge:1 communication:2 committed:1 frame:4 varied:2 specified:1 optimized:1 barna:1 learnillg:1 framing:3 hour:1 inaccuracy:1 below:1 pattern:11 reliable:1 explanation:1 critical:1 difficulty:2 predicting:2 historically:1 picture:1 perf:1 text:1 review:2 checking:1 asymptotic:8 synchronization:1 fully:1 filtering:1 switched:1 imposes:1 principle:2 thresholding:1 prone:1 course:1 supported:1 free:7 curve:2 dimension:1 opper:1 author:2 collection:4 made:4 outlying:1 excess:1 approximate:1 keep:1 overfitting:1 receiver:1 alternatively:1 continuous:1 channel:6 contributes:1 hornik:1 investigated:1 complex:2 necessarily:1 constructing:2 big:1 noise:9 paul:1 benchmarking:1 ny:1 lc:2 ib:2 minute:4 highrate:1 cortes:7 intrinsic:6 quantization:1 false:1 adding:1 vapnik:2 phd:1 illustrates:1 suited:1 likely:1 constantly:1 extracted:8 harris:1 month:1 noel:1 lvq:4 change:1 typical:5 infinite:2 averaging:2 preset:1 degradation:2 engineer:1 called:1 catastrophic:1 experimental:2 la:3 blackwood:1 perceptrons:1 indicating:1 formally:1 internal:1 mark:1
7,156
919
Diffusion of Credit in Markovian Models Yoshua Bengio? Dept. I.R.O., Universite de Montreal, Montreal, Qc, Canada H3C-3J7 bengioyCIRO.UMontreal.CA Paolo Frasconi Dipartimento di Sistemi e Informatica Universita di Firenze, Italy paoloCmcculloch.ing.unifi.it Abstract This paper studies the problem of diffusion in Markovian models, such as hidden Markov models (HMMs) and how it makes very difficult the task of learning of long-term dependencies in sequences. Using results from Markov chain theory, we show that the problem of diffusion is reduced if the transition probabilities approach 0 or 1. Under this condition, standard HMMs have very limited modeling capabilities, but input/output HMMs can still perform interesting computations. 1 Introduction This paper presents an important new element in our research on the problem of learning long-term dependencies in sequences. In our previous work [4J we found theoretical reasons for the difficulty in training recurrent networks (or more generally parametric non-linear dynamical systems) to learn long-term dependencies. The main result stated that either long-term storing or gradient propagation would be harmed, depending on whether the norm of the Jacobian of the state to state function was greater or less than 1. In this paper we consider a special case in which the norm of the Jacobian of the state to state function is constrained to be exactly 1 because this matrix is a stochastic matrix. We consider both homogeneous and non-homogeneous Markovian models. Let n be the number of states and At be the transition matrices (constant in the homogeneous case): Aij(ut} = P(qt = j I qt-l = i, Ut; e) where Ut is an external input (constant in the homogeneous case) and e is a vector of parameters. In the homogeneous case (e.g., standard HMMs), such models can learn the distribution of output sequences by associating an output distribution to each state. In ?also, AT&T Bell Labs, Holmdel, NJ 07733 554 Yoshua Bengio. Paolo Frasconi the non-homogeneous case, transition and output distributions are conditional on the input sequences, allowing to model relationships between input and output sequences (e.g. to do sequence 'regression or classification as with recurrent networks). We thus called Input/Output HMM (IOHMM) this kind of non-homogeneous HMM . In [3, 2] we proposed a connectionist implementation of IOHMMs. In both cases, training requires propagating forward probabilities and backward probabilities, taking products with the transition probability matrix or its transpose. This paper studies in which conditions these products of matrices might gradually converge to lower rank, thus harming storage and learning of long-term context. However, we find in this paper that IOHMMs can better deal with this problem than homogeneous HMMs . 2 2.1 Mathematical Preliminaries Definitions A matrix A is said to be non-negative, written A 2:: 0, if Aij 2:: 0 Vi, j . Positive matrices are defined similarly. A non-negative square matrix A E R nxn is called row stochastic (or simply stochastic in this paper) if 'L,'l=1 Aij = 1 Vi = 1 . . . n. A non-negative matrix is said to be row [column} allowable if every row [column] sum is positive. An allowable matrix is both row and column allowable. A nonnegative matrix can be associated to the directed transition graph 9 that constrains the Markov chain. An incidence matrix A corresponding to a given non-negative matrix A replaces all positive entries of A by 1. The incidence matrix of A is a connectivity matrix corresponding to the graph 9 (assumed to be connected here). Some algebraic properties of A are described in terms of the topology of g. Definition 1 (Irreducible Matrix) A non-negative n x n matrix A is said to be irreducible if for every pair i,j of indices, :3 m = m(i,j) positive integer s.t. (Amhj > O. A matrix A is irreducible if and only if the associated graph is strongly connected (i.e., there exists a path between any pair of states i,j) . If :3k s.t . (Ak)ii > 0, d(i) is called the period of index i ifit is the greatest common divisor (g .c.d.) of those k for which (Ak)ii > O. In an irreducible matrix all the indices have the same period d, which is called the period of the matrix. The period of a matrix is the g .c.d. of the lengths of all cycles in the associated transition graph. Definition 2 (Primitive matrix) A non-negative matrix A is said to be primitive if there exists a positive integer k S.t. Ak > O. An irreducible matrix is either periodic or primitive (i.e. of period 1). A primitive stochastic matrix is necessarily allowable. 2.2 The Perron-Frobenius Theorem Theorem 1 (See [6], Theorem 1.1.) Suppose A is an n x n non-negative primitive matrix. Then there exists an eigenvalue r such that: 1. r is real and positive; 2. with r can be associated strictly positive left and right eigenvectors; 3. r> 1>'1 for any eigenvalue>. 1= r; 4? the eigenvectors associated with r are unique to constant multiples. 5. If 0 S B A and f3 is an eigenvalue of B, then 1f31 r . Moreover, 1f31 = r implies B = A. s s 555 Diffusion of Credit in Markovian Models 6. r is simple root of the characteristic equation of A. A simple consequence of the theorem for stochastic matrices is the following: Corollary 1 Suppose A is a primitive stochastic matrix. Then its largest eigehvalue is 1 and there is only one corresponding right eigenvector, which is 1 = [1, 1 .. ?1]'. Furthermore, all other eigenvalues < 1. Proof. A1 = 1 by definition of stochastic matrices. This eigenvector is unique and all other eigenvalues < 1 by the Perron-Frobenius Theorem. If A is stochastic but periodic with period d, then A has d eigenvalues of module 1 which are the d complex roots of 1. 3 Learning Long-Term Dependencies with HMMs In this section we analyze the case of a primitive transition matrix as well as the general case with a canonical re-ordering of the matrix indices. We discuss how ergodicity coefficients can be used to measure the difficulty in learning long-term dependencies. Finally, we find that in order to avoid all diffusion, the transitions should be deterministic (0 or 1 probability). 3.1 Training Standard HMMs Theorem 2 (See [6], Theorem 4.2.) If A is a primitive stochastic matrix, then as t -+ 00, At -+ 1V' where v' is the unique stationary distribution of the Markov chain. The rate of approach is geometric. Thus if A is primitive, then liIDt-+oo At converges to a matrix whose eigenvalues are all 0 except for ,\ = 1 (with eigenvector 1), i.e. the rank of this product converges to 1, i.e. its rows are equal. A consequence oftheorem 2 is that it is very difficult to train ordinary hidden Markov models, with a primitive transition matrix, to model long-term dependencies in observed sequences. The reason is that the distribution over the states at time t > to becomes gradually independent of the distribution over the states at time to as t increases. It means that states at time to become equally responsible for increasing the likelihood of an output at time t. This corresponds in the backward phase of the EM algorithm for trainin~ HMMs to a diffusion of credit over all the states. In practice we train HMMs WIth finite sequences. However , training will become more and more numerically ill-conditioned as one considers longer term dependencies. Consider two events eo (occurring at to) and et (occurring at t), and suppose there are also "interesting" events occurring in between. Let us consider the overall influence of states at times 1" < t upon the likelihood of the outputs at time t. Because of the phenomenon of diffusion of credit, and because gradients are added together, the influence of intervening events (especially those occurring shortly before t) will be much stronger than the influence of eo . Furthermore, this problem gets geometrically worse as t increases. Clearly a positive matrix is primitive. Thus in order to learn long-term dependencies, we would like to have many zeros in the matrix of transition probabilities. Unfortunately, this generally supposes prior knowledge of an appropriate connectivity graph . 3.2 Coefficients of ergodicity To study products of non-negative matrices and the loss of information about initial state in Markov chains (particularly in the non-homogeneous case), we introduce the projective distance between vectors x and y: x?y? d(x',y') = ~~ In(--.:..l.). I ,) XjYi Clearly, some contraction takes place when d(x'A,y'A) ::; d(x',y'). 556 Yoshua Bengio, Paolo Frasconi Definition 3 BirkhofJ's contraction coefficient TB(A), for a non-negative columnallowable matrix A, is defined in terms of the projective distance: TB(A) = sup x ,y> Ojx;t>.y d(x' A, y' A) d(x', y') Dobrushin's coefficient Tl(A), for a stochastic matrix A, is defined as follows: Tl(A) = 1 2 s~p L I,) laik - ajkl? k = Both are proper ergodicity coefficients: 0 ~ T(A) ~ 1 and T(A) 0 if and only if A has identical rows. Furthermore, T(AIA2) ~ T(Al)T(A2)(see [6]). 3.3 Products of Stochastic Matrices Let A (1 ,t) = A 1 A 2 ??? A t - 1 A t denote a forward product of stochastic matrices AI, A 2, ... At. From the properties of TB and Tl, if T(At} < 1, t > 0 then limt-l-oo T(A(l,t?) = 0, i.e. A(l,t) has rank 1 and identical rows. Weak ergodicity is then defined in terms of a proper ergodic coefficient T such as TB and Tl: Definition 4 (Weak Ergodicity) The products of stochastic matrices weakly ergodic if and only if for all to ~ 0 as t -+ 00, T(A(to,t?) -+ O. A(p,r) are Theorem 3 (See [6], Lemma 3.3 and 3.4.) Let A(l,t) a forward product of non-negative and allowable matrices, then the products A(l,t) are weakly ergodic if and only if the following conditions both hold: 1. 3to S.t. A(to,t) > 0 Vt > to A(to,t) 2. A (;~,t) - -+ Wij (t) > 0 as t -+ 00, i. e. rows of A (to,t) tend to proportionality. ),k For stochastic matrices, row-proportionality is equivalent to row-equality since rows sum to 1. limt-l-oo ACto,t) does not need to exist in order to have weak ergodicity. 3.4 Canonical Decomposition and Periodic Graphs Any non-negative matrix A can be rewritten by relabeling its indices in the following canonical decomposition [6], with diagonal blocks B i , C i and Q: A= Bl 0 0 B2 ( ..... . ..... . 0 . .. . . . ... 0 0 Ll L2 C'+ 1 0 0 0 ... 0 . 0 . . .. . . . " . .... . .. Cr Lr 0 0 Q ) (1 ) where Bi and Ci are irreducible, Bi are primitive and Ci are periodic. Define the corresponding sets of states as SBi' Se" Sq. Q might be reducible, but the groups of states in Sq leak into the B or C blocks, i.e., Sq represents the transient part of the state space. This decomposition is illustrated in Figure 1a. For homogeneous and non-homogeneous Markov models (with constant incidence matrix At = Ao), because P(qt E Sqlqt-l E Sq) < 1, liIl1t-l-oo P(qt E Sqlqo E Sq) = O. Furthermore, because the Bi are primitive, we can apply Theorem 1, and starting from a state in SB" all information about an initial state at to is gradually lost. 557 Diffusion of Credit in Markovian Models (b) Figure 1: (a): Transition graph corresponding to the canonical decomposition. (b): Periodic graph 91 becomes primitive (period 1) 92 when adding loop with states 4,5. A more difficult case is the one of (A(to ,t))jk with initial state j ESc, . Let d i be the period of the ith periodic block Cj. It can be shown r6] that taking d products of periodic matrices with the same incidence matrix and period d yields a blockdiagonal matrix whose d blocks are primitive. Thus C(to ,t) retains information about the initial block in which qt was. However, for every such block of size > 1, information will be gradually lost about the exact identity of the state within that block. This is best demonstrated through a simple example. Consider the incidence matrix represented by the graph 91 of Figure lb. It has period 3 and the only non-deterministic transition is from state 1, which can yield into either one of two loops. When many stochastic matrices with this graph are multiplied together, information about the loop in which the initial state was is gradually lost (i.e. if the initial state was 2 or 3, this information is gradually lost). What is retained is the phase information, i.e. in which block ({O}, {I}, or {2,3}) of a cyclic chain was the initial state. This suggests that it will be easy to learn about the type of outputs associated to each block of a cyclic chain, but it will be hard to learn anything else. Suppose now that the sequences to be modeled are slightly more complicated, requiring an extra loop of period 4 instead of 3, as in Figure lb. In that case A is primitive: all information about the initial state will be gradually lost. 3.5 Learning Long-Term Dependencies: a Discrete Problem? We might wonder if, starting from a positive stochastic matrix, the learning algorithm could learn the topology, i.e. replace some transition probabilities by zeroes. Let us consider the update rule for transition probabilities in the EM algorithm: A oL ij 8A;j (2) A ij ~ " oL . wj Aij oA.j Starting from Aij > 0 we could obtain a new Aij = 0 only if O~~j = 0, i.e. on a local maximum of the likelihood L. Thus the EM training algorithm will not exactly obtain zero probabilities. Transition probabilities might however approach O. It is also interesting to ask in which conditions we are guaranteed that there will not be any diffusion (of influence in the forward phase, and credit in the backward phase of training). It requires that some of the eigenvalues other than Al = 1 have a norm that is also 1. This can be achieved with periodic matrices C (of period 558 Yoshua Bengio, PaoLo Frasconi 5 -: Periodic_ . .~:;:~. -"~::::-~:-~,~~:.~~:-~:-~~:~--.-.-.~~~~~"'-"" -10 :l:~; %?.';:"'?.:: 1: .??.?:.??. II ...: ill) Left-to-right- '-.'-, .. l?:t """ '".,''. 'I d ?.?..?.?.:.' ? .??..:?.. ?: .?: .. ? .? '/,.. .. -15 ~ --20 / Full connected -25 , , Left-to-right (triangular) -30 5 10 15 T 20 25 30 (a) II t=4 t=3 (b) Figure 2: (a) Convergence of Dobrushin's coefficient (see Definition 3. (b) Evolution of products A(l,t) for fully connected graph. Matrix elements are visualized with gray levels. d), which have d eigenvalues that are the d roots of 1 on the complex unit circle. To avoid any loss of information also requires that Cd = I be the identity, since any diagonal block of Cd with size more than 1 will yield to a loss of information (because of diffusion in primitive matrices) . This can be generalized to reducible = I. matrices whose canonical form is composed of periodic blocks Ci with ct The condition we are describing actually corresponds to a matrix with only 1 's and O's_ If At is fixed, it would mean that the Markov chain is also homogeneous. It appears that many interesting computations can not be achieved with such constraints (i.e. only allowing one or more cycles of the same period and a purely deterministic and homogeneous Markov chain). Furthermore, if the parameters of the system are the transition probabilities themselves (as in ordinary HMMs), such solutions correspond to a subset of the corners of the 0-1 hypercube in parameter space. Away from those solutions, learning is mostly influenced by short term dependencies, because of diffusion of credit. Furthermore, as seen in equation 2, algorithms like EM will tend to stay near a corner once it is approached. This suggests that discrete optimization algorithms, rather continuous local algorithms, may be more appropriate to explore the (legal) corners of this hypercube. 4 Experiments 4.1 Diffusion: Numerical Simulations Firstly, we wanted to measure how (and if) different kinds of products of stochastic matrices converged, for example to a matrix of equal rows. We ran 4 simulations, each with an 8 states non-homogeneous Markov chain but with different constraints on the transition graph: 1) 9 fully connected; 2) 9 is a left-to-right model (i.e. A is upper triangular); 3) 9 is left-to-right but only one-state skips are allowed (i.e. A is upper bidiagonal); 4) At are periodic with period 4. Results shown in Figure 2 confirm the convergence towards zero of the ergodicity coefficient 1 , at a rate that depends on the graph topology. In Figure 2, we represent visually the convergence of fully connected matrices, in only 4 time steps, towards equal columns. lexcept for the experiments with periodic matrices, as expected Diffusion of Credit in Markovian Models 559 100,----~-..-???-~/~???~~~\-.~-:-/~:--~-~-~--~~-~--yg-iV-en~ 80 / \? __ -----'\ \\ ,' ..,\ .........- ~ 40stales \ ~ Fully <XlOIIeCIed, .. 16 llitale.1II \ " \ Randomly co",,..;ted. \ \. 24" ... , \ ?./ .... \ ./ 'I Fully conrect.:ted. \ \. \ ~_. \ Fully connected, ',\ 20 .1 ? \\ \"--24 sl.ate." \ ,, \ - _ ?? -yoo,........ Cb.1 (a) 10 Span 1000 (b) Figure 3: (a): Generating HMM. Numbers out of state circles denote output symbols. (b): Percentage of convergence to a good solution (over 20 trials) for various series o( experiments as the span of dependencies is increased. 4.2 Training Experiments To evaluate how diffusion impairs training, a set of controlled experiments were performed, in which the training sequences were generated by a simple homogeneous HMM with long-term dependencies, depicted in Figure 3a. Two branches generate similar sequences except for the first and last symbol. The extent of the longterm context is controlled by the self transition probabilities of states 2 and 5, A = P(qt = 2lqt-l - 2) = P(qt = 5lqt-l = 5). Span or "half-life" is log(.5)/ log(A), i.e. Aspan = .5). Following [4], data was generated for various span of long-term dependencies (0.1 to 1000). For each series of experiments, varying the span, 20 different training trials were run per span value, with 100 training sequences 2 . Training was stopped either after a maximum number of epochs (200), of after the likelihood did not improve significantly, i.e., (L(t) - L(t - l))/IL(t)1 < 10- 5 , where L(t) is the logarithm of the likelihood of the training set at epoch t. If the HMM is fully connected (except for the final absorbing state) and has just the right number of states, trials almost never converge to a good solution (1 in 160 did). Increasing the number of states and randomly putting zeroes helps. The randomly connected HMMs had 3 times more states than the generating HMM and random connections were created with 20% probability. Figure 3b shows the average number of converged trials for these different types of HMM topology. A trial is considered successful when it yields a likelihood almost as good or better than the likelihood of the generating HMM on the same data. In all cases the number of successful trials rapidly drops to zero beyond some value of span. 5 Conclusion and Future Work In previous work on recurrent networks we had found that propagating credit over the long term was incompatible with storing information for the long term. For Markovian models, we found that when the transition probabilities are close to 1 and 0, information can be stored for the long term AND credit can be prop2it appeared sufficient since the likelihood of the generating HMM did not improve much when trained on this data 560 Yoshua Bengio, PaoLo Frasconi agated over the long term. However, like for recurrent networks, this makes the problem of learning long-term dependencies look more like a discrete optimization problem. Thus it appears difficult for local learning algorithm such as EM to learn optimal transition probabilities near 1 or 0, i.e. to learn the topology, while taking into account long-term dependencies. The arguments presented are essentially an application of established mathematical results on Markov chains to the problem of learning long term dependencies in homogeneous and non-homogeneous HMMs. These arguments were also supported by experiments on artificial data, studying the phenomenon of diffusion of credit and the corresponding difficulty in training HMMs to learn long-term dependencies . IOHMMs [1] introduce a reparameterization of the problem: instead of directly learning the transition probabilities, we learn parameters of a function of an input sequence. Even with a fully connected topology, transition probabilities computed at each time step might be very close to and 1. Because of the non-stationarity, more interestin~ computations can emerge than the simple cycles studied above. For example in l3] we found IOHMMs effective in grammar inference tasks. In [1] comparative experiments were performed with a preliminary version of IOHMMs and other algorithms such as recurrent networks, on artificial data on which the span of long-term dependencies was controlled. IOHMMs were found much better than the other algorithms at learning these tasks. ? Based on the analysis presented here, we are also exploring another approach to learning long-term dependencies that consists in building a hierarchical representation of the state. This can be achieved by introducing several sub-state variables whose Cartesian product corresponds to the system state. Each of these sub-state variables can operate at a different time scale, thus allowing credit to propagate over long temporal spans for some of these variables . Another interesting issue to be investigated is whether techniques of symbolic prior knowledge injection (such as in (5]) can be exploited to choose good topologies. One advantage, compared to traditIOnal neural network approaches, is that the model has an underlying finite state structure and is thus well suited to inject discrete transition rules . Acknowledgments We would like to thank Leon Bottou for his many useful comments and suggestions, and the NSERC and FCAR Canadian funding agencies for support. References [1] Y. Bengio and P. Frasconi. Credit assignment through time: Alternatives to backpropagation. In J. D. Cowan, et al., eds., Advances in Neural Information Processing Systems 6. Morgan Kaufmann, 1994. [2] Y. Bengio and P. Frasconi. An Input Output HMM Architecture. In this volume: J. D. Cowan, et al., eds., Advances in Neural Information Processing Systems 7. Morgan Kaufmann, 1994. [3] Y. Bengio and P. Frasconi. An EM approach to learning sequential behavior. Technical Report RT-DSI-ll/94 , University of Florence, 1994. [4] Y. Bengio, P. Simard, and P. Frasconi. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Networks, 5(2):157- 166, 1994. [5] P. Frasconi, M. Gori, M. Maggini, and G . Soda. Unified integration of explicit rules and learning by example in recurrent networks. IEEE Trans. on Knowledge and Data Engineering, 7(1), 1995. [6] E. Seneta. Nonnegative Matrices and Markov Chains. Springer, New York, 1981.
919 |@word trial:6 version:1 longterm:1 norm:3 stronger:1 aia2:1 proportionality:2 simulation:2 propagate:1 contraction:2 decomposition:4 initial:8 cyclic:2 series:2 incidence:5 written:1 numerical:1 wanted:1 drop:1 update:1 stationary:1 half:1 ith:1 short:1 lr:1 firstly:1 mathematical:2 become:2 consists:1 introduce:2 expected:1 behavior:1 themselves:1 ol:2 increasing:2 becomes:2 moreover:1 underlying:1 what:1 kind:2 eigenvector:3 unified:1 nj:1 temporal:1 every:3 exactly:2 unit:1 positive:9 before:1 engineering:1 local:3 consequence:2 ak:3 path:1 might:5 studied:1 suggests:2 co:1 hmms:13 limited:1 projective:2 bi:3 directed:1 unique:3 responsible:1 acknowledgment:1 practice:1 block:11 lost:5 firenze:1 backpropagation:1 sq:5 bell:1 significantly:1 symbolic:1 get:1 close:2 storage:1 context:2 influence:4 equivalent:1 deterministic:3 demonstrated:1 primitive:17 starting:3 ergodic:3 qc:1 rule:3 his:1 reparameterization:1 xjyi:1 suppose:4 exact:1 homogeneous:17 element:2 particularly:1 jk:1 observed:1 module:1 reducible:2 wj:1 connected:10 cycle:3 ordering:1 sbi:1 ran:1 agency:1 leak:1 constrains:1 trained:1 weakly:2 purely:1 upon:1 represented:1 various:2 train:2 effective:1 artificial:2 approached:1 whose:4 triangular:2 grammar:1 h3c:1 aspan:1 final:1 sequence:13 eigenvalue:9 advantage:1 product:13 loop:4 rapidly:1 intervening:1 frobenius:2 convergence:4 generating:4 comparative:1 converges:2 help:1 depending:1 recurrent:6 montreal:2 propagating:2 oo:4 ij:2 qt:7 skip:1 implies:1 stochastic:17 transient:1 ao:1 preliminary:2 lqt:2 dipartimento:1 strictly:1 exploring:1 hold:1 credit:13 considered:1 visually:1 cb:1 a2:1 largest:1 clearly:2 j7:1 harmed:1 rather:1 avoid:2 cr:1 varying:1 corollary:1 rank:3 likelihood:8 inference:1 iohmm:1 sb:1 hidden:2 wij:1 overall:1 classification:1 ill:2 issue:1 constrained:1 special:1 integration:1 equal:3 once:1 f3:1 frasconi:10 ted:2 never:1 identical:2 represents:1 look:1 future:1 yoshua:5 connectionist:1 report:1 irreducible:6 randomly:3 composed:1 relabeling:1 phase:4 divisor:1 stationarity:1 chain:11 iv:1 logarithm:1 re:1 circle:2 theoretical:1 stopped:1 increased:1 column:4 modeling:1 markovian:7 retains:1 assignment:1 ordinary:2 introducing:1 entry:1 subset:1 wonder:1 successful:2 stored:1 dependency:20 supposes:1 periodic:11 stay:1 harming:1 together:2 yg:1 connectivity:2 choose:1 worse:1 external:1 corner:3 inject:1 simard:1 account:1 de:1 b2:1 coefficient:8 vi:2 depends:1 performed:2 root:3 lab:1 analyze:1 sup:1 capability:1 complicated:1 florence:1 square:1 il:1 kaufmann:2 characteristic:1 yield:4 correspond:1 weak:3 converged:2 influenced:1 ed:2 definition:7 universite:1 associated:6 di:2 proof:1 ask:1 knowledge:3 ut:3 cj:1 actually:1 appears:2 strongly:1 furthermore:6 ergodicity:7 just:1 propagation:1 gray:1 building:1 requiring:1 evolution:1 equality:1 illustrated:1 deal:1 ll:2 self:1 anything:1 generalized:1 allowable:5 funding:1 umontreal:1 common:1 absorbing:1 volume:1 numerically:1 ai:1 similarly:1 had:2 l3:1 longer:1 italy:1 vt:1 life:1 exploited:1 seen:1 morgan:2 greater:1 eo:2 converge:2 period:14 ii:5 branch:1 multiple:1 full:1 ing:1 technical:1 long:24 dept:1 maggini:1 equally:1 a1:1 controlled:3 regression:1 essentially:1 represent:1 limt:2 achieved:3 else:1 extra:1 operate:1 comment:1 tend:2 cowan:2 integer:2 near:2 canadian:1 bengio:9 easy:1 architecture:1 associating:1 topology:7 whether:2 impairs:1 algebraic:1 york:1 generally:2 useful:1 se:1 eigenvectors:2 ifit:1 visualized:1 informatica:1 reduced:1 generate:1 sl:1 exist:1 percentage:1 canonical:5 sistemi:1 per:1 discrete:4 paolo:5 group:1 putting:1 diffusion:15 backward:3 graph:13 geometrically:1 sum:2 run:1 soda:1 place:1 almost:2 incompatible:1 holmdel:1 ct:1 guaranteed:1 replaces:1 nonnegative:2 constraint:2 argument:2 span:9 leon:1 injection:1 ate:1 slightly:1 em:6 gradually:7 legal:1 equation:2 discus:1 describing:1 studying:1 rewritten:1 multiplied:1 apply:1 hierarchical:1 away:1 appropriate:2 alternative:1 shortly:1 gori:1 esc:1 especially:1 universita:1 hypercube:2 bl:1 added:1 parametric:1 rt:1 diagonal:2 traditional:1 said:4 gradient:3 distance:2 thank:1 oa:1 hmm:10 s_:1 considers:1 extent:1 reason:2 length:1 index:5 relationship:1 retained:1 modeled:1 difficult:5 unfortunately:1 mostly:1 seneta:1 stated:1 negative:11 implementation:1 proper:2 perform:1 allowing:3 upper:2 markov:12 finite:2 iohmms:6 descent:1 lb:2 canada:1 pair:2 perron:2 connection:1 established:1 trans:2 beyond:1 dynamical:1 appeared:1 tb:4 greatest:1 event:3 difficulty:3 improve:2 created:1 prior:2 geometric:1 l2:1 blockdiagonal:1 epoch:2 nxn:1 loss:3 fully:8 dsi:1 interesting:5 suggestion:1 sufficient:1 storing:2 cd:2 row:12 supported:1 last:1 transpose:1 aij:6 taking:3 emerge:1 f31:2 transition:23 forward:4 confirm:1 assumed:1 continuous:1 learn:10 ca:1 investigated:1 necessarily:1 complex:2 bottou:1 did:3 main:1 allowed:1 tl:4 en:1 sub:2 explicit:1 trainin:1 r6:1 unifi:1 jacobian:2 theorem:9 symbol:2 exists:3 adding:1 sequential:1 ci:3 conditioned:1 occurring:4 cartesian:1 suited:1 depicted:1 simply:1 explore:1 nserc:1 springer:1 corresponds:3 conditional:1 identity:2 towards:2 replace:1 hard:1 except:3 lemma:1 called:4 support:1 dobrushin:2 evaluate:1 yoo:1 phenomenon:2
7,157
92
544 MURPHY: A Robot that Learns by Doing Bartlett W. Mel Center for Complex Systems Research University of Illinois 508 South Sixth Street Champaign, IL 61820 January 2, 1988 Abstract MURPHY consists of a camera looking at a robot arm, with a connectionist network architecture situated in between. By moving its arm through a small, representative sample of the 1 billion possible joint configurations, MURPHY learns the relationships, backwards and forwards, between the positions of its joints and the state of its visual field. MURPHY can use its internal model in the forward direction to "envision" sequences of actions for planning purposes, such as in grabbing a visually presented object, or in the reverse direction to "imitate", with its arm, autonomous activity in its visual field. Furthermore, by taking explicit advantage of continuity in the mappings between visual space and joint space, MURPHY is able to learn non-linear mappings with only a single layer of modifiable weights. Background Current Focus Of Learning Research Most connectionist learning algorithms may be grouped into three general catagories, commonly referred to as supenJised, unsupenJised, and reinforcement learning. Supervised learning requires the explicit participation of an intelligent teacher, usually to provide the learning system with task-relevant input-output pairs (for two recent examples, see [1,2]). Unsupervised learning, exemplified by "clustering" algorithms, are generally concerned with detecting structure in a stream of input patterns [3,4,5,6,7]. In its final state, an unsupervised learning system will typically represent the discovered structure as a set of categories representing regions of the input space, or, more generally, as a mapping from the input space into a space of lower dimension that is somehow better suited to the task at hand. In reinforcement learning, a "critic" rewards or penalizes the learning system, until the system ultimately produces the correct output in response to a given input pattern [8]. It has seemed an inevitable tradeoff that systems needing to rapidly learn specific, behaviorally useful input-output mappings must necessarily do so under the auspices of an intelligent teacher with a ready supply of task-relevant training examples. This state of affairs has seemed somewhat paradoxical, since the processes of Rerceptual and cognitive development in human infants, for example, do not depend on the moment by moment intervention of a teacher of any sort. Learning by Doing The current work has been focused on a fourth type of learning algorithm, i.e. learning-bydoing, an approach that has been very little studied from either a connectionist perspective ? American Institute of Physics 1988 545 or in the context of more traditional work in machine learning. In its basic form, the learning agent ? begins with a repertoire of actions and some form of perceptual input, ? exercises its repertoire of actions, learning to predict i) the detailed sensory consequences of its actions, and, in the other direction, ii) its actions that are associated with incoming sensory patterns, and ? runs its internal model (in one or both directions) in a variety of behaviorally-relevant tasks, e.g. to "envision" sequences of actions for planning purposes, or to internally "imitate" , via its internal action representation, an autonomously generated pattern of perceptual activity. In comparison to standard supenJised learning algorithms, the crucial property of learning-by-doing is that no intelligent teacher is needed to provide input-output pairs for learning. Laws of physics simply translate actions into their resulting percepts, both of which are represented internally. The learning agent need only notice and record this relationship for later use. In contrast to traditional unsupervised learning approaches, learning-by-doing allows the acquisition of specific, task-relevant mappings, such as the relationship between a simultaneously represented visual and joint state. Learning-bydoing differs as well from reinforcement paradigms in that it can operate in the absence of a critic, i.e. in situations where reward or penalty for a particular training instance may be inappropriate. Learning by doing may therefore by described as an unsupeMJised associative algorithm, capable of acquiring rich, task-relevant associations, but without an intelligent teacher or critic. Abridged History of the Approach The general concept of leaning by doing may be attributed at least to Piaget from the 1940's (see [9] for review). Piaget, the founder of the "constructivist" school of cognitive development, argued that knowledge is not given to a child as a passive observer, but is rather discovered and constructed by the child, through active manipulation of the environment. A handful of workers in artificial intelligence have addressed the issue of learning-by-doing, though only in highly schematized, simulated domains, where actions and sensory states are represented as logical predicates [10,11,12,13]. Barto & Sutton [14] discuss learning-by-doing in the context of system identification and motor control. They demonstrated how a simple simulated automaton with two actions and three sensory states can build a model of its environment through exploration, and subsequently use it to choose among behavioral alternatives. In a similar vein, Rumelhart [15] has suggested this same approach could be used to learn the behavior of a robot arm or a set of speech articulators. Furthermore, the forward-going "mental model" , once learned, could be used internally to train an inverse model using back-propagation. In previous work, this author [16] described a connectionist model (VIPS) that learned to perform 3-D visual transformations on simulated wire-frame objects. Since in complex sensory-motor environments it is not possible, in general, to learn a direct relationship between an outgoing command state and an incoming sensory state, VIPS was designed to predict changes in the state of its visual field as a function of its outgoing motor command. VIPS could then use its generic knowledge of motor-driven visual transformations to "mentally rotate" objects through a series of steps. 546 Recinolopic Visual RepresemaDOII -- - Value-Coded loint RepresenlatiOll - -- I Figure 1: MURPHY's Connectionist Architecture. 4096 coarsely-tuned visual units are organized in a square, retinotopic grid. These units are bi-directionally interconnected with a population of 273 joint units. The joint population is subdivided into 3 sUbpopulations, each one a value-coded representation of joint angle for one of the three joints. During training, activity in the joint unit population determines the physical arm configuration. Inside MURPY The current work has sought to further explore the process of learning-by-doing in a complex sensory-motor domain, extending previous work in three ways. First, the learning of mappings between sensory and command (e.g. motor) representations should be allowed to proceed in both directions simultaneously during exploratory behavior, where each mapping may ultimately subserve a very different behavioral goal. Secondly, MURPHY has been implemented with a real camera and robot arm in order to insure representational realism to the greatest degree possible. Third, while the specifics of MURPHY's internal structures are not intended as a model of a specific neural system, a serious attempt has been made to adhere to architectural components and operations that have either been directly suggested by nervous system structures, or are at least compatible with what is currently known. Detailed biological justification on this point awaits further work. MURPHY's Body MURPHY consists of a 512 x 512 JVC color video camera pointed at a Rhino XR-3 robotic arm. Only the shoulder, elbow, and wrist joints are used, such that the arm can move only in the image plane of the camera. (A fourth, waist joint is not used). White spots are stuck to the arm in convenient places; when the image is thresholded, only the white spots appear in the image. This arrangement allows continuous control over the complexity of the visual image of the arm, which in turn affects time spent both in computing visual features and processing weights during learning. A Datacube image processing system is used for the thresholding operation and to "blur" the image in real time with a gaussian mask. The degree of blur is variable and can be used to control the degree of coarse-coding (i.e. receptive field overlap) in the camera-topic array of visual units. The arm is software controllable, with a stepper motor for each joint. Arm dynamics are not considered in this work. 547 MURPHY's Mind MURPHY is currently based on two interconnected populations of neuron-like units. The first is organized as a rectangular, visuotopically-mapped 64 x 64 grid of coarsely-tuned visual units that each responds when a visual feature (such as a white spot on the arm) falls into its receptive field (fig. 1). Coarse coding insures that a single visual feature will activate a small population of units whose receptive fields overlap the center of stimulation. The upper trace in figure 2 shows the unprocessed camera view, and the center trace depicts the resulting pattern of activation over the grid of visual units. The second population of 273 units consists of three subpopulations, representing the angles of each of the three joints. The angle of each joint is value-coded in a line of units dedicated to that joint (fig. 1). Each unit in the population is "centered" at a some joint angle, and is maximally activated when the joint is to be sent to that angle. Neighboring joint units within a joint sub population have overlapping "projective fields" and progressively increasing joint-angle centers. It may be noticed that both populations of units are coarsely tuned, that is, the units have overlapping receptive fields whose centers vary in an orderly fashion from unit to neighboring unit. This style of representation is extremely common in biological sensory systems [17,18,19]' and has been attributed a number ofrepresentational advantages (e.g. fewer units needed to encode range of stimuli, increased immunity to noise and unit malfunction, and finer stimulus discriminations). A number of additional advantages of this type of encoding scheme are discussed in section, in relation to ease of learning, speed of learning, and efficacy of generalization. MURPHY's Education By moving its arm through a small, representative sample (approximately 4000) of the 1 billion possible joint configurations, MURPHY learns the relationships, backwards and forwards, between the positions of its joints and the state of its visual field. During training, the physical environment to which MURPHY's visual and joint representations are wired enforces a particular mapping between the states of these two representations. The mapping comprises both the kinematics of the arm as well as the optical parameters and global geometry of the camera/imaging system. It is incrementally learned as each unit in population B comes to "recognize" , through a process of weight modification, the states of population A in which it has been strongly activated. After sufficient training experience therefore, the state of popUlation A is sufficient to generate a "mental image" on population B, that is, to predictively activate the units in B via the weighted interconnections developed during training. In its current configuration, MURPHY steps through its entire joint space in around 1 hour, developing a total of approximately 500,000 weights between the two popUlations . The Learning Rule Tradeoffs in Learning and Representation It is well known in the folklore of connectionist network design that a tradeoff exists between the choice of representation (i.e. the "semantics") at the single unit level and the consequent ease or difficulty of learning within that representational scheme. At one extreme, the single-unit representation might be completely decoded, calling for a separate unit for each possible input pattern. While this scheme requires a combinatorially explosive number of units, and the system must "see" every possible input pattern during training, the actual weight modification rule is rendered very simple. At another extreme, the single unit representation might be chosen in a highly encoded fashion with complex interactions among input units. In this case, the activation of an output unit 548 may be a highly non-linear or discontinuous function of the input pattern, and must be learned and represented in mUltiple layers of weights. Research in connectionism has often focused on Boolean functions [20,21,1,22,23], typified by the encoder problem [20], the shifter problem [21] and n-bit parity [22]. Since Boolean functions are in general discontinuous, such that two input patterns that are close in the sense of Hamming distance do not in general result in similar outputs, much effort has been directed toward the development of sophisticated, multilayer weight-modification rules (e.g. back-propagation) capable of learning arbitrary discontinuous functions. The complexity of such learning procedures has raised troubling questions of scaling behavior and biological plausibility. The assumption of continuity in the mappings to be learned, however, can act to significationly simplify the learning problem while still allowing for full generalization to novel input patterns. Thus, by relying on the continuity assumption, MURPHY's is able to learn continuous non-linear functions using a weight modification procedure that is simple, locally computable, and confined to a single layer of modifiable weights. How MURPHY learns For sake of concrete illustration, MURPHY's representation and learning scheme will be described in terms of the mapping learned from joint units to visual units during training. The output activity of a given visual unit may be described as a function over the 3dimensional joint space, whose shape is determined by the kinematics of the arm, the location of visual features (i.e. white spots) on the arm, the global properties of the camera/imaging system, and the location of the visual unit's receptive field. In order for the function to be learned, a visual unit must learn to "recognize" the regions of joint space in which it has been visually activated during training. In effect, each visual unit learns to recognize the global arm configurations that happen to put a white spot in its receptive field. It may be recalled that MURPHY's joint unit population is value-coded by dimension, such that each unit is centered on a range of joint angles (overlapping with neighboring units) for one of the 3 joints. In this representation, a global arm configuration can be represented as the conjunctive activity of the k (where k 3) most active joint units. MURPHY's visual units can therefore learn to recognize the regions of joint space in which they are strongly activated by simply ''memorizing'' the relevant global joint configurations as conjunctive clusters of input connections from the value-coded joint unit population. To realize this conjunctive learning scheme, MURPHY's uses sigma-pi units (see [24]), as described below. At training step S, the set of k most active joint units are first identified. Some subset of visual units is also strongly activated in state S, each one signalling the presence of a visual feature (such as a white spot) in its receptive fields. At the input to each active visual unit, connections from the k most highly active joint units are formed as a multiplicative k-tuple of synaptic weights. The weights Wi on these connections are initially chosen to be of unit strength. The output Cj of a given synaptic conjunction is computed by multiplying the k joint unit activation values Xi together with their weights: = The output y of the entire unit is computed as a weighted sum of the outputs of each conjunction and then applying a sigmoidal nonlinearity: y = u(L WjCj). i Sigma-pi units of this kind may be thought of as a generalization of a logical disjunction of conjunctions (OR of AND's). The multiplicative nature of the conjunctive clusters insures 549 that every input to the conjunct is active in order for the conjunct to have an effect on the unit as a whole. If only a single input to a conjunct is inactive, the effect of the conjunction is nullified. Specific-Instance Learning in Continuous Domains MURPHY's learning scheme is directly reminiscent of specific-instance learning as discussed by Hampson &: Vol per [23] in their excellent review of Boolean learning and representational schemes. Specific-instance learning requires that each unit simply ''memorize" all relevant input states, i.e. those states in which the unit is intended to fire. Unfortunately, simple specific-instance learning allows for no generalization to novel inputs, implying that each desired system responses will have to have been explicitly seen during training. Such a state of affairs is clearly impractical in natural learning contexts. Hampson &: Volper [23] have further shown that random Boolean functions will require an exponential number of weights in this scheme. For continous functions on the other hand, two kinds of generalization are possible within this type of specific-instance learning scheme. We consider each in turn, once again from the perspective of MURPHY's visual units learning to recognize the regions in joint space in which they are activated. Generalization by Coarse-Coding When a visual unit is activated in a given joint configuration, and acquires an appropriate conjunct of weights from the set of highly active units in the joint popUlation, by continuity the unit may assume that it should be at least partially activated in nearby joint configurations as well. Since MURPHY's joint units are coarse-coded in joint angle, this will happen automatically: as the joints are moved a small distance away from the specific training configuration, the output of the conjunct encoding that training configuration will decay smoothly from its maximum. Thus, a visual unit can "fire" predictively in joint configurations that it has never specifically seen during training, by interpolating among conjuncts that encode nearby joint configurations. This scheme suggests that training must be sufficiently dense in joint space to have seen configurations ''nearby'' to all points in the space by some criterion. In practice, the training step size is related to the degree of coarse-coding in the joint population, which is chosen in tUrn such that a joint pertubation equal to the radius of a joint unit's projective field (i.e. the range of joint angles over which the unit is active) should on average push a feature in the visual field a distance of about one visual receptive field radius. As a rule of thumb, the visual receptive field radius is chosen small enough so as to contain only a single feature on average. Generalization by Extrapolation The second type of generalization is based on a heuristic principle, again illustrated in terms of learning in the visual population. If a visual unit has during training been very often activated over a large, easy-to-specify region of joint space, such as a hyperrectangular region, then it may be assumed that the unit is activated over the entire region of joint space, i.e. even at points not yet seen. At the synaptic level, "large regions" can be represented as conjuncts with fewer terms. In its simplest form, this kind of generalization amounts to simply throwing out one or more joints as irrelevant to the activation of a given visual unit. What synaptic mechanism can achieve this effect? Competition among joint unit afferents can be used to drive irrelevant variables from the sigma-pi conjuncts. Thus, if a visual unit is activated repeatedly during training, and the elbow and shoulder angle units are constantly active while the most active wrist unit varies from step to step, then the weighted connections from the repeatedly activated elbow and shoulder units 550 -- . ? ? ? -- ? , ?? ? ?? ? ? ? ,.. til I .; , ?.. ? ? ,., I Figure 2: Three Visual Traces. The top trace shows the unprocessed camera view of MURPHY's arm. White spots have been stuck to the arm at various places, such that a thresholded image contains only the white spots. This allows continuous control over the visual complexity of the image. The center trace represents the resulting pattern of activation over the 64 x 64 grid of coarsely-tuned visual units. The bottom trace depicts an internally-produced "mental" image of the arm in the same configuration as driven by weighted connections from the joint population. Note that the "mental" trace is a sloppy, but highly recognizable approximation to the camera-driven trace. will become progressively and mutually reinforced at the expense of the set of wrist unit connections, each of which was only activated a single time. This form of generalization is similar in function to a number of "covering" algorithms designed to discover optimal hyper-rectangular decompositions (with possible overlap) of a set of points in a multi-dimensional space (e.g. [25,26]). The competitive feature has not yet been implemented explicitly at the synaptic level, rather, the full set of conjuncts acquired during training are currently collapsed en masse into a more compact set of conjuncts, according to the above heuristics. In a typical run, MURPHY is able to eliminate between 30% and 70% of its conjuncts in this way. What MURPHY Does Grabbing A Visually Presented Target Once MURPHY has learned to image its arm in an arbitrary joint configuration, it can use heuristic search to guide its arm "mentally" to a visually specified goal. Figure 3(a) depicts a hand trajectory from an initial position to the location of a visually presented target. Each step in the trajectory represents the position of the hand (large blob) at an intermediate joint configuration. MURPHY can visually evaluate the remaining distance to the goal at each position and use best-first search to reduce the distance. Once a complete trajectory has been found, MURPHY can move its arm to the goal in a single physical step, dispensing with all backtracking dead-ends, and other wasted operations (fig. 3(b? . It would also be possible to use the inverse model, i.e. the map from a desired visual into an internal joint image, to send the arm directly to its final position. Unfortunately, MURPHY has no means in its current early state of development to generate a full-blown 551 - _... - M1RPI!Y'. HInt&! Tra jeetory Figure 3: Grabbing a.n Object. (a) Irregular trajectory represents sequence of "mental" steps taken by MURPHY in attempting to "grab" a visually-presented target (shown in (b) as white cross). Mental image depicts MURPHY's arm in its final goal configuration, i.e. with hand on top of object. Coarse-coded joint activity is shown at right. (b) Having mentally searched a.nd found the target through a series of steps, MURPHY moves its arm phY6ically in a single step to the target, discarding the intermediate states of the trajectory that are not relevant in this simple problem. visual image of its arm in one of the final goal positions, of which there are many possible. Sending the tip of a robot arm to a given point in space is a classic task in robotics. The traditional approach involves first writing explicit kinematic equations for the arm based on the specific geometric details of the given arm. These equations take joint angles as inputs and produce manipUlator coordinates as outputs. In general, however, it is most often useful to specify the coordinates of the manipulator tip (i.e. its desired final position), and compute the joint angles necessary to achieve this goal. This involves the solution of the kinematic equations to generate an inverse kinematic model. Deriving such expressions has been called "the most difficult problem we will encounter" in vision-based robotics [27]. For this reason, it is highly desirable for a mobile agent to learn a model of its sensory-motor environment from scratch, in a way that depends little or not at all on the specific parameters of the motor apparatus, the sensory apparatus, or their mutual interactions. It is interesting to note that in this reaching task, MURPHY appears from the outside to be driven by an inverse kinematic model of its arm, since its first official act after training is to reach directly for a visually-presented object. While it is clear that best-first search is a weak method whose utility is limited in complex problem solving domains, it may be speculated that given the ability to rapidly image arm configurations, combined with a set of simple visual heuristics and various mechanism for escaping local minima (e.g. send the arm home), a number of more interesting visual planning problems may be within MURPHY's grasp, such as grabbing an object in the presence of obstacles. Indeed, for problems that are either difficult to invert, or for which the goal state is not fully known a priori, the technique of iterating a forward-going model has a long history (e.g. Newton's Method). 552 Imitating Autonomous Arm Activity A particularly interesting feature of "learning-by-doing" is that for every pair of unit populations present in the learning system, a mapping can be learned between them both backwards and forwards. Each such mapping may enable a unique and interesting kind of behavior. In MURPHY's case, we have seen that the mapping from a joint state to a visual image is useful for planning arm trajectories. The reverse mapping from a visual state to ajoint image has an altogether different use, i.e. that of "imitation". Thus, if MURPHY's arm is moved passively, the model can be used to ''follow'' the motion with an internal command (i.e. joint) trace. Or, if a substitute arm is positioned in MURPHY's visual field, MURPHY can "assume the position", i.e . imitate the model arm configuration by mapping the afferent visual state into a joint image, and using the joint image to move the arm. As of this writing, the implementation of this behavior is still somewhat unreliable. Discussion and Future Work In short, this work has been concerned with learning-by-doing in the domain of visionbased robotics. A number of features differentiate MURPHY from most other learning systems, and from other approaches to vision-based robotics: ? No intelligent teacher is needed to provide input-ouput pairs. MURPHY learns by exercising its repertoire of actions and learning the relationship between these actions and the sensory images that result. ? Mappings between populations of units, regardless of modality, can be learned in both directions simultaneously during exploratory behavior. Each mapping learned can support a distinct class of behaviorally useful functions. ? MURPHY uses its internal models to first solve problems "mentally". Plans can therefore be developed and refined before they are actually executed. ? By taking explicit advantage of continuity in the mappings between visual and joint spaces, and by using a variant of specific-instance learning in such a way as to allow generalization to novel inputs, MURHPY can learn "difficult" non-linear mappings with only a single layer of modifiable weights. Two future steps in this endeavor are as follows: ? Provide MURPHY with direction-selective visual and joint units both, so that it may learn to predict relationships between rates of change in the visual and joint domains. In this way, MURPHY can learn how to perturb its joints in order to send its hand in a particular direction, greatly reducing the current need to search for hand trajectories. ? Challenge MURPHY to grab actual objects, possibly in the presence of obstacles, where path of approach is crucial. The ability to readily envision intermediate arm configurations will become critical for such tasks. . Acknowledgements Particular thanks are due to Stephen Omohundro for his unrelenting scientific and moral support, and for suggesting vision and robotic kinematics as ideal domains for experimentation. 553 References [1] T.J. Sejnowski & C.R. Roaenberg, Complex Systems, 1, 145, (1969). [2] G.J. Tesauro & T.J. Sejnowski. A parallel network that leMIUI to play backgammon. Submitted for publication. [3] S. Grossberg, BioI. Cybern., f3, 187, (1976). [4] T. Kohonen, Self organization and auociati"e memory., (Springer-Verlag, Berlin 1984). [5] D.E. Rumelhart & D. Zipser. In Parallel diatri6uted proceuing: e:rplorationa in tA.e microatructure oj cognition, "01. 1, D.E. Rumelhart, J.L. McClelland, Eds., (Bradford: Cambridge, MA, 1986), p. 151. [6] R. Linsker, Proc. Natl. Acad. Sci., 83, 8779, (1986). [7] G.E. Hinton & J.L. McClelland. Learning representations by recirculation. Oral presentation, IEEE conference on Neural Information Processing Systems, Denver, 1987. [8] A.G. Barto, R.S. Sutton, & C.W. Anderson, IEEE Trans. on Sy ?. , Man, Cybern., amc-13, 834, (1983). [9] H. Ginsburg & S. Opper, Piaget'a tA.eor, oj intellectual de"elopment., (Prentice Hall, New Jersey, 1969). [10] J.D. Becker. In Computer modela Jor tA.ougA.t and language., R. Schank & K.M. Colby, Eds., (FreelllAIl, San Francisco, 1973). [11] R.L. Rivest & R.E. Schapire. In Proc. of the 4th into workshop on ma.ch.ine learning, 364-375, (1987). [12] J .G. Carbonell & Y. Gil. In Proc. of the 4th into workshop on machlne learning, 256-266, (1987). [13] K. Chen, Tech Report, Dept. of Computer Science, University of illinois, 1987. [14] A.G. Barto & R.S. Sutton, AFWAL-TR-81-1070, Avionics Laboratory, Air Force Wright Aeronautical Laboratories, Wright-Patterson AFB, Ohio 45433, 1981. [15] D.E. Rumelhart, "On learning to do what you want". Talk given at CMU Connectionist Summer School,1986a. [16] B.W. Melin Proc. of 8th Ann. Con!. of the Cognitive Science Soc., 562-571, (1986). [17] D.H. Ballard, G.E. Hinton, & T.J Sejnowski, Nature, 306, 21, (1983). [18] R.P. Erikson, American Scientist, May-June 1984, p. 233. [19] G.E. Hinton, J.L. McClelland, & D.E. Rumelhart. In Parallel diatri6uted proceuing: e:rplorationa in tA.e microatructure oj cognition, "01. 1, D.E. Rumelhart, J .L. McClelland, Eds., (Bradford, Cambridge, 1986), p. 77. [20] D.H. Ackley, G.E. Hinton, & T.J. Sejnowski, Cognitive Science, 9, 147, (1985). [21] G.E. Hinton & T .J. Sejnowski. In Parallel diatri6uted proceuing: e:rplorationa in tA.e microatructure oj cognition, "01. 1, D.E. Rumelhart, J.L. McClelland, Eds., (Bradford, Cambridge, 1986), p. 282. [22] G.J. Tes&llro, Complex Systems, 1,367, (1987). [23] S.E. Hampson & D.J. Volper, Biological Cybernetics, 56, 121, (1987). [24] D.E. Rumelhart, G.E. Hinton, & J.L. McClelland. In Parallel diatri6uted proceuing: e:rplorationa in tA.e microatructure oj cognition, "01. 1, D.E. Rumelhart, J.L. McClelland, Eds., (Bradford, Cambridge, 1986), p. 3. [25] R.S. Michalski, J.G. Carbonell, & T.M. Mitchell, Eds., MacA.ine learning: an artificial intelligence approacA., Vois. I and II, (Morgan KauflllAll, Los Altos, 1986). [26] S. Omohundro, Complex Systems, 1, 273, (1987). [27] Paul, R. R060t manipulatora: matA.ematica, programming, and control., (MIT Press, Cambridge, 1981).
92 |@word eor:1 hampson:3 nd:1 decomposition:1 tr:1 colby:1 moment:2 initial:1 configuration:21 series:2 efficacy:1 contains:1 tuned:4 envision:3 current:6 activation:5 conjunct:5 conjunctive:4 must:5 reminiscent:1 yet:2 realize:1 readily:1 happen:2 blur:2 shape:1 motor:9 designed:2 progressively:2 discrimination:1 infant:1 intelligence:2 fewer:2 implying:1 nervous:1 imitate:3 signalling:1 plane:1 affair:2 realism:1 short:1 record:1 mental:6 detecting:1 coarse:6 intellectual:1 location:3 sigmoidal:1 constructed:1 direct:1 become:2 supply:1 ouput:1 consists:3 behavioral:2 inside:1 recognizable:1 acquired:1 mask:1 indeed:1 behavior:6 planning:4 multi:1 relying:1 automatically:1 little:2 actual:2 inappropriate:1 elbow:3 increasing:1 begin:1 retinotopic:1 insure:1 discover:1 rivest:1 alto:1 what:4 kind:4 developed:2 transformation:2 impractical:1 every:3 act:2 control:5 unit:73 internally:4 intervention:1 appear:1 before:1 scientist:1 local:1 apparatus:2 consequence:1 acad:1 sutton:3 encoding:2 proceuing:4 path:1 approximately:2 might:2 studied:1 suggests:1 ease:2 limited:1 projective:2 bi:1 range:3 directed:1 unique:1 camera:10 enforces:1 wrist:3 grossberg:1 practice:1 differs:1 xr:1 spot:8 procedure:2 thought:1 convenient:1 subpopulation:2 close:1 put:1 context:3 applying:1 collapsed:1 writing:2 cybern:2 prentice:1 map:1 demonstrated:1 center:6 send:3 regardless:1 automaton:1 focused:2 rectangular:2 rule:4 array:1 deriving:1 his:1 population:22 classic:1 exploratory:2 autonomous:2 justification:1 coordinate:2 target:5 play:1 programming:1 us:2 rumelhart:9 particularly:1 vein:1 bottom:1 ackley:1 region:8 autonomously:1 environment:5 complexity:3 reward:2 dynamic:1 ultimately:2 depend:1 solving:1 oral:1 patterson:1 completely:1 joint:70 represented:6 various:2 jersey:1 talk:1 train:1 distinct:1 activate:2 sejnowski:5 artificial:2 abridged:1 hyper:1 outside:1 refined:1 disjunction:1 whose:4 encoded:1 heuristic:4 solve:1 interconnection:1 encoder:1 ability:2 pertubation:1 final:5 associative:1 directionally:1 differentiate:1 sequence:3 advantage:4 blob:1 michalski:1 interconnected:2 interaction:2 ajoint:1 neighboring:3 relevant:8 kohonen:1 rapidly:2 translate:1 achieve:2 representational:3 ine:2 moved:2 competition:1 los:1 billion:2 cluster:2 extending:1 produce:2 wired:1 object:8 spent:1 school:2 soc:1 implemented:2 involves:2 come:1 memorize:1 direction:8 radius:3 correct:1 discontinuous:3 subsequently:1 exploration:1 human:1 centered:2 enable:1 exercising:1 education:1 argued:1 subdivided:1 require:1 generalization:11 repertoire:3 biological:4 connectionism:1 secondly:1 around:1 considered:1 sufficiently:1 wright:2 visually:8 hall:1 mapping:20 predict:3 cognition:4 sought:1 vary:1 early:1 purpose:2 proc:4 currently:3 grouped:1 combinatorially:1 weighted:4 mit:1 clearly:1 behaviorally:3 gaussian:1 rather:2 schematized:1 reaching:1 mobile:1 barto:3 command:4 publication:1 conjunction:4 encode:2 focus:1 june:1 articulator:1 backgammon:1 greatly:1 contrast:1 tech:1 sense:1 typically:1 entire:3 eliminate:1 initially:1 rhino:1 relation:1 going:2 selective:1 semantics:1 issue:1 among:4 priori:1 development:4 plan:1 raised:1 mutual:1 amc:1 field:17 once:4 never:1 equal:1 masse:1 having:1 f3:1 represents:3 unsupervised:3 inevitable:1 linsker:1 future:2 report:1 connectionist:7 stimulus:2 intelligent:5 serious:1 simplify:1 hint:1 simultaneously:3 recognize:5 murphy:47 intended:2 geometry:1 fire:2 explosive:1 attempt:1 organization:1 highly:7 kinematic:4 grasp:1 extreme:2 activated:13 natl:1 tuple:1 capable:2 worker:1 necessary:1 experience:1 penalizes:1 desired:3 stepper:1 recirculation:1 maca:1 instance:7 increased:1 boolean:4 typified:1 obstacle:2 visuotopically:1 subset:1 predicate:1 teacher:6 varies:1 combined:1 thanks:1 hyperrectangular:1 physic:2 tip:2 together:1 concrete:1 again:2 choose:1 possibly:1 cognitive:4 dead:1 american:2 style:1 til:1 suggesting:1 de:1 coding:4 tra:1 explicitly:2 afferent:2 depends:1 stream:1 later:1 view:2 observer:1 multiplicative:2 extrapolation:1 doing:11 competitive:1 sort:1 parallel:5 air:1 il:1 square:1 formed:1 percept:1 reinforced:1 sy:1 waist:1 weak:1 identification:1 thumb:1 nullified:1 produced:1 multiplying:1 trajectory:7 drive:1 finer:1 cybernetics:1 history:2 submitted:1 reach:1 synaptic:5 ed:6 sixth:1 acquisition:1 associated:1 attributed:2 hamming:1 con:1 modela:1 logical:2 mitchell:1 knowledge:2 color:1 organized:2 malfunction:1 cj:1 positioned:1 sophisticated:1 actually:1 back:2 appears:1 ta:6 supervised:1 follow:1 afb:1 response:2 maximally:1 specify:2 though:1 strongly:3 anderson:1 furthermore:2 until:1 hand:7 overlapping:3 propagation:2 incrementally:1 somehow:1 continuity:5 scientific:1 manipulator:2 effect:4 concept:1 contain:1 laboratory:2 illustrated:1 white:9 during:14 self:1 acquires:1 covering:1 mel:1 criterion:1 complete:1 omohundro:2 dedicated:1 motion:1 passive:1 image:20 novel:3 ohio:1 common:1 mentally:4 stimulation:1 physical:3 denver:1 avionics:1 association:1 discussed:2 cambridge:5 subserve:1 grid:4 pointed:1 illinois:2 nonlinearity:1 predictively:2 language:1 moving:2 robot:5 recent:1 perspective:2 irrelevant:2 driven:4 reverse:2 manipulation:1 tesauro:1 verlag:1 seen:5 minimum:1 additional:1 somewhat:2 morgan:1 paradigm:1 ii:2 stephen:1 multiple:1 full:3 needing:1 desirable:1 champaign:1 plausibility:1 cross:1 long:1 coded:7 variant:1 basic:1 multilayer:1 vision:3 cmu:1 represent:1 confined:1 robotics:4 irregular:1 invert:1 background:1 want:1 addressed:1 adhere:1 grabbing:4 crucial:2 modality:1 operate:1 volper:2 south:1 sent:1 zipser:1 presence:3 backwards:3 intermediate:3 visionbased:1 enough:1 concerned:2 easy:1 variety:1 affect:1 ideal:1 architecture:2 identified:1 escaping:1 reduce:1 tradeoff:3 computable:1 loint:1 unprocessed:2 inactive:1 expression:1 bartlett:1 utility:1 becker:1 effort:1 moral:1 penalty:1 speech:1 proceed:1 action:12 repeatedly:2 generally:2 useful:4 detailed:2 clear:1 iterating:1 amount:1 locally:1 situated:1 category:1 simplest:1 mcclelland:7 generate:3 datacube:1 schapire:1 notice:1 blown:1 gil:1 per:1 modifiable:3 vol:1 coarsely:4 thresholded:2 imaging:2 wasted:1 grab:2 aeronautical:1 sum:1 run:2 inverse:4 angle:12 fourth:2 you:1 place:2 architectural:1 home:1 scaling:1 bit:1 layer:4 summer:1 activity:7 strength:1 handful:1 throwing:1 software:1 auspex:1 calling:1 sake:1 nearby:3 speed:1 extremely:1 attempting:1 passively:1 optical:1 rendered:1 developing:1 according:1 wi:1 founder:1 modification:4 memorizing:1 imitating:1 taken:1 equation:3 mutually:1 discus:1 turn:3 kinematics:3 mechanism:2 needed:3 mind:1 vip:3 end:1 sending:1 piaget:3 operation:3 awaits:1 experimentation:1 away:1 generic:1 appropriate:1 ginsburg:1 alternative:1 encounter:1 altogether:1 substitute:1 top:2 clustering:1 remaining:1 paradoxical:1 newton:1 folklore:1 perturb:1 build:1 move:4 noticed:1 arrangement:1 question:1 receptive:9 traditional:3 responds:1 distance:5 separate:1 mapped:1 simulated:3 berlin:1 street:1 sci:1 carbonell:2 topic:1 evaluate:1 toward:1 reason:1 shifter:1 relationship:7 illustration:1 troubling:1 unfortunately:2 difficult:3 executed:1 expense:1 trace:9 sigma:3 design:1 implementation:1 perform:1 allowing:1 upper:1 wire:1 neuron:1 january:1 situation:1 hinton:6 looking:1 shoulder:3 frame:1 discovered:2 constructivist:1 arbitrary:2 pair:4 specified:1 immunity:1 connection:6 continous:1 recalled:1 learned:11 hour:1 trans:1 able:3 suggested:2 usually:1 exemplified:1 pattern:11 below:1 challenge:1 oj:5 memory:1 video:1 greatest:1 overlap:3 critical:1 difficulty:1 natural:1 force:1 participation:1 arm:42 representing:2 scheme:10 ready:1 review:2 geometric:1 acknowledgement:1 law:1 fully:1 interesting:4 sloppy:1 agent:3 degree:4 sufficient:2 thresholding:1 principle:1 leaning:1 critic:3 pi:3 compatible:1 jvc:1 parity:1 guide:1 allow:1 institute:1 fall:1 taking:2 dimension:2 opper:1 jor:1 rich:1 seemed:2 sensory:12 forward:6 commonly:1 reinforcement:3 author:1 made:1 stuck:2 san:1 compact:1 unreliable:1 orderly:1 global:5 active:10 incoming:2 robotic:2 assumed:1 francisco:1 xi:1 imitation:1 continuous:4 search:4 learn:11 nature:2 ballard:1 controllable:1 conjuncts:6 excellent:1 complex:8 necessarily:1 interpolating:1 domain:7 official:1 dense:1 whole:1 noise:1 paul:1 child:2 allowed:1 body:1 fig:3 representative:2 referred:1 en:1 depicts:4 fashion:2 dispensing:1 sub:1 position:9 comprises:1 explicit:4 decoded:1 exponential:1 exercise:1 perceptual:2 third:1 learns:6 specific:13 discarding:1 decay:1 consequent:1 exists:1 workshop:2 te:1 push:1 chen:1 suited:1 smoothly:1 backtracking:1 simply:4 explore:1 insures:2 visual:52 partially:1 speculated:1 acquiring:1 springer:1 ch:1 determines:1 constantly:1 ma:2 bioi:1 goal:8 endeavor:1 presentation:1 ann:1 absence:1 man:1 change:2 determined:1 specifically:1 typical:1 reducing:1 total:1 called:1 bradford:4 internal:7 searched:1 support:2 rotate:1 dept:1 outgoing:2 scratch:1
7,158
921
The NilOOO: High Speed Parallel VLSI for Implementing Multilayer Perceptrons Michael P. Perrone Thomas J. Watson Research Center P.O. Box 704 Yorktown Heights, NY 10598 mppGwatson.ibm.com Leon N Cooper Institute for Brain and Neural Systems Brown University Providence, Ri 02912 IncGcns.brown.edu Abstract In this paper we present a new version of the standard multilayer perceptron (MLP) algorithm for the state-of-the-art in neural network VLSI implementations: the Intel Ni1000. This new version of the MLP uses a fundamental property of high dimensional spaces which allows the 12 -norm to be accurately approximated by the It -norm. This approach enables the standard MLP to utilize the parallel architecture of the Ni1000 to achieve on the order of 40000, 256-dimensional classifications per second. 1 The Intel NilOOO VLSI Chip The Nestor/Intel radial basis function neural chip (Ni1000) contains the equivalent of 1024 256-dimensional artificial digital neurons and can perform at least 40000 classifications per second [Sullivan, 1993]. To attain this great speed, the Ni1000 was designed to calculate "city block" distances (Le. the II-norm) and thus to avoid the large number of multiplication units that would be required to calculate Euclidean dot products in parallel. Each neuron calculates the city block distance between its stored weights and the current input: neuron activity = w, L IWi - :eil (1) where is the neuron's stored weight for the ith input and :ei is the ith input. Thus the Nil000 is ideally suited to perform both the RCE [Reillyet al., 1982] and 748 Michael P. Perrone. Leon N. Cooper PRCE [Scofield et al., 1987] algorithms or any of the other commonly used radial basis function (RBF) algorithms. However, dot products are central in the calculations performed by most neural network algorithms (e.g. MLP, Cascade Correlation, etc.). Furthermore, for high dimensional data, the dot product becomes the computation bottleneck (i.e. most ofthe network's time is spent calculating dot products). If the dot product can not be performed in parallel there will be little advantage using the NilOOO for such algorithms. In this paper, we address this problem by showing that we can extend the NilOOO to many of the standard neural network algorithms by representing the Euclidean dot product as a function of Euclidean norms and by then using a city block norm approximation to the Euclidean norm. Section 2, introduces the approximate dot productj Section 3 describes the City Block MLP which uses the approximate dot productj and Section 4 presents experiments which demonstrate that the City Block MLP performs well on the NIST OCR data and on human face recognition data. 2 Approximate Dot Product Consider the following approximation [Perrone, 1993]: 1 11Z11 ~ y'n1Z1 (2) z where is some n-dimensional vector, II? II is the Euclidean length (i.e. the 12 norm) and I? I is the City Block length (i.e. the 11-norm). This approximation is motivated by the fact that in high dimensional spaces it is accurate for a majority of the points in the space. In Figure 1, we suggest an intuitive interpretation of why this approximation is reasonable. It is clear from Figure 1 that the approximation is reasonable for about 20% of the points on the arc in 2 dimensions. 1 As the dimensionality of the data space increases, the tangent region in Figure 1 expands asymptotically to fill the entire space and thus the approximation becomes more valid. Below we examine how accurate this approximation is and how we can use it with the NilOOO, particularly in the MLP context. Given a set of vectors, V, all with equal city block length, we measure the accuracy of the approximation by the ratio of the variance of the Euclidean lengths in V to the squared mean Euclidean lengths in V. If the ratio is low, then the approximation is good and all we must do is scale the city block length to the mean Euclidean length to get a good fit. 2 In particular, it can be shown that assuming all the vectors of the space are equally likely, the following equation holds [Perrone, 1993]: O'~ < (a~(!n+ 1) -1)ILfower, (3) where n is the dimension of the space; ILn is the average Euclidean length of the set of vectors with fixed city block length Sj O'~ is the variance about the average Euclidean length; ILlower is the lower bound for ILn and is given by ILlower == an S / Vnj 1 In fact, approximately 20% of the points are within 1% of each other and 40% of the points are within 5% of each other. 2 Note that in Equation 2 we scale by 1/ fo. For high dimensional spaces this is a good approximation to the ratio of the mean Euclidean length to the City Block length. VLSI for Implementing Multilayer Perceptrons 749 Figure 1: Two dimensional interpretation of the city block approximation. The circle corresponds to all of the vectors with the same Euclidean length. The inner square corresponds to all of the vectors with city block length equal the Euclidean length of the vectors in the circle. The outer square (tangent to the circle) corresponds to the set of vectors over which we will be making our approximation. In order to scale the outer square to the inner square, we multiple by 11 Vn where n is the dimensionality of the space. The outer square approximates the circle in the regions near the tangent points. In high dimensional spaces, these tangent regions approximate a large portion of the total hypersphere and thus the city block distance is a good approximation along most of the hypersphere. and an is defined by an - nn = --1 +-1 J1+ 2?r( -2 ~ +-nen- 1) (n) n+2 1 . (4) From this equation we see that the ratio of O"~ to /-L~ower in the large n limit is bounded above by 0.4. This bound is not very tight due to the complexity of the calculations required; however Figure 3 suggests that a much tighter bound must exist. A better bound exists if we are willing to add a minor constraint to our high dimensional space [Perrone, 1993]. In the case in which each dimension of the vector is constrained such that the entire vector cannot lie along a single axis, 3 we can show that 0"2 :::::: 2(n -1) ( ~ _ 1)2 /-L~ower (5) n (n + 1) 2 S a~ , V where S is the city block length of the vector in question. Thus in this case, the ratio of O"~ to /-L~ower decreases at least as fast as lin since nl S will be some fixed constant independent of n. 4 This dependency on nand S is shown in Figure 2. This result suggests that the approximation will be very accurate for many real-world pattern 3For example, when the axes are constrained to be in the range [D, 1] and the city block length of the vector is greater than 1. Note that this is true for the majority of the points in a n dimensional unit hypercube. ~Thus the accuracy improves as S increases towards its maximum value. 750 Michael P. Perrone, Leon N. Cooper recognition tasks such as speech and high resolution image recognition which can typically have thousand or even tens of thousands of dimensions. 1 0.8 SIn - 0.025 SIn - 0.05 SIn - 0.1 SIn - 0 .2 SIn - 0 .3 0.6 0.4 0 .2 ,----.---~~:~?~~~::~~ --------o L\.--__ :? ?~~:2~:~ ??--u -- ~--?_-?_??_-?_ ? -~--_?-_ . ._-._-._ . ..~.-~.-~-._ ..----------------_.. ~ .. -_.._.._._--_.. ~.-._.-_ : :_~~:-~.:_~_~-_:;~~_ : :~~:~~~-:~~:~ --~7 o 100 200 300 400 500 Figure 2: Plot of unj I-'lower vs. n for constrained vectors with varying values of Sin. As S grows the ratio shrinks and consequently, accuracy improves. If we assume that all of the vectors are uniformly distributed in an n-dimensional unit hypercube, it is easy to show that the average city block length is nj2 and the variance of the city block length is n/12. Since Sjn will generally be within one standard deviation ofthe mean, we find that typically 0.2 < Sjn < 0.8. We can use the same analysis on binary valued vectors to derive similar results. We explore this phenomenon further by considering the following Monte Carlo simulation. We sampled 200000 points from a uniform distribution over an n-dimensional cube. The Euclidean distance of each of these points to a fixed corner of the cube was calculated and all the lengths were normalized by the largest possible length, ~. Histograms of the resulting lengths are shown in Figure 3 for four different values of n. Note that as the dimension increases the variance about the mean drops. From Figure 3 we see that for as few as 100 dimensions, the standard deviation is approximately 5% of the mean length. 3 The City Block MLP In this section we describe how the approximation explained in Section 2 can be used by the NilOOO to implement MLPs in parallel. Consider the following formula for the dot product (6) VLSI for Implementing Multilayer Perceptrons 751 0.45 0 . 4 0.35 g ~ ~ooo D~mQg~ons ~ 100 10 :2 D.1.mes.1.ons D;l.mas1.ons D.1.mas1.ons - ..... -- -c:.--- ...... - .- 0.3 0.25 {} e 0.2 t>. 0 . 15 0.1 0.05 0.5 0.45 0.55 0.6 Norma1izad 0.65 LenQth 0.7 0.75 O.B Figure 3: Probability distributions for randomly draw lengths. Note that as the dimension increases the variance about the mean length drops. where II? II is the Euclidean length (i.e. 12-norm).5 Using Equation 2, we can approximation Equation 6 by ...... Y ~ - 1(1"';;'12 :t! + YI 4n :t! ? I'":t! - ;;'12) YI (7) where n is the dimension of the vectors and I . I is the city block length. The advantage to the approximation in Equation 7 is that it can be implemented in parallel on the Ni1000 while still behaving like a dot product. Thus we can use this approximation to implement MLP's on an Ni1000. The standard functional form for MLP's is given by [Rumelhart et al., 1986] N h(:t!;a,f3) d = u(aok + Lajk u (!30j + Lf3ij:t!i?) j=1 (8) i=1 were u is a fixed ridge function chosen to be u(:t!) = (1 + e N is the number of hidden units; k is the class label; d is the dimensionality of the data space; and a and f3 are adjustable parameters. The alternative which we propose, the City Block MLP, is given by [Perrone, 1993] -:t) 1 N gk(:t!; a, f3) = u(aok + d 1 - \ d L ajk u (f30j + 4(L lf3ij + :t!i1)2 - 4(L lf3ij -:t!i 1)2?) n i=1 j=1 n i=1 riNote also that depending on the information available to us, we could use either i? y= }(IIi + yW -lIiW -11?7W) or i? y= }<lIiIl2 + IIYlI2 -Iii - Y112). (9) 752 Michael P. Perrone, Leon N. Cooper DATA SET Faces Numbers Lowercase Uppercase HIDDEN UNITS 12 10 20 20 STANDARD CITYBLOCK % CORRECT % CORRECT 94.6?1.4 98.4?0.17 88.9?0.31 90.5?0.39 92.2?1.9 97.3?0.26 84.0?0.48 85.6?O.89 ENSEMBLE CITYBLOCK 96.3 98.3 88.6 90.7 Table 1: Comparison of MLPs classification performance with and with out the city block approximation to the dot product. The final column shows the effect of function space averaging. where the two city block calculation would be performed by neurons on the Nil000 chip. 6 The City Block MLP learns in the standard way by minimizing the mean square error (MSE), (10) MSE = (glc (Xi; 0:, (3) - tlci) 2 2: ilc where tic; is the value of the data at Xi for a class k. The MSE is minimized using the backpropagation stochastic gradient descent learning rule [Werbos, 1974]: For a fixed stepsize f'/ and each k, randomly choose a data point Xi and change 'Y by the amount A _ o(MSEi) L.l.'Y - -f'/ o'Y , (11) where 'Y is either 0: or {3 and MSEi is the contribution to the MSE of the ith data point. Note that although we have motivated the City Block MLP above as an approximation to the standard MLP, the City Block MLP can also be thought of as special case of radial basis function network. 4 Experimental Results This section describes experiments using the City Block MLP on a 120-dimensional representation of the NIST Handwritten Optical Character Recognition database and on a 2294-dimensional grayscale human face image database. The results indicate that the performance of networks using the approximation is as good as networks using the exact dot product [Perrone, 1993]. In order to test the performance of the City Block MLP, we simulated the behavior of the NilOOO on a SPARC station in serial. We used the approximation only on the first layer of weights (i.e. those connecting the inputs to the hidden units) where the dimensionality is highest and the approximation is most accurate. The approximation was not used in the second layer of weights (i.e. those connecting the hidden units to the output units were calculated in serial) since the number of hidden units was low and therefore do not correspond to a major computational bottleneck. It should be noted that for a 2 layer MLP in which the number of hidden units and output units are much lower than the input dimensionality, the 6The dot product between the hidden and the output layers may also be approximated in the same way but it is not shown here. In fact, the NilOOO could be used to perform all of the functions required by the network. VLSI for Implementing Multilayer Perceptrons DATA SET Numbers Lowercase Uppercase HIDDEN UNITS 10 20 20 STANDARD FOM 92.1?0.57 59.7?1.7 60.0?1.8 753 CITYBLOCK FOM 87.4?0.83 44.4?2.0 44.6?4.5 ENSEMBLE CITYBLOCK 92.5 62.7 66.4 Table 2: Comparison of MLPs FOM. The FOM is defined as the 100 minus the number rejected minus 10 time the number incorrect . majority of the computation is in the calculation of the dot products in the first weight layer. So even using the approximation only in the first layer will significantly accelerate the calculation. Also, the Nil000 on-chip math coprocessor can perform a low-dimensional, second layer dot product while the high-dimensional, first layer dot product is being approximated in parallel by the city block units. In practice, if the number of hidden units is large, the approximation to the dot product may also be used in the second weight layer. In the simulations, the networks used the approximation when calculating the dot product only in the feedforward phase of the algorithm. For the feedback ward phase (i.e. the error backpropagation phase), the algorithm was identical to the original backward propagation algorithm. In other words the approximation was used to calculate the network activity but the stochastic gradient term was calculated as if the network activity was generated with the real dot product. This simplification does not slow the calculation because all the terms needed for the backpropagation phase are calculated in the forward propagation phase In addition, it allows us to avoid altering the backpropagation algorithm to incorporate the derivative of the city block approximation. We are currently working on simulations which use city block calculations in both the forward and backward passes. Since these simulations will use the correct derivative for the functional form of the City Block MLP, we expect that they will have better performance. In practice, the price we pay for making the approximation is reduced performance. We can avoid this problem by increasing the number of hidden units and thereby allow more flexibility in the network. This increase in size will not significantly slow the algorithm since the hidden unit activities are calculated in parallel. In Table 1 and Table 2, we compare the performance of a standard MLP without the city block approximation to a MLP using the city block approximation to calculate network activity. In all cases, a population of 10 neural networks were trained from random initial weight configurations and the means and standard deviations were listed. The number of hidden units was chosen to give a reasonable size network while at the same time reasonably quick training. Training was halted by cross-validating on an independent hold-out set. From these results, one can see that the relative performances with and with out the approximation are similar although the City Block is slightly lower. We also perform ensemble averaging [Perrone, 1993, Perrone and Cooper, 1993] to further improve the performance of the approximate networks. These results are given in the last column of the table. From these data we see that by combining the city block approximation with the averaging method, we can generate networks which have comparable and sometimes better performance than the standard MLPs. In addition, because the Nil000 is running in parallel, there is minimal additional computational overhead for using 754 Michael P. Perrone, Leon N. Cooper the averaging. 7 5 Discussion We have described a new radial basis function network architecture which can be used in high dimensional spaces to approximate the learning characteristics of a standard MLP without using dot products. The absence of dot products allows us to implement this new architecture efficiently in parallel on an NilOOO; thus enabling us to take advantage of the Ni1000's extremely fast classification rates. We have also presented experimental results on real-world data which indicate that these high classifications rates can be achieved while maintaining or improving classification accuracy. These results illustrate that it is possible to use the inherent high dimensionality of real-world problems to our advantage. References [Perrone, 1993] Perrone, M. P. (1993). Improving Regression Estimation: Averaging Methods for Variance Reduction with Eztensions to General Convez Measure Optimization. PhD thesis, Brown University, Institute for Brain and Neural Systems; Dr. Leon N Cooper, Thesis Supervisor. [Perrone and Cooper, 1993] Perrone, M. P. and Cooper, L. N. (1993). When networks disagree: Ensemble method for neural networks. In Mammone, R. J., editor, Artificial Neural Networks for Speech and Vision. Chapman-Hall. Chapter 10. [Reillyet al., 1982] Reilly, D. L., Cooper, L. N., and Elbaum, C. (1982). A neural model for category learning. Biological Cybernetics, 45:35-41. [Rumelhart et al., 1986] Rumelhart, D. E., McClelland, J. L., and the PDP Research Group (1986). Parallel Distributed Processing, Volume 1: Foundations. MIT Press. [Scofield et al., 1987] Scofield, C. L., Reilly, D. L., Elbaum, C., and Cooper, L. N. (1987). Pattern class degeneracy in an unrestricted storage density memory. In Anderson, D. Z., editor, Neural Information Processing Systems. American Institute of Physics. [Sullivan, 1993] Sullivan, M. (1993). Intel and Nestor deliver second-generation neural network chip to DARPA: Companies launch beta site program. Intel Corporation News Release. Feb. 12. [Werbos, 1974] Werbos, P. (1974). Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. PhD thesis, Harvard University. 7The averaging can also be applied to the standard MLPs with a corresponding improvement in performance. However, for serial machines averaging slows calculations by a factor equal to the number of averaging nets.
921 |@word coprocessor:1 version:2 norm:9 willing:1 simulation:4 thereby:1 minus:2 reduction:1 initial:1 configuration:1 contains:1 current:1 com:1 must:2 j1:1 enables:1 designed:1 plot:1 drop:2 v:1 ith:3 hypersphere:2 math:1 height:1 along:2 beta:1 incorrect:1 overhead:1 behavioral:1 behavior:1 examine:1 brain:2 eil:1 company:1 little:1 considering:1 increasing:1 becomes:2 bounded:1 tic:1 elbaum:2 corporation:1 expands:1 unit:17 convez:1 limit:1 approximately:2 suggests:2 range:1 practice:2 block:35 implement:3 backpropagation:4 sullivan:3 attain:1 cascade:1 thought:1 significantly:2 word:1 reilly:2 radial:4 suggest:1 get:1 cannot:1 storage:1 context:1 equivalent:1 quick:1 center:1 resolution:1 rule:1 fill:1 glc:1 population:1 exact:1 us:2 harvard:1 rumelhart:3 approximated:3 recognition:4 particularly:1 werbos:3 database:2 calculate:4 thousand:2 region:3 news:1 decrease:1 highest:1 complexity:1 ideally:1 trained:1 tight:1 deliver:1 basis:4 accelerate:1 darpa:1 chip:5 chapter:1 fast:2 describe:1 monte:1 artificial:2 mammone:1 valued:1 ward:1 final:1 advantage:4 net:1 propose:1 product:20 combining:1 flexibility:1 achieve:1 intuitive:1 spent:1 derive:1 depending:1 illustrate:1 minor:1 implemented:1 launch:1 indicate:2 correct:3 stochastic:2 human:2 implementing:4 tighter:1 biological:1 hold:2 hall:1 great:1 major:1 estimation:1 label:1 currently:1 largest:1 city:35 tool:1 mit:1 avoid:3 varying:1 ax:1 release:1 improvement:1 lowercase:2 nn:1 entire:2 typically:2 nand:1 hidden:12 vlsi:6 i1:1 classification:6 art:1 constrained:3 special:1 cube:2 equal:3 f3:3 chapman:1 identical:1 minimized:1 inherent:1 few:1 randomly:2 nestor:2 phase:5 mlp:22 introduces:1 nl:1 uppercase:2 accurate:4 euclidean:15 circle:4 minimal:1 vnj:1 column:2 halted:1 altering:1 deviation:3 uniform:1 supervisor:1 stored:2 providence:1 dependency:1 density:1 fundamental:1 physic:1 michael:5 connecting:2 squared:1 central:1 thesis:3 choose:1 dr:1 corner:1 american:1 derivative:2 performed:3 portion:1 parallel:11 iwi:1 contribution:1 mlps:5 square:6 accuracy:4 variance:6 characteristic:1 efficiently:1 ensemble:4 correspond:1 ofthe:2 handwritten:1 accurately:1 carlo:1 cybernetics:1 fo:1 degeneracy:1 sampled:1 dimensionality:6 improves:2 ooo:1 box:1 shrink:1 anderson:1 furthermore:1 rejected:1 correlation:1 working:1 ei:1 propagation:2 grows:1 effect:1 brown:3 true:1 normalized:1 sin:6 iln:2 noted:1 yorktown:1 ridge:1 demonstrate:1 performs:1 ilc:1 image:2 functional:2 volume:1 extend:1 interpretation:2 approximates:1 cityblock:4 dot:22 behaving:1 etc:1 add:1 feb:1 sparc:1 binary:1 watson:1 yi:2 unrestricted:1 greater:1 additional:1 ii:5 multiple:1 z11:1 calculation:8 cross:1 lin:1 serial:3 equally:1 calculates:1 prediction:1 regression:2 multilayer:5 vision:1 histogram:1 sometimes:1 achieved:1 addition:2 pass:1 validating:1 nj2:1 near:1 feedforward:1 iii:2 easy:1 fit:1 architecture:3 inner:2 y112:1 bottleneck:2 motivated:2 speech:2 generally:1 clear:1 yw:1 listed:1 amount:1 ten:1 category:1 mcclelland:1 reduced:1 generate:1 exist:1 per:2 group:1 four:1 utilize:1 backward:2 asymptotically:1 reasonable:3 vn:1 draw:1 comparable:1 bound:4 layer:9 pay:1 simplification:1 activity:5 constraint:1 ri:1 speed:2 extremely:1 leon:6 optical:1 perrone:16 describes:2 slightly:1 character:1 making:2 explained:1 equation:6 fom:4 needed:1 available:1 nen:1 ocr:1 stepsize:1 alternative:1 thomas:1 original:1 running:1 maintaining:1 rce:1 calculating:2 hypercube:2 question:1 gradient:2 distance:4 simulated:1 majority:3 outer:3 me:1 assuming:1 length:27 ratio:6 minimizing:1 gk:1 slows:1 implementation:1 adjustable:1 perform:5 disagree:1 neuron:5 arc:1 nist:2 enabling:1 descent:1 pdp:1 station:1 required:3 address:1 beyond:1 below:1 pattern:2 program:1 memory:1 representing:1 improve:1 axis:1 tangent:4 multiplication:1 relative:1 expect:1 lenqth:1 generation:1 digital:1 foundation:1 editor:2 ibm:1 last:1 scofield:3 allow:1 perceptron:1 institute:3 face:3 distributed:2 feedback:1 dimension:8 calculated:5 valid:1 world:3 forward:2 commonly:1 sj:1 approximate:6 ons:4 xi:3 grayscale:1 why:1 table:5 reasonably:1 improving:2 mse:4 sjn:2 site:1 intel:5 ny:1 cooper:11 slow:2 lie:1 learns:1 formula:1 showing:1 exists:1 ower:3 phd:2 suited:1 likely:1 explore:1 corresponds:3 consequently:1 rbf:1 towards:1 price:1 ajk:1 absence:1 change:1 uniformly:1 averaging:8 total:1 experimental:2 perceptrons:4 incorporate:1 phenomenon:1
7,159
922
PCA-Pyramids for Image Compression* Horst Bischof Department for Pattern Recognition and Image Processing Technical University Vienna Treitlstraf3e 3/1832 A-1040 Vienna, Austria bis@prip.tuwien.ac.at Kurt Hornik Institut fur Statistik und Wahrscheinlichkeitstheorie Technische UniversiUit Wien Wiedner Hauptstraf3e 8-10/1071 A-1040 Vienna, Austria Kurt.Hornik@ci .tuwien.ac.at Abstract This paper presents a new method for image compression by neural networks. First, we show that we can use neural networks in a pyramidal framework, yielding the so-called PCA pyramids. Then we present an image compression method based on the PCA pyramid, which is similar to the Laplace pyramid and wavelet transform. Some experimental results with real images are reported. Finally, we present a method to combine the quantization step with the learning of the PCA pyramid. 1 Introduction In the past few years, a lot of work has been done on using neural networks for image compression, d . e.g. (Cottrell et al., 1987; Sanger, 1989; Mougeot et al., 1991; Schweizer et al., 1991)). Typically, networks which perform a Principal Component Analysis (PCA) were employed; for a recent overview of PCA networks, see (Baldi and Hornik, 1995). A well studied and thoroughly understood PCA network architecture is the linear autoassociative network, see (Baldi and Hornik, 1989; Bourlard and Kamp, 1988). This network consists of N input and output units and M < N hidden units, and is *This work was supported in part by a grant from the Austrian National Fonds zur Forderung der wissenschaftlichen Forschung (No. S7002MAT) to Horst Bischof. 942 Horst Bischof, Kurt Hornik trained (usually by back-propagation) to reproduce the input at the output units. All units are linear. Bourlard & Kamp (Bourlard and Kamp, 1988) have shown that at the minimum of the usual quadratic error function ?, the hidden units project the input on the space spanned by the first M principal components of the input distribution. In fact, as long as the output units are linear, nothing is gained by using non-linear hidden units. On average, all hidden units have equal variance. However, peA is not the only method for image compression. Among many others, the Laplace Pyramid (Burt and Adelson, 1983) and wavelets (Mallat, 1989) have successfully been used to compress images. Of particular interest is the fact that these techniques provide a hierarchical representation of the image which can be used for progressive image transmission. However, these hierarchical methods are not adaptive. In this paper, we present a combination of autoassociative networks with hierarchical methods. We propose the so-called peA pyramids, which can be seen as an extension of image pyramids with a learning algorithm as well as cascaded locally connected autoassociative networks. In other words, we combine the structure of image pyramids and neural network learning algorithms, resulting in learning pyramids. The structure of this paper is as follows . We first present image pyramids and, in particular, the peA pyramid. Then, we discuss how these pyramids can be used for image compression, and present some experimental results. Next, we discuss a method to combine the quantization step of compression with the transformation. Finally, we give some conclusions and an outline of further research. 2 The peA Pyramid Before we introduce the peA pyramid, let us describe regular image pyramids. For a discussion of irregular pyramids and their relation to neural networks, see (Bischof, 1993). In the simplest case, each successive level ofthe pyramid is obtained from the previous level by a filtering operation followed by a sampling operator. More general functions can be used to achieve the desired reduction. We therefore call them reduction functions. The structure of a pyramid is determined by the neighbor relations within the levels of the pyramid and by the "father-son" relations between adjacent levels. A cell (if it is not at the base level) has a set of children (sons) at the level directly below which provide input to the cell, a set of neighbors (brothers/sisters) at the same level, and (if it is not the apex of the pyramid) a set of parents (fathers) at the level directly above. We denote the structure of a (regular) pyramid by the expression n x nlr, where n x n (the number of sons) is the size of the reduction window and r the reduction factor which describes how the number of cells decreases from level to level. 2.1 peA Pyramids Since a pyramid reduces the information content of an image level by level, an objective for the reduction function would be to preserve as much information as possible, given the restrictions imposed by the structure of the pyramid, or equivalently, to minimize the information loss by the reduction function. This naturally PCA-Pyramids for Image Compression 943 leads to the idea of representing the pyramid by a suitable peA network. Among the many alternatives for such networks, we have chosen the autoassociative networks for two reasons. First, the analysis of Hornik & Kuan (Hornik and Kuan, 1992) shows that these networks are more stable than competing models. Second, autoassociative networks have the nice feature that they automatically provide us with the expansion function (weights from the hidden layer to output layer). Since the neural network should have the same connectivity as the pyramid (i.e., the same father-son relations), its topology is determined by the structure of the pyramid. In this paper, we confine ourselves to the 4 x 4/4 pyramid for two reasons. First, the 4 x 4/4 pyramid has the nice property that every cell has the same number of fathers, which results in homogeneous networks. Second, as experiments have shown (Bischof, 1993) the results achieved with this pyramid are similar to other structures, e.g. the 5 x 5/4 pyramid, using fewer weights. E R L-..J rn Ii. = E(I n+l ) = E(R(ln? (a) General Setting (b) 4/2 pyramid (c) Corresponding network Figure 1: From the structure of the pyramid to the topology of the network Figure 1 depicts the one-dimensional situation of a 4/2 pyramid (this is the onedimensional counterpart of the two-dimensional 4 x 4/2 pyramid). Figure 1a shows the general goal to be achieved and the notations employed; Figure 1b shows a 4/2 pyramid. When constructing the corresponding network, we start at the output layer (Le., I~). For an n/r pyramid we typically choose the size of the output layer as n. Next, we have to include all fathers of the cells in the output layer as hidden units. Finally, we have to include all sons of the hidden layer cells in the input layer. For the 4/2 pyramid, this results in an 8-3-4 network as shown in Figure 1c. A similar construction yields an 8 x 8-3 x 3-4 x 4 network for the 4 x 4/4 pyramid. The next thing to consider are the constraints on the network weights due to the overlaps in the pyramid. To completely cover the input image with output units, we can shift the network only by four cells in each direction. Therefore, the hidden units at the borders overlap. For the 4/2 pyramid, the left and right hidden units must have identical weights. In the case of the 4 x 4/4 pyramid, the network has four independent units. The thus constructed network can be trained by some suitable learning algorithm, typically of the back-propagation type, using batches of an image as input for trai- 944 Horst Bischof, Kurt Hornik ning the first pyramid level. After that, the second level of the pyramid can be trained in the same way using the first pyramid level as training data, and so on. 2.2 PeA-Laplace Pyramid and Image Compression Thus far, we have introduced a network which can learn the reduction function R and the expansion function E of a pyramid. Analogously to the Laplace pyramid and the wavelet transform we can now introduce the level Li of the PCA-Laplace pyramid, given by Li = Ii - I: = Ii - E(R(Ii)) It should be noted that during learning we exactly minimize the squared Laplace (a) First 2 levels of a Laplace pyramid (upper half) and peA-Laplace pyramid (lower half) (grey = 0) (9) Reconstruction error of house image with quantization of 3 bits, 4 bits, 7 bits, and reconstructed image Figure 2: Results of PCA-Laplace-Pyramid level. The original image 10 can be completely recovered from level In and the Laplace levels L o, ... ,Ln - 1 by 10 = E(??? E(E(In) + L n- 1 ) + L n- 2 )???) + Lo? Since the level In is rather small (e.g., 32 x 32 pixels) and the levels of the PCALaplace pyramid are typically sparse (i.e., many pixels are zero, see Figure 2a) and can therefore be compressed considerably by a conventional compression algorithm PCA-Pyramids for Image Compression 945 (e.g. Lempel-Ziv (Ziv and Lempel, 1977)), this image representation results in a lossless image compression algorithm . In order to achieve higher compression ratios we can quantize the levels of the PCALaplace pyramid. In this case, the compression is lossy, because the original image cannot be recovered exactly. The compression ratio and the amount of loss can be controlled by the number of bits used to quantize the levels of the PCA-Laplacian. To measure the difference between the compressed and the original image, we use the normalized mean squared error (NMSE) as in (Cottrell et al., 1987; Sanger, 1989) . The NMSE is given by the mean squared error divided by the average squared intensity of the image, i.e., NMSE = MSE = ((10 - C(10))2) (I~) (I~)' where 10 and C(lo) are the original and the compressed image, respectively. The compression ratio is measured by the amount of bits used to store 10 , divided by the amount of bits used to store C(1o). 2.3 Results For the results reported here we trained the networks by a conjugate gradient algorithm for 100 steps! and used a uniform quantization which is fixed for all levels of the pyramid. As was shown in (Burt and Adelson, 1983; Mayer and Kropatsch, 1989), the results could be improved by gradually increasing the quantization from bottom to top. Figure 2b shows the error images when the levels of the PCA-Laplacian pyramid are quantized with 3, 4, and 7 bits and the reconstructed image from the 7 bit Laplacian. Note that we used the same lookup-table for the error images. To compress the levels of the PCA-Laplacian pyramid, we employed the standard UNIX compress program which implements a Lempel-Ziv algorithm. iFrom these images one can see that the results with the 4 and 7 bit quantization are very good. Visually, no difference between the reconstructed and the original image can be perceived. Table 1 shows the compression ratios and the NMSEs on these images. We have performed experiments on 20 different images, the results on these images are comparable to the ones reported here. These results compare favorably with the results in the literature (see Table 1). We have also applied a 5 x 5/4 Laplace pyramid to the house image which gave a compression ratio of 3.42 with an NMSE of 0.000087 for quantization with four bits of the Laplace levels. We have also included results achieved with JPEG. One can see that our method gives considerably better results . We have also demonstrated experimentally what happens if we train a pyramid on one image and then apply this pyramid to another image without retraining. These experiments indicate that the errors are only a little bit larger for images not trained on. With five additional steps of training the errors are almost the same. iFrom lIn all our experiments the training algorithm converged (i.e. usually after 200 steps, however the improvements between steps 20 and convergence are negligible). 946 Horst Bischof, Kurt Hornik Quant. 3 Bit 4 Bit 7 Bit no Quant. Cottrell,~Cottrell et al., 1987) Sanger (Sanger, 1989) 5 x 5/4 Laplace JPEG JPEG Compression ratio 37.628 24.773 8.245 3.511 8.0 22.0 3.420 8.290 15.774 Bits/Pixel 0.212 0.323 0.970 2.279 1.000 0.360 2.339 0.965 0.507 NMSE 0.0172 0.0019 0.0000215 0.0 0.0059 0.043 0.000087 0.00139 0.00348 Table 1: Compression ratios and NMSE for various compression methods this results we can conclude that we do not need to retrain the pyramid for each new image. 3 Integration of Quantization For the results reported in the previous section we have used a fixed and uniform quantization scheme which can be improved by using adaptive quantizers like the Lloyd I algorithm, Kohonen's Feature Maps, learning vector quantization, or something similar. Such an approach as taken by Schweizer (Schweizer et al., 1991) who combined a Cottrell-type network with self-organizing feature maps. However, we can go further. With the PCA network we minimize the squared Laplace level which does not necessarily yield low compression errors. What we really want to minimize are the quantized Laplace levels. Usually, the Laplace levels have an unimodally shaped histogram centered at zero. However, for the result of the compression (i.e., compression ratio and NMSE), it is irrelevant if we shift the histogram to the left or the right as long as we shift the quantization intervals in the same way. The best results could be achieved if we have a multimodal histogram with peaks centered at the quantization points. Using neural networks for both PCA and quantization, this goal could e.g. be achieved by a modular network as in Figure 3 for the 4/2 pyramid. For quantization, we could either apply a vector quantizer to a whole patch of the Laplace level, or use a scalar quantizer (as depicted in Figure 3) for each pixel of the Laplace level. In the second case, we have to constrain the weights of the quantization network to be identical for every Laplace pixel. Since scalar quanti.zation is simpler to analyze and uses less free parameters, we only consider this case. As each quantization subnetwork can be treated separately (we only have to average the weight changes over all subnetworks), the following only considers the case of one output unit of the PCA network. PCA-Pyramids for Image Compression Quantization 947 peA Figure 3: PCA network and Quantization network The error to be minimized is the squared quantization error where p refers to the patterns in the training set, Ck is the kth weight of the quantization network, and 1 is the output of the PCA-Laplace unit. Changing the weights of the quantization network by gradient descent leads to the LVQl rule of Kohonen Ac - { 2a(lp k 0, Ck), if k = kp is the winning unit, otherwise. For the PCA network we can proceed similarly to back-propagation to obtain the rule AWij = -K 8Ep = _K 8Ep 8Wij alp = _K 8Ep alp 8Wij alp 8~p 8i~ 8z~ 8Wij = -2K(lp _ Ck) 8i~ . 8Wij Of course, this is only one out of many possible algorithms. More elaborate minimization techniques than gradient descent could be used; similarly, LVQl could be replaced by a different quantization algorithm. But the basic idea of letting the quantization step and the the compression step adapt to each other remains unchanged. 4 Conclusions In this paper, we presented a new image compression scheme based on neural networks. The PCA and PCA-Laplace pyramids were introduced, which can be seen as both an extension of image pyramids to learning pyramids and as cascaded, locally connected autoassociators. The results achieved are promising and compare favorably to work reported in the literature. A lot of work remains to be done to analyze these networks analytically. The convergence properties of the PCA pyramid are not known; we expect results similar 948 Horst Bischof, Kurt Hornik to the ones (Baldi and Hornik, 1989) for the autoassociative network. Also, for the PCA network it would be desirable to characterize the features which are extracted. Similarly, the integrated network needs to be analyzed. It is clear that for such networks, the usual error function has local minima, but maybe they can be avoided by a proper training regime (i.e. start training the PCA pyramid, then train the vector quantizer, and finally train them together). References Baldi, P. and Hornik, K. (1989). Neural Networks and principal component analysis: Learning from examples without local minima. Neural Networks, 2:53-58. Baldi, P. and Hornik, K. (1995). Learning in Linear Neural Networks: a Survey. IEEE Transactions on Neural Networks, to appear. Bischof, H. (1993). Pyramidal Neural Networks. PhD thesis, TU-Vienna, Inst. f. Automation, Dept. f. Pattern Recognition and Image Processing. Bourlard, H. and Kamp, Y. (1988). Auto-Association by Multilayer Perceptrons and Singular Value Decomposition. Biological Cybernetics, 59:291-294. Burt, P. J. and Adelson, E. H. (1983). The Laplacian pyramid as a compact image code. IEEE Transactions on Communications, Vol. COM-31(No.4):pp.532-540. Cottrell, G., Munro, P., and Zipser, D. (1987). Learning Internal Representations from Grey-Scale Images: An Example of Extensional Programming. In Ninth Annual Conference of the Cognitive Science Society, pages 462-473. Hillsdale Erlbaum. Hornik, K. and Kuan, C. (1992). Convergence analysis of local feature extraction algorithms. Neural Networks, 5(2):229-240. Mallat, S. G. (1989). A Theory for Multiresolution Signal Decomposition: The Wavelet Representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. PAMI-ll(No. 7):pp. 674-693. Mayer, H. and Kropatsch, W. G. (1989). Progressive Bildubertragung mit der 3x3/2 Pyramide. In Burkhardt, H., H6hne, K., and Neumann, B., editors, Informatik Fachberichte 219: Mustererkennung 1989, pages 160-167, Hamburg. l1.DAGM - Symposium, Springer Verlag. Mougeot, M., Azencott, R., and Angeniol, B. (1991). Image Compression with Back Propagation: Improvement of the Visual Restoration using different Cost Functions. Neural Networks, 4:467-476. Sanger, T. (1989). Optimal Unsupervised learning in a Single-Layer Linear Feedforward Neural Network. Neural Networks, 2:433-459. Schweizer, L., Parladori, G., Sicranza, G., and Marsi, S. (1991). A fully neural approach to image compression. In Kohonen, T., Makissara, K., Simula, 0., and Kangas, J., editors, Artificial Neural Networks, volume I, pages 815-820. Ziv, J. and Lempel, A. (1977). A universal algorithm for sequential data compression. IEEE Trans. on Information Theory, 23(5):337 - 343.
922 |@word compression:30 retraining:1 grey:2 decomposition:2 awij:1 reduction:7 kurt:6 past:1 recovered:2 com:1 must:1 cottrell:6 extensional:1 half:2 fewer:1 intelligence:1 quantizer:3 quantized:2 successive:1 simpler:1 five:1 constructed:1 symposium:1 consists:1 combine:3 baldi:5 introduce:2 tuwien:2 automatically:1 little:1 window:1 increasing:1 project:1 notation:1 what:2 transformation:1 every:2 exactly:2 unit:16 grant:1 appear:1 before:1 negligible:1 understood:1 local:3 pami:1 studied:1 bi:1 implement:1 x3:1 universal:1 word:1 regular:2 refers:1 cannot:1 operator:1 restriction:1 conventional:1 imposed:1 demonstrated:1 map:2 go:1 survey:1 rule:2 spanned:1 laplace:21 construction:1 mallat:2 programming:1 homogeneous:1 us:1 simula:1 recognition:2 bottom:1 ep:3 connected:2 decrease:1 und:1 trained:5 completely:2 multimodal:1 various:1 train:3 describe:1 universiuit:1 kp:1 artificial:1 modular:1 larger:1 otherwise:1 compressed:3 transform:2 schweizer:4 kuan:3 propose:1 reconstruction:1 kohonen:3 tu:1 organizing:1 achieve:2 multiresolution:1 parent:1 convergence:3 transmission:1 neumann:1 ac:3 measured:1 indicate:1 direction:1 ning:1 pea:10 centered:2 alp:3 hillsdale:1 really:1 biological:1 extension:2 confine:1 visually:1 perceived:1 nlr:1 successfully:1 minimization:1 mit:1 rather:1 ck:3 ifrom:2 improvement:2 fur:1 inst:1 dagm:1 typically:4 integrated:1 hidden:9 relation:4 reproduce:1 wij:4 pixel:5 among:2 ziv:4 integration:1 equal:1 shaped:1 extraction:1 sampling:1 identical:2 progressive:2 adelson:3 unsupervised:1 minimized:1 others:1 few:1 preserve:1 national:1 replaced:1 ourselves:1 interest:1 analyzed:1 yielding:1 institut:1 desired:1 zation:1 wiedner:1 cover:1 jpeg:3 restoration:1 cost:1 technische:1 uniform:2 father:5 erlbaum:1 marsi:1 characterize:1 reported:5 autoassociators:1 considerably:2 combined:1 thoroughly:1 peak:1 analogously:1 together:1 connectivity:1 squared:6 thesis:1 nm:1 choose:1 wahrscheinlichkeitstheorie:1 cognitive:1 li:2 wien:1 lookup:1 lloyd:1 automation:1 performed:1 lot:2 analyze:2 start:2 minimize:4 variance:1 who:1 azencott:1 yield:2 ofthe:1 kamp:4 angeniol:1 informatik:1 cybernetics:1 converged:1 pp:2 naturally:1 austria:2 back:4 higher:1 improved:2 done:2 propagation:4 lossy:1 normalized:1 counterpart:1 analytically:1 mougeot:2 adjacent:1 ll:1 during:1 self:1 noted:1 outline:1 l1:1 image:51 overview:1 volume:1 association:1 onedimensional:1 similarly:3 apex:1 stable:1 base:1 something:1 recent:1 irrelevant:1 store:2 hamburg:1 verlag:1 der:2 seen:2 minimum:3 additional:1 employed:3 forderung:1 signal:1 ii:4 desirable:1 reduces:1 technical:1 adapt:1 long:2 lin:1 divided:2 controlled:1 laplacian:5 basic:1 austrian:1 multilayer:1 histogram:3 pyramid:72 achieved:6 cell:7 zur:1 irregular:1 want:1 separately:1 interval:1 pyramidal:2 singular:1 thing:1 prip:1 call:1 zipser:1 feedforward:1 gave:1 architecture:1 competing:1 topology:2 quant:2 idea:2 shift:3 expression:1 pca:26 munro:1 proceed:1 wissenschaftlichen:1 autoassociative:6 clear:1 burkhardt:1 maybe:1 amount:3 locally:2 simplest:1 brother:1 sister:1 vol:2 four:3 changing:1 year:1 unix:1 almost:1 patch:1 comparable:1 bit:15 layer:8 followed:1 quadratic:1 annual:1 constraint:1 constrain:1 statistik:1 department:1 combination:1 conjugate:1 describes:1 son:5 lp:2 happens:1 gradually:1 taken:1 ln:2 remains:2 discus:2 letting:1 subnetworks:1 operation:1 apply:2 hierarchical:3 lempel:4 alternative:1 batch:1 original:5 compress:3 top:1 include:2 vienna:4 sanger:5 society:1 unchanged:1 objective:1 usual:2 subnetwork:1 gradient:3 kth:1 considers:1 reason:2 code:1 ratio:8 equivalently:1 quantizers:1 favorably:2 proper:1 perform:1 upper:1 descent:2 situation:1 communication:1 rn:1 kangas:1 ninth:1 intensity:1 burt:3 introduced:2 mayer:2 bischof:9 trans:1 usually:3 pattern:4 below:1 regime:1 program:1 suitable:2 overlap:2 treated:1 bourlard:4 cascaded:2 representing:1 scheme:2 lossless:1 auto:1 nice:2 literature:2 loss:2 expect:1 fully:1 filtering:1 editor:2 lo:2 course:1 supported:1 free:1 neighbor:2 sparse:1 horst:6 adaptive:2 avoided:1 far:1 transaction:3 reconstructed:3 compact:1 conclude:1 table:4 promising:1 learn:1 hornik:14 expansion:2 quantize:2 mse:1 necessarily:1 constructing:1 quanti:1 border:1 whole:1 nothing:1 child:1 nmse:7 retrain:1 depicts:1 elaborate:1 winning:1 house:2 wavelet:4 trai:1 quantization:23 sequential:1 gained:1 ci:1 forschung:1 phd:1 fonds:1 depicted:1 visual:1 scalar:2 springer:1 extracted:1 goal:2 content:1 experimentally:1 change:1 included:1 determined:2 principal:3 called:2 experimental:2 perceptrons:1 internal:1 mustererkennung:1 dept:1
7,160
923
Learning Saccadic Eye Movements Using Multiscale Spatial Filters Rajesh P.N. Rao and Dana H. Ballard Department of Computer Science University of Rochester Rochester, NY 14627 {rao)dana}~cs.rochester.edu Abstract We describe a framework for learning saccadic eye movements using a photometric representation of target points in natural scenes. The representation takes the form of a high-dimensional vector comprised of the responses of spatial filters at different orientations and scales. We first demonstrate the use of this response vector in the task of locating previously foveated points in a scene and subsequently use this property in a multisaccade strategy to derive an adaptive motor map for delivering accurate saccades. 1 Introduction There has been recent interest in the use of space-variant sensors in active vision systems for tasks such as visual search and object tracking [14]. Such sensors realize the simultaneous need for wide field-of-view and good visual acuity. One popular class of space-variant sensors is formed by log-polar sensors which have a small area near the optical axis of greatly increased resolution (the fovea) coupled with a peripheral region that witnesses a gradual logarithmic falloff in resolution as one moves radially outward. These sensors are inspired by similar structures found in the primate retina where one finds both a peripheral region of gradually decreasing acuity and a circularly symmetric area centmlis characterized by a greater density of receptors and a disproportionate representation in the optic nerve [3]. The peripheral region, though of low visual acuity, is more sensitive to light intensity and movement. The existence of a region optimized for discrimination and recognition surrounded by a region geared towards detection thus allows the image of an object of interest detected in the outer region to be placed on the more analytic center for closer scrutiny. Such a strategy however necessitates the existence of (a) methods to determine which location in the periphery to foveate next, and (b) fast gaze-shifting mechanisms to achieve this 894 Rajesh P. N. Rao, Dana H. Ballard foveation. In the case of humans, the "where-to-Iook-next" issue is addressed by both bottom-up strategies such as motion or salience clues from the periphery as well as topdown strategies such as search for a particular form or color. Gaze-shifting is accomplished via very rapid eye movements called saccades. Due to their high velocities, guidance through visual feedback is not possible and hence, saccadic movement is preprogrammed or ballistic: a pattern of muscle activation is calculated in advance that will direct the fovea almost exactly to the desired position [3]. In this paper, we describe an iconic representation of scene points that facilitates topdown foveal targeting. The representation takes the form of a high-dimensional vector comprised of the responses of different order Gaussian derivative filters, which are known to form the principal components of natural images [5], at variety of orientations and scales. Such a representation has been recently shown to be useful for visual tasks ranging from texture segmentation [7] to object indexing using a sparse distributed memory [11]. We describe how this photometric representation of scene points can be used in locating previously foveated points when a log-polar sensor is being used. This property is then used in a simple learning strategy that makes use of multiple corrective saccades to adaptively form a retinotopic motor map similar in spirit to the one known to exist in the deep layers of the primate superior colliculus [13]. Our approach differs from previous strategies for learning motor maps (for instance, [12]) in that we use the visual modality to actively supply the necessary reinforcement signal required during the motor learning step (Section 3.2) . 2 The Multiscale Spatial Filter Representation In the active vision framework, vision is seen as subserving a larger context of the encompassing behaviors that the agent is engaged in. For these behaviors, it is often possible to use temporary, iconic descriptions of the scene which are only relatively insensitive to variations in the view. Iconic scene descriptions can be obtained, for instance, by employing a bank of linear spatial filters at a variety of orientations and scales. In our approach, we use derivative of Gaussian filters since these are known to form the dominant eigenvectors of natural images [5] and can thus be expected to yield reliable results when used as basis functions for indexingl . The exact number of Gaussian derivative basis functions used is motivated by the need to make the representations invariant to rotations in the image plane (see [11] for more details). This invariance can be achieved by exploiting the property of steerability [4] which allows filter responses at arbitrary orientations to be synthesized from a finite set of basis filters. In particular, our implementation uses a minimal basis set of two firstorder directional derivatives at 0? and 90?, three second-order derivatives at 0?, 60? and 120?, and four third-order derivatives oriented at 0?, 45?, 90?, and 135?. The response of an image patch J centered at (xo, Yo) to a particular basis filter G~j can be obtained by convolving the image patch with the filter : ri,j(XO, Yo) = (G/g . * I)(xo, Yo) = If 9 (XO - x, Yo - y)J(x, y)dx dy G/ (1) lIn addition, these filters are endorsed by recent physiological studies [15] which show that derivative-of-Gaussians provide the best fit to primate cortical receptive field profiles among the different functions suggested in the literature. Learning Saccadic Eye Movements Using Multiscale Spatial Filters 895 The iconic representation for the local image patch centered at (xo, Yo) is formed by combining into a single high-dimensional vector the responses from the nine basis filters, each (in our current implementation) at five different scales: r(xo, Yo) =(ri,j,s) , i = 1,2, 3;j = 1, . .. , i + 1; S = Smin , . .. , Smax (2) where i denotes the order of the filter, j denotes the number of filters per order, and denotes the number of different scales. S The use of multiple scales increases the perspicuity of the representation and allows interpolation strategies for scale invariance (see [9] for more details). The entire representation can be computed using only nine convolutions done at frame-rate within a pipeline image processor with nine constant size 8 x 8 kernels on a five-level octave-separated low-passfiltered pyramid of the input image. The 45-dimensional vector representation described above shares some of the favorable matching properties that accrue to high-dimensional vectors (d. [6]). In particular, the distribution of distances between points in the 45-dimensional space of these vectors approximates a normal distribution; most of the points in the space lie at approximately the mean distance and are thus relatively uncorrelated to a given point [11]. As a result, the multiscale filter bank tends to generate almost unique location-indexed signatures of image regions which can tolerate considerable noise before they are confused with other image regions. 2.1 Localization Denote the response vector from an image point as fi and that from a previously foveated model point as Tm. Then one metric for describing the similarity between the two points is simply the square of the Euclidean distance (or the sum-of-squared-differences) between 112. The algorithm for locating model points in a their response vectors dim = llfi new scene can then be described as follows : r.n 1. For the response vector representing a model point m, create a distance image I m defined by Im(x,y) = min [Imax - t3dim , 0] (3) where t3 is a suitably chosen constant (this makes the best match the brightest point in Im). 2. Find the best match point (Xb~, Yb~) in the image using the relation (4) Figure 1 shows the use of the localization algorithm for targeting the optical axis of a uniform-resolution sensor in an example scene. 2.2 Extension to Space-Variant Sensing The localization algorithm as presented above will obviously fail for sensors exhibiting nonuniform resolution characteristics. However, the multiscale structure of the response vectors can be effectively exploited to obtain a modified localization algorithm. Since decreasing radial resolution results in an effective reduction in scale (in addition to some 896 Rajesh P. N. Rao, Dana H. Ballard (a) (b) (c) (d) Figure 1: Using response vectors to saccade to previously foveated positions. (a) Initial gaze point. (b) New gaze point; (c) To get back to the original point, the "distance image" is computed: the brightest spot represents the point whose response vector is closest to that of the original gaze point; (d) Location of best match is marked and an oculomotor command at that location can be executed to foveate that point. other minor distortions) of previously foveated regions as they move towards the periphery, the filter responses previously occuring at larger scales now occur at smaller scales. Responses usually vary smoothly between scales; it is thus possible to establish a correspondence between the two response vectors of the same point on an object imaged at different scales by using a simple interpolate-and-eompare scale matching strategy. That is, in addition to comparing an image response vector and a model response vector directly as outlined in the previous section, scale interpolated versions of the image vector are also compared with the original model response vector. In the simplest case, interpolation amounts to shifting image response vectors by one scale and thus, responses from a new image are compared with original model responses at second, third, .. , scales, then with model responses at third, fourth, ... scales, and so on upto some threshold scale. This is illustrated in Figure 2 for two discrete movements of a simulated log-polar sensor. 3 The M ultisaccade Learning Strategy Since the high speed of saccades precludes visual guidance, advance knowledge of the precise motor command to be sent to the extraocular muscles for fixation of a desired retinal location is required. Results from neurophysiological and psychophysical studies suggest that in humans, this knowledge is acquired via learning: infants show a gradual increase in saccadic accuracy during their first year [1, 2] and adults can adapt to changes (caused for example by weakening of eye-muscles) in the interrelation between visual input and the saccades needed for centering. An adaptive mechanism for automatically learning the transfer function from retinal image space into motor space is also desirable in the context of active vision systems since an autonomous calibration of the saccadic system would (a) avoid the need for manual calibration, which can sometimes be complicated, and (b) provide resilience amidst changing circumstances caused by, for instance, changes in the camera lens mechanisms or degradation of the motor apparatus. 3.1 Motor Maps In primates, the superior eollieulus (SC), a multilayered neuron complex located in the upper regions of the brain stem, is known to playa crucial role in the saccade generation [13]. The upper layers of the SC contain a retinotopie sensory map with inputs from Learning Saccadic Eye Movements Using MuLtiscaLe SpatiaL Filters (a) Scale I (a) (b) Scale 2 III' I " 111'1"" Scille I (b) Scale 3 II" Scale 4 (c) Scale 5 III, ' II " ,li'I"" 111'1 Scale 2 s1I3 seNe 4 Scale 5 111 ' 1 "'1 1" 111 " 11,, 1 " 1' ,11 ,,1 ' 11 1,1 Scale I (c) 897 Scale 2 Scaje 3 Scale 4 Scale 5 11"1 11 , II ' ,11'''1, ,111,, 111 11 1 ,1'11'1 I (d) Figure 2: Using response vectors with a log-polar sensor, (a) through (c) represent a sequence of images (in Cartesian coordinates) obtained by movement of a simulated log-polar sensor from an original point (marked by '+') in the foveal region (indicated by a circle) towards the right. (d) depicts the process of interpolating (in this case, shifting) and matching response vectors of the same point as it moves towards the periphery of the sensor (Positive responses are represented by proportional upward bars and negative ones by proportional downward bars with the nine smallest scale responses at the beginning and the nine largest ones at the end). the retina while the deeper layers contain a motor map approximately aligned with the sensory map. The motor map can be visualized as a topologically-organized network of neurons which reacts to a local activation caused by an input signal with a vectorial output quantity that can be transcoded into a saccadic motor command. The alignment of the sensory and motor maps suggests the following convenient strategy for foveation: an excitation in the sensory layer (signaling a foveal target) is transferred to the underlying neurons in the motor layer which deliver the required saccade. In our framework, the excitation in the sensory layer before a goal-directed saccade corresponds to the brightest spot (most likely match) in the distance image (Figure 1 (c) for example), The formation of sensory map can be achieved using Kohonen's well-known stochastic learning algorithm by using a Gaussian input density function as described in [12]. Our primary interest lies not in the formation of the sensory map but in the development of a learning algorithm that assigns appropriate motor vectors to each location in the corresponding retinotopically-organized motor map. In particular, our algorithm employs a visual reinforcement signal obtained using iconic scene representations to determine the error vector during the learning step. 3.2 Learning the Motor Map Our multisaccade learning strategy is inspired by the following observations in [2] : During the first few weeks after birth, infants appear to fixate randomly. At about 3 months of age, infants are able to fixate stimuli albeit with a number of corrective saccades of relatively large dispersion. There is however a gradual decrease in both the dispersion 898 Rajesh P. N. Rao, Dana H. Ballard and the number of saccades required for foveation in subsequent months (Figure 3 (a) depicts a sample set of fixations). After the first year, saccades are generally accurate, requiring at most one corrective saccade2 ? The learning method begins by assigning random values to the motor vectors at each location. The response vector for the current fixation point is first stored and a random saccade is executed to a different point. The goal then is to refixate the original point with the help of the localization algorithm and a limited number of multiple corrective saccades. The algorithm keeps track of the motor vector with minimum error during each run and updates the motor vectors for the neighborhood around the original unit whenever an improvement is observed. The current run ends when either the original point was successfully foveated or the limit MAX for the maximum number of allowable corrective saccades was exceeded. A more detailed outline of the algorithm is as follows: 1. Initialize the motor map by assigning random values (within an appropriate range) to the saccadic motor vectors at each location. Align the optical axis of the sensor so that a suitable salient point falls on the fovea. Initialize the run number to t := O. 2. Store in memory the filter response vector of the point p currently in the center of the foveal region. Let t := t + 1. 3. Execute a random saccade to move the fovea to a different location in the scene. 4. Use the localization algorithm described in Section 2.2 and the stored response vector to find the location [ of the previously foveated point in the current retinal image. Execute a saccade using the motor vector St stored in this location in the motor map. 5. If the currently foveated region contains the original point p, return to 2 (SI is accurate); otherwise, s:= St. (a) Initialize the number of corrective saccades N := 0 and let (b) Determine the new location /' of p in the new image as in (4) and let emin be the error vector, i.e. the vector from the foveal center to /', computed from the output of the localization algorithm. (c) Execute a saccade using the motor vector Stl stored at [' and let ebe the error vector (computed from the output of the localization algorithm) from the foveal center to the new location [II of point p found as in 4. Let N := N + 1 and let SI' . (d) If lie'll < lliminll, then let emin := e and update the motor vectors for the units k given by the neighborhood function N(l, t) according to the wellknown Kohonen rule: (5) where 'Y(t) is an appropriate gain function (0 < 'Y(t) < 1). (e) If the currently foveated region contains the original point p, return to 2; otherwise, if N < MAX, then determine the new location [' of p in the new image as in (4) and go to 5(c) (i.e. execute the next saccade); otherwise, return to 2. s:= s+ 2Large saccades in adults are usually hypometric i.e. they undershoot, necessitating a slightly slower corrective saccade. There is currently no universally accepted explanation for the need for such a two-step strategy. 899 Learning Saccadic Eye Movements Using Multiscale Spatial Filters !" 1 I I .. ? ......, + IMX-IO f.! ' ..... 1 N!.IBfIll _ _ (a) (b) ",~;:;!l-;;:;;;--l-;;;; ll1I)"'---;;:ll1Ol;;;--;;"=-lI)---;",=-,C:;;,.,:---;;:; ...;;-;:;!,,", N_la rillelllltN (c) Figure 3: (a) Successive saccades executed by a 3-month old (left) and a 5-month old (right) infant when presented with a single illuminated stimulus (Adapted from [2]) . (b) Graph showing % of saccades that end directly in the fovea plotted against the number of iterations of the learning algorithm for different values of MAX. (c) An enlarged portion of the same graph showing points when convergence was achieved. The algorithm continues typically until convergence or the completion of a maximum number of runs. The gain term -y(t) and the neighborhood N(l, t) for any location l are gradually decreased with increasing number of iterations t. 4 Results and Discussion The simulation results for learning a motor map comprising of 961 units are shown in Figures 3 (b) and (c) which depict the variation in saccadic accuracy with the number of iterations of the algorithm for values of MAX (maximum number of corrective saccades) of 1, 5 and 10. From the graphs, it can be seen that starting with an initially random assignment of vectors, the algorithm eventually assigns accurate saccadic vectors to all units. Fewer iterations seem to be required if more corrective saccades are allowed but then each iteration itself takes more time. The localization algorithm described in Section 2.1 has been implemented on a Datacube MaxVideo 200 pipeline image processing system and takes 1-2 seconds for location of points. Current work includes the integration of the multisaccade learning algorithm described above with the Datacube implementation and further evaluation of the learning algorithm. One possible drawback of the proposed algorithm is that for large retinal spaces, learning saccadic motor vectors for every retinal location can be time-consuming and in some cases, even infeasible [1]. In order to address this problem, we have recently proposed a variation of the current learning algorithm which uses a sparse motor map in conjunction with distributed coding of the saccadic motor vectors. This organization bears some striking similarities to Kanerva's sparse distributed memory model [6] and is in concurrence with recent neurophysiological evidence [8] supporting a distributed population encoding of saccadic movements in the superior colliculus. We refer the interested reader to [10] for more details. 900 Rajesh P. N. Rao, Dana H. Ballard Acknowledgments We thank the NIPS*94 referees for their helpful comments. This work was supported by NSF research grant no. CDA-8822724, NIH/PHS research grant no. 1 R24 RR06853, and a grant from the Human Science Frontiers Program. References [1] Richard N. Aslin. Perception of visual direction in human infants. In C. Granlund, editor, Visual Perception and Cognition in Infancy, pages 91-118. Hillsdale, NJ: Lawrence Erlbaum Associates, 1993. [2] Gordon W. Bronson. The Scanning Patterns of Human Infants: Implications for Visual Learning. Norwood, NJ : Ablex, 1982. [3] Roger H.S. Carpenter. Movements of the Eyes. London: Pion, 1988. [4] William T . Freeman and Edward H. Adelson. The design and use of steerable filters. IEEE Transactions on Pattern Analysis and Machine Intelligence, 13(9):891-906, September 1991. [5] Peter J.B. Hancock, Roland J. Baddeley, and Leslie S. Smith. The principal components of natural images. Network, 3:61-70, 1992. [6] Pentti Kanerva. Sparse Distributed Memory. Bradford Books, Cambridge, MA, 1988. [7] Jitendra Malik and Pietro Perona. A computational model of texture segmentation. In IEEE Conference on Computer Vision and Pattern Recognition, pages 326-332, June 1989. [8] James T. McIlwain. Distributed spatial coding in the superior colliculus: A review. Visual Neuroscience, 6:3-13, 1991. [9] Rajesh P.N . Rao and Dana H. Ballard. An active vision architecture based on iconic representations. Technical Report 548, Department of Computer Scienc~, University of Rochester, 1995. [10] Rajesh P.N. Rao and Dana H. Ballard. A computational model for visual learning of saccadic eye movements. Technical Report 558, Department of Computer Science, University of Rochester, January 1995. [11] Rajesh P.N. Rao and Dana H. Ballard. Object indexing using an iconic sparse distributed memory. Technical Report 559, Department of Computer Science, University of Rochester, January 1995. [12] Helge Ritter, Thomas Martinetz, and Klaus Schulten. Neural Computation and SelfOrganizing Maps: An Introduction. Reading, MA: Addison-Wesley, 1992. [13] David L. Sparks and Rosi Hartwich-Young. The deep layers of the superior collicuIus. In R.H. Wurtz and M.E. Goldberg, editors, The Neurobiology of Saccadic Eye Movements, pages 213-255. Amsterdam: Elsevier, 1989. [14] Massimo Tistarelli and Giulio Sandini. Dynamic aspects in active vision. Computer Vision, Graphics, and Image Processing: Image Understanding, 56(1):108-129, 1992. [15] R.A. Young. The Gaussian derivative theory of spatial vision: Analysis of cortical cell receptive field line-weighting profiles. General Motors Research Publication GMR4920, 1985.
923 |@word version:1 suitably:1 simulation:1 gradual:3 reduction:1 initial:1 foveal:6 contains:2 current:6 comparing:1 activation:2 assigning:2 dx:1 si:2 realize:1 subsequent:1 analytic:1 motor:30 update:2 depict:1 discrimination:1 infant:6 intelligence:1 fewer:1 plane:1 scienc:1 beginning:1 smith:1 location:17 successive:1 five:2 direct:1 supply:1 fixation:3 acquired:1 expected:1 rapid:1 behavior:2 brain:1 inspired:2 freeman:1 decreasing:2 automatically:1 increasing:1 confused:1 retinotopic:1 underlying:1 begin:1 nj:2 every:1 firstorder:1 amidst:1 exactly:1 giulio:1 unit:4 grant:3 scrutiny:1 appear:1 before:2 positive:1 local:2 resilience:1 tends:1 apparatus:1 io:1 limit:1 receptor:1 interrelation:1 encoding:1 interpolation:2 approximately:2 suggests:1 limited:1 range:1 directed:1 unique:1 camera:1 acknowledgment:1 differs:1 signaling:1 spot:2 steerable:1 area:2 matching:3 convenient:1 radial:1 suggest:1 get:1 targeting:2 context:2 map:18 center:4 go:1 starting:1 resolution:5 spark:1 assigns:2 rule:1 imax:1 population:1 variation:3 autonomous:1 coordinate:1 target:2 exact:1 us:2 goldberg:1 associate:1 velocity:1 referee:1 recognition:2 located:1 continues:1 bottom:1 role:1 observed:1 region:14 extraocular:1 movement:14 decrease:1 dynamic:1 preprogrammed:1 signature:1 ablex:1 deliver:1 localization:9 basis:6 necessitates:1 represented:1 corrective:9 separated:1 fast:1 describe:3 effective:1 london:1 hancock:1 detected:1 sc:2 klaus:1 formation:2 neighborhood:3 birth:1 whose:1 larger:2 distortion:1 otherwise:3 precludes:1 itself:1 obviously:1 sequence:1 kohonen:2 aligned:1 combining:1 achieve:1 description:2 exploiting:1 convergence:2 smax:1 object:5 help:1 derive:1 completion:1 minor:1 edward:1 implemented:1 c:1 disproportionate:1 exhibiting:1 direction:1 drawback:1 filter:21 subsequently:1 stochastic:1 centered:2 human:5 hillsdale:1 im:2 extension:1 frontier:1 around:1 normal:1 brightest:3 lawrence:1 cognition:1 week:1 vary:1 smallest:1 polar:5 favorable:1 ballistic:1 currently:4 sensitive:1 largest:1 create:1 successfully:1 sensor:13 gaussian:5 modified:1 avoid:1 command:3 publication:1 conjunction:1 acuity:3 yo:6 june:1 iconic:7 improvement:1 greatly:1 dim:1 helpful:1 elsevier:1 entire:1 weakening:1 typically:1 initially:1 perona:1 relation:1 comprising:1 interested:1 upward:1 issue:1 among:1 orientation:4 development:1 spatial:9 integration:1 initialize:3 field:3 represents:1 adelson:1 photometric:2 report:3 stimulus:2 aslin:1 richard:1 employ:1 retina:2 few:1 oriented:1 randomly:1 gordon:1 interpolate:1 william:1 detection:1 organization:1 interest:3 evaluation:1 alignment:1 light:1 xb:1 implication:1 accurate:4 rajesh:8 closer:1 necessary:1 indexed:1 euclidean:1 old:2 desired:2 circle:1 accrue:1 guidance:2 plotted:1 cda:1 minimal:1 increased:1 instance:3 rao:9 assignment:1 leslie:1 perspicuity:1 uniform:1 comprised:2 erlbaum:1 graphic:1 stored:4 scanning:1 adaptively:1 st:2 density:2 ritter:1 gaze:5 squared:1 book:1 convolving:1 derivative:8 return:3 actively:1 li:2 retinal:5 coding:2 includes:1 jitendra:1 caused:3 view:2 imx:1 portion:1 mcilwain:1 complicated:1 rochester:6 formed:2 square:1 accuracy:2 characteristic:1 yield:1 t3:1 directional:1 processor:1 simultaneous:1 falloff:1 manual:1 whenever:1 centering:1 against:1 james:1 fixate:2 gain:2 radially:1 popular:1 color:1 knowledge:2 segmentation:2 organized:2 subserving:1 back:1 nerve:1 exceeded:1 wesley:1 tolerate:1 response:29 emin:2 yb:1 done:1 though:1 execute:4 roger:1 until:1 multiscale:7 indicated:1 contain:2 requiring:1 undershoot:1 hence:1 symmetric:1 imaged:1 illustrated:1 ll:1 during:5 excitation:2 octave:1 allowable:1 outline:1 occuring:1 demonstrate:1 necessitating:1 motion:1 image:29 ranging:1 recently:2 fi:1 nih:1 superior:5 rotation:1 retinotopically:1 insensitive:1 approximates:1 synthesized:1 refer:1 cambridge:1 outlined:1 calibration:2 geared:1 similarity:2 align:1 dominant:1 playa:1 closest:1 recent:3 wellknown:1 periphery:4 store:1 accomplished:1 exploited:1 muscle:3 seen:2 minimum:1 greater:1 determine:4 signal:3 ii:4 multiple:3 desirable:1 stem:1 technical:3 match:4 characterized:1 adapt:1 ebe:1 lin:1 roland:1 variant:3 vision:9 metric:1 circumstance:1 wurtz:1 iteration:5 kernel:1 sometimes:1 represent:1 pyramid:1 achieved:3 cell:1 addition:3 addressed:1 decreased:1 modality:1 crucial:1 martinetz:1 comment:1 facilitates:1 sent:1 spirit:1 seem:1 near:1 iii:2 reacts:1 variety:2 fit:1 architecture:1 tm:1 motivated:1 locating:3 peter:1 nine:5 deep:2 useful:1 generally:1 delivering:1 eigenvectors:1 endorsed:1 detailed:1 selforganizing:1 outward:1 amount:1 ph:1 visualized:1 simplest:1 generate:1 datacube:2 exist:1 nsf:1 neuroscience:1 per:1 track:1 discrete:1 smin:1 four:1 salient:1 threshold:1 changing:1 graph:3 pietro:1 sum:1 colliculus:3 year:2 run:4 fourth:1 striking:1 topologically:1 almost:2 reader:1 patch:3 dy:1 illuminated:1 layer:7 correspondence:1 r24:1 adapted:1 occur:1 optic:1 vectorial:1 scene:10 ri:2 interpolated:1 aspect:1 speed:1 min:1 optical:3 relatively:3 transferred:1 department:4 according:1 peripheral:3 smaller:1 slightly:1 helge:1 primate:4 gradually:2 indexing:2 invariant:1 xo:6 pipeline:2 kanerva:2 previously:7 describing:1 eventually:1 mechanism:3 fail:1 needed:1 addison:1 end:3 gaussians:1 upto:1 appropriate:3 slower:1 existence:2 original:10 thomas:1 denotes:3 establish:1 psychophysical:1 move:4 malik:1 quantity:1 strategy:12 saccadic:17 receptive:2 primary:1 september:1 fovea:5 distance:6 thank:1 simulated:2 outer:1 executed:3 negative:1 implementation:3 design:1 upper:2 convolution:1 neuron:3 observation:1 dispersion:2 finite:1 supporting:1 january:2 witness:1 neurobiology:1 precise:1 frame:1 nonuniform:1 arbitrary:1 intensity:1 david:1 required:5 optimized:1 temporary:1 nip:1 adult:2 able:1 suggested:1 topdown:2 usually:2 pattern:4 bar:2 address:1 perception:2 reading:1 oculomotor:1 program:1 reliable:1 memory:5 max:4 explanation:1 shifting:4 suitable:1 natural:4 representing:1 eye:10 axis:3 coupled:1 review:1 literature:1 understanding:1 encompassing:1 bear:1 generation:1 proportional:2 dana:9 age:1 norwood:1 agent:1 sandini:1 editor:2 bank:2 uncorrelated:1 surrounded:1 share:1 placed:1 supported:1 infeasible:1 salience:1 deeper:1 concurrence:1 wide:1 fall:1 steerability:1 sparse:5 distributed:7 feedback:1 calculated:1 cortical:2 sensory:7 adaptive:2 clue:1 reinforcement:2 universally:1 employing:1 transaction:1 keep:1 active:5 consuming:1 search:2 ballard:8 transfer:1 complex:1 interpolating:1 multilayered:1 noise:1 profile:2 allowed:1 carpenter:1 enlarged:1 depicts:2 ny:1 position:2 schulten:1 lie:3 infancy:1 third:3 weighting:1 young:2 showing:2 sensing:1 physiological:1 stl:1 evidence:1 circularly:1 albeit:1 effectively:1 texture:2 foveated:9 downward:1 cartesian:1 smoothly:1 logarithmic:1 simply:1 likely:1 neurophysiological:2 visual:14 amsterdam:1 tracking:1 saccade:26 corresponds:1 foveate:2 ma:2 marked:2 goal:2 month:4 towards:4 massimo:1 considerable:1 change:2 foveation:3 sene:1 principal:2 degradation:1 called:1 lens:1 pentti:1 engaged:1 invariance:2 accepted:1 bradford:1 baddeley:1
7,161
924
Extracting Rules from Artificial Neural Networks with Distributed Representations Sebastian Thrun University of Bonn Department of Computer Science III Romerstr. 164, D-53117 Bonn, Germany E-mail: thrun@carbon.informatik.uni-bonn.de Abstract Although artificial neural networks have been applied in a variety of real-world scenarios with remarkable success, they have often been criticized for exhibiting a low degree of human comprehensibility. Techniques that compile compact sets of symbolic rules out of artificial neural networks offer a promising perspective to overcome this obvious deficiency of neural network representations. This paper presents an approach to the extraction of if-then rules from artificial neural networks. Its key mechanism is validity interval analysis, which is a generic tool for extracting symbolic knowledge by propagating rule-like knowledge through Backpropagation-style neural networks. Empirical studies in a robot arm domain illustrate the appropriateness of the proposed method for extracting rules from networks with real-valued and distributed representations. 1 Introduction In the last few years artificial neural networks have been applied successfully to a variety of real-world problems. For example, neural networks have been successfully applied in the area of speech generation [12] and recognition [18], vision and robotics [8], handwritten character recognition [5], medical diagnostics [11], and game playing [13]. While in these and other approaches neural networks have frequently found to outperform more traditional approaches, one of their major shortcomings is their low degree of human comprehensibility. In recent years, a variety of approaches for compiling rules out of networks have been proposed. Most approaches [1, 3,4,6, 7, 16, 17] compile networks into sets of rules with equivalent structure: Each processing unit is mapped into a separate rule-or a smal1 set of rules-, and the ingoing weights are interpreted as preconditions to this rule. Sparse connectivity facilitates this type rule extraction, and so do binary activation values. In order to enforce such properties, which is a necessary prerequisite for these techniques to work effectively, some approaches rely on specialized training procedures, network initializations 506 Sebastian Thrun and/or architectures. While such a methodology is intriguing, as it draws a clear one-to-one correspondence between neural inference and rule-based inference, it is not universally applicable to arbitrary Backpropagation-style neural networks. This is because artificial neural networks might not meet the strong representational and structural requirements necessary for these techniques to work successfully. When the internal representation of the network is distributed in nature, individual hidden units typically do not represent clear, logical entities. One might argue that networks, if one is interested in extracting rules, should be constructed appropriately. But this would outrule most existing network implementation~, as such considerations have barely played a role. In addition, such an argument would suppress the development of distributed, non-discrete internal representations, which have often be attributed for the generalization properties of neural networks. It is this more general class of networks that is at stake in this paper. This paper presents a rule extraction method which finds rules by analyzing networks as a whole. The rules are of the type "if X then y," where both x and y are described by a linear set of constraints. The engine for proving the correspondence of rule and network classification is VI-Analysis. Rules extracted by VI-Analysis can be proven to exactly describe the network. 2 Validity-Interval Analysis Validity Interval Analysis (in short: VI-Analysis) is a generic tool for analyzing the inputoutput behavior of Backpropagation-style neural networks. In short, they key idea of VIAnalysis is to attach intervals to the activation range of each unit (or a subset of all units, like input and output units only), such that the network's activations must lie within these intervals. These intervals are called validity intervals. VI-Analysis checks whether such a set of intervals is consistent, i.e., whether there exists a set of network activations inside the validity intervals. It does this by iteratively refining the validity intervals, excluding activations that are provably inconsistent with other intervals. In what follows we will present the general VI-Analysis algorithm, which can be found in more detail elsewhere [14], Let n denote the total number of units in the network, and let Xi denote the (output) activation of unit i (i 1, ... , n). If unit i is an input unit, its activation value will simply be the external input value. If not, i.e., if i refers to a hidden or an output unit, let P( i) denote the set of units that are connected to unit i through a link. The activation Xi is computed in two steps: with WikXk + Oi = L kEP(i) The auxiliary variable neti is the net-input of unit i, and Wik and Oi are the weights and biases, respectively. O'j denotes the transfer function (squashing function), which usually is given by 1 + e- net , Validity intervals for activation values Xi are denoted by [ai, bi ]. If necessary, validity intervals are projected into the net-input space of unit i, where they will be denoted by [a~, b~]. Let T be a set of validity intervals for (a subset of) all units. An activation vector (XI, .. " xn) is said to be admissible with respect to T, if all activations lie in T. A set of intervals T is consistent, if there exists an admissible activation vector. Otherwise T is inconsistent. Assume an initial set of intervals, denoted by T, is given (in the next section we will present a procedure for generating initial intervals). VI-Analysis refines T iteratively using linear Extracting Rules from Artificial Neural Networks with Distributed Representations 507 non-linear .<;quashing functioll' CJ linear equations Figure 1: VI-Analysis in a single weight layer. Units in layer P are connected to the units in layer S. A validity interval [aj, bj ] is assigned to each unit j E PuS. By projecting the validity intervals for all i E S, intervals [a~, b~] for the net-inputs netj are created. These, plus the validity intervals for all units k E P, form a set of linear constraints on the activations x k in layer P. Linear programming is now employed to refine all interval bounds one-by-one. programming [9], so that those activation values which are inconsistent with other intervals are excluded. In order to simplify the presentation, let us assume without loss of ?enerality (a) that the network is layered and fully connected between two adjacent layers, and (b) that there is an interval [aj, bj ] ~ [0,1] in I for every unit in P and S.2 Consider a single weight layer, connecting a layer of preceding units, denoted by p, to a layer of succeeding units, denoted by S (cf Fig. 1). In order to make linear programming techniques applicable, the non-linearity of the transfer function must be eliminated. This is achieved by projecting [ai, bi ] back to the corresponding net-input intervals 3 [ai, biJ = {T-I([ai' biD E ~2 for all i E S. The resulting validity intervals in P and S form the foIIowing set of linear constraints on the activation values in P: Vk E P: Xk > ak and Xk < bk Vi E S: WikXk + ()d WjkXk + ()j > a~, [by substituting neti = (1) kEP kEP WikXk + ()i] WikXk + ()j < b~, [by substituting netj = L L L L kEP kEP Notice that all these constraints are linear in the activation values Xk (k E P). Linear programming allows to maximize or minimize arbitrary linear combinations of the variables x j while not violating a set of linear constraints [9]. Hence, linear programming can be applied to refine lower and upper bounds for validity intervals one-by-one. In VI-Analysis, constraints are propagated in two phases: 1. Forward phase. To refine the bounds aj and bj for units i E S, new bounds iii and hi are 'This assumption simplifies the description of VI-Analysis, although VI-Analysis can also be applied to arbitrary non-layered, partially connected network architectures, as well as recurrent networks not examined here. 2The canonical interval [0, I] corresponds to the state of maximum ignorance about the activation of a unit, and hence is the default interval if no more specific interval is known. 3Here ~ denotes the set of real numbers extended by ?oo. Notice that this projection assumes that the transfer function is monotonic. 508 derived: Sebastian Thrun with aA'z? min neti L: max L: min WikXk + Oi WikXk + OJ kE1' with = max neti kE1' If o'i > ai, a tighter lower bound is found and ai is updated by o'i . Likewise, b i is set to hi if hi < bi . Notice that the minimax operator is computed within the bounds imposed by Eq. I, using the Simplex algorithm (linear programming) [9]. 2. Backward phase. In the backward phase the bounds ak and b k of all units k E Pare refined. li k minxk and hk = max Xk As in the forward phase, ak is updated by o'k if li k > ak, and h is updated by hk if hk < bk. If the network has multiple weight layers, this process is applied to all weight layers one-byone. Repetitive refinement results in the propagation of interval constraints through multiple layers in both directions. The convergence of VI-Analysis follows from the fact that the update rule that intervals are changed monotonically, since they can only shrink or stay the same. Recall that the "input" of VI-Analysis is a set of intervals I ~ [0, l]n that constrain the activations of the network. VI-Analysis generates a refined set of intervals, I' ~ I, so that all admissible activation values in the original intervals I are also in the refined intervals I'. In other words, the difference between the original set of intervals and the refined set of intervals I - I' is inconsistent. In summary, VI-Analysis analyzes intervals I in order to detect inconsistencies. If I is found to be inconsistent, there is provably no admissible activation vector in I . Detecting inconsistencies is the driving mechanism for the verification and extraction of rules presented in turn. 3 Rule Extraction The rules considered in this paper are propositional if-then rules. Although VI-Analysis is able to prove rules expressed by arbitrary linear constraints [14], for the sake of simplicity we will consider only rules where the precondition is given by a set of intervals for the individual input values, and the output is a single target category. Rules of this type can be written as: !linput E some hypercube I then class is C (or short: I - - C) for some target class C. The compliance of a rule with the network can be verified through VI-Analysis. Assume, without loss of generality, the network has a single output unit, and input patterns are classified as members of class C if and only if the output activation, Xout, is larger than a threshold e (see [14] for networks with multiple output units). A rule conjecture I - - C is then verified by showing that there is no input vector i E I that falls into the opposite class, ,C. This is done by including the (negated) condition Xout E [0, e] into the set of intervals: Ineg = 1+ {xout E [0, e]}. If the rule is correct, Xout will never be in [0, e]. Hence, if VI-Analysis finds an inconsistency in Ineg, the rule I - - ,C is proven to be incorrect, and thus the original rule I - - C holds true for the network at hand. This illustrates how rules are verified using VI-Analysis. It remains to be shown how such conjectures can be generated in a systematic way. Two major classes of approaches can be distinguished, specific-to-general and general-to-specific. Extracting Rules from Artificial Neural Networks with Distributed Representations 509 Figure 2: Robot Ann. (a) Front view of two arm configurations. (b) Two-dimensional side view. The grey area indicates the workspace, which partially intersects with the table. 1. Specific-to-general. A generic way to generate rules, which forms the basis for the experimental results reported in the next section, is to start with rather specific rules which are easy to verify, and gradually generalize those rules by enlarging the corresponding validity intervals. Imagine one has a training instance that, without loss of generality, falls into a class C. The input vector of the training instance already forms a (degenerate) set of validity intervals I. VI-Analysis will, applied to I, trivially confirm the membership in C, and hence the single-point rule I ~ C. Starting with I, a sequence of more general rule preconditions I C II C I2 C ... can be obtained by enlarging the precondition of the rule (i.e., the input intervals I) by small amounts, and using VI-Analysis to verify if the new rule is still a member of its class. In this way randomly generated instances can be used as "seeds" for rules, which are then generalized via VI-Analysis. 2. General-to-specific. An alternative way to extract rules, which has been studied in more detail elsewhere [14], works from general to specific. General-to-specific rule search maintains a list of non-proven conjectures, R. R is initialized with the most general rules (like "everything is in C" and "nothing is in C"). VI-Analysis is then applied to prove rules in R. If it successfully confirms a rule, the rule and its complement is removed from R. If not, the rule is removed, too, but instead new rules are added to R. These new rules form a specialized version of the old rule, so that their disjunct is exactly the old rule. For example, new rules can be generated by splitting the hypercube spanned by the old rule into disjoint regions, one for each new rule. Then, the new set R is checked with VI-Analysis. The whole procedure continues till R is empty and the whole input domain is described by rules. In discrete domains, such a strategy amounts to searching directed acyclic graphs in breadth-first manner. Obviously, there is a variety of alternative techniques to generate meaningful rule hypotheses. For example, one might employ a symbolic learning technique such as decision tree learning [10] to the same training data that was used for training the network. The rules, which are a result of the symbolic approach, constitute hypotheses that can be checked using VI-Analysis. 4 Empirical Results In this section we will be interested in extracting rules in a real-valued robot arm domain. We trained a neural network to model the forward kinematics function of a 5 degree-of-freedom robot arm. The arm, a Mitsubishi RV-Ml, is depicted in Fig. 2. Its kinematic function determines the position of the tip of the manipulator in (x, y, z) workspace coordinates and 510 Sebastian Thrun coverage first 10 rules first 100 rules first 1 000 rules first 10000 rules average (per rule) 9.79% 2.59% 1.20% 0.335% cumulative 30.2% 47.8% 61.6% 84.4% Table 1: Rule coverage in the robot arm domain. These numbers inc1ude rules for both concepts, SAFE and UNSAFE. the angle of the manipulator h to the table based on the angles of the five joints. As can be seen in Fig. 2, the workspace intersects with the table on which the arm is mounted. Hence, some configurations of the joints are safe, namely those for which z ~ 0, whiJe others can physically not be reached without a col1ision that would damage the robot (unsafe). When operating the robot arm one has to be able to tell safe from unsafe. Henceforth, we are interested in a set of rules that describes the subspace of safe and unsafe joint configurations. A total of 8192 training examples was used for training the network (four input, five hidden and four output units), resulting in a considerably accurate model of the kinematics of the robot arm. Notice that the network operates in a continuous space. Obviously, compiling the network into logical rules node-by-node, as frequently done in other approaches to rule extraction, is difficult due to the real-valued and distributed nature of the internal representation. Instead, we applied VI-Analysis using a specific-to-general mechanism as described above. More specifically, we incrementally constructed a collection of rules that gradually covered the workspace of the robot arm. Rules were generated whenever a (random) joint configuration was not covered by a previously generated rule. Table 1 shows average results that characterize the extraction of rules. Initially, each rule covers a rather large fraction of the 5-dimensional joint configuration space. As few as 11 rules, on average, suffice to cover more than 50% (by volume) of the whole input space. However, these 50% are the easy half. As the domain gets increasingly covered by rules, gradually more specific rules are generated in regions closer to the c1ass boundary. After extracting 10,000 rules, only 84.4% of the input space is covered. Since the decision boundary between the two c1asses is highly non-linear, finitely many rules will never cover the input space completely. How general are the rules extracted by VI-Analysis? Genera])y speaking, for joint configurations c10se to the c1ass boundary, i.e., where the tip of the manipulator is close to the table, we observed that the extracted rules were rather specific. If instead the initial configuration was closer to the center of a class, VI-Analysis was observed to produce more general rules that had a larger coverage in the workspace. Here VI-Analysis managed to extract surprisingly general rules. For example, the configuration a = (30 0 ,80 0 ,200 ,600 , -200 ), which is depicted in Fig. 3, yields the rule !!.a2 ~ 90.5 0 and a3 ~ 27.3 0 then SAFE. Notice that out of 10 initial constraints, 8 were successfully removed by VI-Analysis. The rule lacks both bounds on a), a4 and as and the lower bounds on a2 and a3. Fig. 3a shows the front view of the initial arm configuration and the generalized rule (grey area). Fig. 3b shows a side view of the arm, along with a slice of the rule (the base joint a) is kept fixed). Notice that this very rule covers 17.1 % of the configuration space (by volume). Such general rules were frequently found in the robot arm domain. This conc1 udes the brief description of the experimental results. Not mentioned here are results with different size networks, and results obtained for the MONK's benchmark problems. For example, in the MONK's problems [15], VI-Analysis successfully extracted compact target Extracting Rules from Artificial Neural Networks with Distributed Representations 511 Figure 3: A single rule, extracted from the network. (a) Front view. (b) Two-dimensional side view. The grey area indicates safe positions for the tip of the manipulator. concepts using the originally published weight sets. These results can be found in [14]. 5 Discussion In this paper we have presented a mechanism for the extraction of rules from Backpropagationstyle neural networks. There are several limitations of the current approach that warrant future research. (a) Speed. While the one-to-one compilation of networks into rules is fast, rule extraction via VI-Analysis requires mUltiple runs of linear programming, each of which can be computationally expensive [9]. Searching the rule space without domain-specific search heuristics can thus be a most time-consuming undertaking. In all our experiments, however, we observed reasonably fast convergence of the VI-Algorithm, and we successfully managed to extract rules from larger networks in reasonable amounts of time. Recently, Craven and Shavlik proposed a more efficient search method which can be applied in conjunction with VI-Analysis [2]. (b) Language. Currently VI-Analysis is limited to the extraction of if-then rules with linear preconditions. While in [14] it has been shown how to generalize VI-Analysis to rules expressed by arbitrary linear constraints, a more powerful rule language is clearly desirable. (c) Linear optimization. Linear programming analyzes multiple weight layers independently, resulting in an overly careful refinement of intervals. This effect can prevent from detecting correct rules. If linear programming is replaced by a non-linear optimization method that considers multiple weight layers simultaneously, more powerful rules can be generated. On the other hand, efficient non-linear optimization techniques might find rules which do not describe the network accurately. Moreover, it is generally questionable whether there will ever exist techniques for mapping arbitrary networks accurately into compact rule sets. Neural networks are their own best description, and symbolic rules might not be appropriate for describing the input-output behavior of a complex neural network. A key feature of of the approach presented in this paper is the particular way rules are extracted. Unlike other approaches to the extraction of rules, this mechanism does not compile networks into structurally equivalent set of rules. Instead it analyzes the input output relation of networks as a whole. As a consequence, rules can be extracted from unstructured networks with distributed and real-valued internal representations. In addition, the extracted rules describe the neural network accurately, regardless of the size of the network. This makes VI-Analysis a promising candidate for scaling rule extraction techniques to deep networks, in which approximate rule extraction methods can suffer from cumulative errors. We conjecture that such properties are important if meaningful rules are to be extracted in today's and tomorrow's successful Backpropagation applications. 512 Sebastian Thrun Acknowledgment The author wishes to express his gratitude to Marc Craven, Tom Dietterich, Clayton McMillan. Tom Mitchell and Jude Shavlik for their invaluable feedback that has influenced this research . References [I] M. W. Craven and J. W. Shavlik. Learning symbolic rules using artificial neural networks. In Paul E. Utgoff, editor, Proceedings of the Tenth International Conference on Machine Learning, 1993. Morgan Kaufmann. [2] M. W. Craven and J. W. Shavlik. Using sampling and queries to extmct rules from tmined neural networks. In Proceedings of the Eleventh International Conference on Machine Learning, 1994. Morgan Kaufmann. [3] L.-M. Fu. Integration of neural heuristics into knowledge-based inference. Connection Science, 1(3):325-339,1989. [4] C. L. Giles and C. W. Omlin. Rule refinement with recurrent neural networks. In Proceedings of the IEEE International Conference on Neural Network, 1993. IEEE Neuml Network Council. [5] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard. W. Hubbard, and L. D. Jackel. Backpropagation applied to handwritten zip code recognition. Neural Computation, 1:541-551. 1990. [6] J. J. Mahoney and R. J. Mooney. Combining neural and symbolic learning to revise probabilistic rule bases. In J. E. Moody, S. J. Hanson, and R. P. Lippmann, editors, Advances in Neural Information Processing Systems 5, 1993. Morgan Kaufmann. [7] C. McMillan, M. C. Mozer, and P. Smolensky. Rule induction through integrated symbolic and subsymbolic processing. In J. E. Moody. S. J. Hanson. and R. P. Lippmann, editors, Advances in Neural Information Processing Systems 4. 1992. Morgan Kaufmann. [8] D. A. Pomerleau. ALVINN: an autonomous land vehicle in a neural network. Technical Report CMU-CS-89-1 07. Computer Science Dept. Carnegie Mellon University, Pittsburgh PA, 1989. [9] W. H. Press. Numerical recipes in C " the art of scientific computing. Cambridge University Press, Cambridge [Cambridgeshire], New York, 1988. [10] J. R. Quinlan. Induction of decision trees. Machine Learning, 1:81-106,1986. [II] J. Rennie. Cancer catcher: Neural net catches errors that slip through pap tests. Scientific American, 262, May 1990. [12] T. J. Sejnowski and C. R. Rosenberg. Nettalk: A parallel network that learns to read aloud. Technical Report JHUIEECS-86/01, Johns Hopkins University, 1986. [13] G. J. Tesauro. Practical issues in tempoml difference learning. Machine Learning. 8, 1992. [14] S. Thrun. Extracting provably correct rules from artificial neuml networks. Technical Report IAI-TR-93-5, University of Bonn. Institut flir Informatik III, D-53117 Bonn, May 1993. [15] S. Thrun, J. Bala,E. Bloedorn, I. Bmtko, B. Cestnik, J. Cheng, K. Dejong, S. Dzeroski, D. Fisher, S. E. Fahlman, R. Hamann, K. Kaufman, S. Keller. I. Kononenko, J. Kreuziger, R. S. Michalski, T.M. Mitchell, P. Pachowicz, Y. Reich, H. Vafaie, W. Van de WeIde, W. Wenzel, J. Wnek, and J. Zhang. The MONK's problems - a performance comparison of different learning algorithms. Technical Report CMU-CS-91-197, Carnegie Mellon University. Pittsburgh, PA, December 1991. [16] G. Towell and J. W. Shavlik. Interpretation of artificial neural networks: Mapping knowledgebased neural networks into rules. In J. E. Moody. S. J. Hanson, and R. P. Lippmann, editors, Advances in Neural Information Processing Systems 4. 1992. Morgan Kaufmann. [17] V. Tresp and J. Hollatz. Network structuring and training using rule-based knowledge. In J. E. Moody, S. J. Hanson. and R. P. Lippmann, editors, Advances in Neural Information Processing Systems 5,1993. Morgan Kaufmann. [18] A. H. Waibel. Modular construction of time-delay neural networks for speech recognition. Neural Computation, 1:39-46, 1989.
924 |@word version:1 grey:3 confirms:1 mitsubishi:1 xout:4 tr:1 c1ass:2 configuration:10 initial:5 existing:1 current:1 activation:21 intriguing:1 must:2 written:1 john:1 refines:1 numerical:1 succeeding:1 update:1 half:1 monk:3 xk:4 short:3 detecting:2 node:2 zhang:1 five:2 along:1 constructed:2 tomorrow:1 incorrect:1 prove:2 eleventh:1 inside:1 manner:1 behavior:2 frequently:3 udes:1 linearity:1 suffice:1 moreover:1 what:1 kaufman:1 interpreted:1 dejong:1 every:1 questionable:1 exactly:2 unit:28 medical:1 consequence:1 ak:4 analyzing:2 meet:1 might:5 plus:1 initialization:1 studied:1 examined:1 compile:3 genus:1 limited:1 range:1 bi:3 directed:1 acknowledgment:1 lecun:1 practical:1 backpropagation:5 procedure:3 area:4 empirical:2 projection:1 word:1 refers:1 symbolic:8 get:1 close:1 layered:2 operator:1 linput:1 equivalent:2 imposed:1 center:1 regardless:1 starting:1 independently:1 keller:1 simplicity:1 splitting:1 unstructured:1 rule:121 spanned:1 his:1 proving:1 searching:2 coordinate:1 autonomous:1 updated:3 target:3 imagine:1 today:1 construction:1 programming:9 slip:1 hypothesis:2 pa:2 recognition:4 expensive:1 continues:1 observed:3 role:1 precondition:5 region:2 connected:4 removed:3 mentioned:1 mozer:1 utgoff:1 trained:1 basis:1 completely:1 joint:7 intersects:2 fast:2 shortcoming:1 describe:3 sejnowski:1 artificial:12 query:1 tell:1 refined:4 aloud:1 heuristic:2 larger:3 valued:4 modular:1 rennie:1 otherwise:1 obviously:2 sequence:1 net:6 michalski:1 combining:1 till:1 degenerate:1 representational:1 description:3 inputoutput:1 recipe:1 convergence:2 empty:1 requirement:1 knowledgebased:1 produce:1 generating:1 illustrate:1 recurrent:2 oo:1 propagating:1 dzeroski:1 finitely:1 eq:1 strong:1 auxiliary:1 coverage:3 c:2 exhibiting:1 appropriateness:1 direction:1 safe:6 undertaking:1 correct:3 human:2 everything:1 generalization:1 tighter:1 hold:1 considered:1 seed:1 mapping:2 bj:3 substituting:2 driving:1 major:2 a2:2 applicable:2 currently:1 jackel:1 council:1 hubbard:1 successfully:7 tool:2 clearly:1 ke1:2 rather:3 rosenberg:1 conjunction:1 structuring:1 derived:1 refining:1 vk:1 check:1 indicates:2 hk:3 detect:1 inference:3 membership:1 typically:1 integrated:1 initially:1 hidden:3 relation:1 kep:5 interested:3 germany:1 provably:3 issue:1 classification:1 denoted:5 development:1 art:1 integration:1 never:2 extraction:13 eliminated:1 sampling:1 warrant:1 future:1 simplex:1 others:1 report:4 simplify:1 few:2 employ:1 randomly:1 simultaneously:1 individual:2 replaced:1 phase:5 freedom:1 highly:1 kinematic:1 henderson:1 mahoney:1 diagnostics:1 compilation:1 netj:2 accurate:1 fu:1 closer:2 necessary:3 institut:1 tree:2 old:3 initialized:1 criticized:1 instance:3 disjunct:1 giles:1 cover:4 subset:2 delay:1 successful:1 front:3 too:1 characterize:1 reported:1 considerably:1 international:3 stay:1 workspace:5 systematic:1 probabilistic:1 tip:3 connecting:1 hopkins:1 moody:4 connectivity:1 henceforth:1 wenzel:1 external:1 american:1 style:3 li:2 de:2 vi:37 vehicle:1 view:6 hollatz:1 reached:1 start:1 maintains:1 parallel:1 minimize:1 oi:3 kaufmann:6 likewise:1 yield:1 generalize:2 flir:1 handwritten:2 accurately:3 informatik:2 published:1 mooney:1 classified:1 influenced:1 sebastian:5 checked:2 whenever:1 obvious:1 attributed:1 propagated:1 revise:1 logical:2 recall:1 knowledge:4 mitchell:2 cj:1 back:1 ingoing:1 originally:1 violating:1 methodology:1 tom:2 iai:1 done:2 shrink:1 generality:2 hand:2 propagation:1 incrementally:1 lack:1 aj:3 scientific:2 manipulator:4 dietterich:1 validity:16 verify:2 true:1 concept:2 managed:2 effect:1 hence:5 assigned:1 excluded:1 read:1 iteratively:2 i2:1 ignorance:1 nettalk:1 adjacent:1 game:1 generalized:2 functioll:1 invaluable:1 consideration:1 recently:1 specialized:2 volume:2 interpretation:1 mellon:2 cambridge:2 ai:6 trivially:1 language:2 had:1 robot:10 reich:1 operating:1 pu:1 base:2 own:1 recent:1 perspective:1 tesauro:1 scenario:1 binary:1 success:1 inconsistency:3 seen:1 analyzes:3 morgan:6 preceding:1 zip:1 employed:1 maximize:1 monotonically:1 ii:2 rv:1 multiple:6 desirable:1 technical:4 offer:1 vision:1 cmu:2 physically:1 repetitive:1 represent:1 jude:1 robotics:1 achieved:1 addition:2 interval:45 appropriately:1 unlike:1 comprehensibility:2 compliance:1 facilitates:1 member:2 december:1 inconsistent:5 extracting:10 structural:1 iii:3 easy:2 variety:4 bid:1 architecture:2 opposite:1 idea:1 simplifies:1 whether:3 suffer:1 speech:2 speaking:1 york:1 constitute:1 deep:1 generally:1 clear:2 covered:4 amount:3 category:1 generate:2 outperform:1 exist:1 canonical:1 notice:6 disjoint:1 per:1 overly:1 towell:1 discrete:2 carnegie:2 express:1 key:3 four:2 threshold:1 prevent:1 verified:3 breadth:1 tenth:1 kept:1 backward:2 graph:1 pap:1 fraction:1 year:2 run:1 angle:2 powerful:2 minxk:1 reasonable:1 draw:1 decision:3 scaling:1 layer:13 bound:9 hi:3 played:1 bala:1 correspondence:2 cheng:1 refine:3 constraint:10 deficiency:1 constrain:1 sake:1 bonn:5 generates:1 speed:1 argument:1 min:2 romerstr:1 conjecture:4 department:1 waibel:1 combination:1 craven:4 describes:1 increasingly:1 character:1 projecting:2 gradually:3 computationally:1 equation:1 remains:1 previously:1 turn:1 kinematics:2 mechanism:5 neti:4 describing:1 prerequisite:1 denker:1 generic:3 enforce:1 appropriate:1 distinguished:1 alternative:2 compiling:2 original:3 neuml:2 denotes:2 assumes:1 cf:1 a4:1 quinlan:1 hypercube:2 already:1 added:1 strategy:1 damage:1 traditional:1 said:1 subspace:1 separate:1 mapped:1 thrun:8 entity:1 link:1 catcher:1 mail:1 argue:1 considers:1 barely:1 induction:2 code:1 difficult:1 carbon:1 suppress:1 implementation:1 pomerleau:1 negated:1 upper:1 benchmark:1 howard:1 extended:1 excluding:1 ever:1 arbitrary:6 mcmillan:2 bk:2 propositional:1 complement:1 namely:1 gratitude:1 clayton:1 connection:1 hanson:4 engine:1 boser:1 unsafe:4 able:2 usually:1 pattern:1 smolensky:1 max:3 oj:1 including:1 weide:1 omlin:1 rely:1 attach:1 arm:13 wik:1 minimax:1 pare:1 brief:1 created:1 catch:1 extract:3 tresp:1 loss:3 fully:1 generation:1 limitation:1 mounted:1 proven:3 acyclic:1 remarkable:1 degree:3 verification:1 consistent:2 editor:5 playing:1 land:1 squashing:1 cancer:1 elsewhere:2 changed:1 summary:1 surprisingly:1 last:1 fahlman:1 bias:1 side:3 shavlik:5 fall:2 sparse:1 distributed:9 slice:1 overcome:1 default:1 xn:1 world:2 cumulative:2 boundary:3 feedback:1 van:1 forward:3 collection:1 refinement:3 universally:1 projected:1 author:1 approximate:1 compact:3 uni:1 lippmann:4 confirm:1 ml:1 pittsburgh:2 kononenko:1 consuming:1 xi:4 search:3 continuous:1 table:6 promising:2 nature:2 transfer:3 reasonably:1 alvinn:1 complex:1 domain:8 marc:1 whole:5 paul:1 nothing:1 fig:6 structurally:1 position:2 wish:1 lie:2 candidate:1 learns:1 bij:1 admissible:4 subsymbolic:1 enlarging:2 specific:12 showing:1 list:1 a3:2 exists:2 effectively:1 illustrates:1 depicted:2 simply:1 expressed:2 partially:2 monotonic:1 aa:1 corresponds:1 determines:1 extracted:9 presentation:1 ann:1 careful:1 wikxk:6 fisher:1 specifically:1 operates:1 called:1 total:2 stake:1 experimental:2 meaningful:2 internal:4 dept:1
7,162
925
From Data Distributions to Regularization in Invariant Learning Todd K. Leen Department of Computer Science and Engineering Oregon Graduate Institute of Science and Technology 20000 N.W. Walker Rd Beaverton, Oregon 97006 tieen@cse.ogi.edu Abstract Ideally pattern recognition machines provide constant output when the inputs are transformed under a group 9 of desired invariances. These invariances can be achieved by enhancing the training data to include examples of inputs transformed by elements of g, while leaving the corresponding targets unchanged. Alternatively the cost function for training can include a regularization term that penalizes changes in the output when the input is transformed under the group. This paper relates the two approaches, showing precisely the sense in which the regularized cost function approximates the result of adding transformed (or distorted) examples to the training data. The cost function for the enhanced training set is equivalent to the sum of the original cost function plus a regularizer. For unbiased models, the regularizer reduces to the intuitively obvious choice a term that penalizes changes in the output when the inputs are transformed under the group. For infinitesimal transformations, the coefficient of the regularization term reduces to the variance of the distortions introduced into the training data. This correspondence provides a simple bridge between the two approaches. Todd Leen 22 4 1 A pproaches to Invariant Learning In machine learning one sometimes wants to incorporate invariances into the function learned. Our knowledge of the problem dictates that the machine outputs ought to remain constant when its inputs are transformed under a set of operations gl. In character recognition, for example, we want the outputs to be invariant under shifts and small rotations of the input image. In neural networks, there are several ways to achieve this invariance 1. The invariance can be hard-wired by weight sharing in the case of summation nodes (LeCun et al. 1990) or by constraints similar to weight sharing in higher-order nodes (Giles et al. 1988). 2. One can enhance the training ensemble by adding examples of inputs transformed under the desired inval"iance group, while maintaining the same targets as for the raw data. 3. One can add to the cost function a regularizer that penalizes changes in the output when the input is transformed by elements of the group (Simard et al. 1992). Intuitively one expects the approaches in 3 and 4 to be intimately linked. This paper develops that correspondence in detail. 2 The Distortion-Enhanced Input Ensemble Let the input data x be distributed according to the density function p( x). The conditional distribution for the corresponding targets is denoted p(tlx). For simplicity of notation we take t E R. The extension to vector targets is trivial. Let f(x; w) denote the network function, parameterized by weights w. The training procedure is assumed to minimize the expected squared error ?(w) = JJdtdx p(tlx) p(x) [t - f(x; w)]2 . (1) vVe wish to consider the effects of adding new inputs that are related to the old by transformations that correspond to the desired invariances. These transformations, or distortions, of the inputs are carried out by group elements g E g. For Lie groups2, the transformations are analytic functions of parameters a E Rk x -t x' = g(x;a) , (2) with the identity transformation corresponding to parameter value zero g(x;O) = x . (3) In image processing, for example, we may want our machine to exhibit invariance with respect to rotation, scaling, shearing and translations of the plane. These 1 We assume that the set forms a group. 2See for example (Sattillger, 1986). From Data Distributions to Regularization in Invariant Learning 225 transformations form a six-parameter Lie group3. By adding distorted input examples we alter the original density p( x). To describe the new density, we introduce a probability density for the transformation parameters p(a). Using this density, the distribution for the distortion-enhanced input ensemble is p(x') = j j dadx p(x'lx,a) p(a) p(x) = j jdadxt5(x'-g(x;a?p(a)p(x) where t5(.) is the Dirac delta function 4 Finally we impose that the targets remain unchanged when the inputs are transformed according to (2) i.e., p(tlx') = p(tlx). Substituting p(x') into (1) and using the invariance of the targets yields the cost function t J = j j dtdxda p(tlx)p(x)p(a) [t - f(g(x;a);w)]2 (4) Equation (4) gives the cost function for the distortion-enhanced input ensemble. 3 Regularization and Hints The remainder of the paper makes precise the connection between adding transformed inputs, as embodied in (4), and various regularization procedures. It is straightforward to show that the cost function for the distortion-enhanced ensemble is equivalent to the cost function for the original data ensemble (1) plus a regularization term. Adding and subtracting f(x; w) to the term in square brackets in (4), and expanding the quadratic leaves t = E + ER , (5) where the regularizer is ER = EH J + Ec da p(a) j dx p(x) [f(x, w) - f(g(x; a); w)]2 - 2 JJ j dtdxda p(tlx) p(x) p(a) x[t - f{x;w)] [f(g(x;a);w) - f(x;w)] (6) 3The parameters for rotations, scaling and shearing completely specify elements of G L2, the four parameter group of 2 x 2 invertible matrices. The translations carry an additional two degrees of freedom. 4 In general the density on 0 might vary through the input space, suggesting the conditional density p(o I :1'). This introduces rather minor changes in the discussion that will not be considered here. 226 ToddLeen Training with the original data ensemble using the cost function (5) is equivalent to adding transformed inputs to the data ensemble. The first term of the regularizer ?H penalizes the average squared difference between I(x;w) and I(g(x;a);w). This is exactly the form one would intuitively apply in order to insure that the network output not change under the transformation x -4 g( x, a). Indeed this is the similar to the form of the invariance "hint" proposed by Abu-Mostafa (1993). The difference here is that there is no arbitrary parameter multiplying the term. Instead the strength of the regularizer is governed by the average over the density pea). The term ?H measures the error in satisfying the invariance hint. The second term ?a measures the correlation between the error in fitting to the data, and the errol' in satisfying the hint. Only when these correlations vanish is the cost function for the enhanced ensemble equal to the original cost function plus the invariance hint penalty. The correlation term vanishes trivially when either 1. The invariance I (g( x; a); w) = I (x; w) is satisfied, or 2. The network function equals the least squares regression on t I(x; w) = J dt p(tlx) t =E[tlx] . (7) The lowest possible ? occurs when I satisfies (7), at which ? becomes the variance in the targets averaged over p( x ). By substituting this into ?a and carrying out the integration over dt p( tlx), the correlation term is seen to vanish. If the minimum of t occurs at a weight for which the invariance is satisfied (condition 1 above). then minimizing t (w) is equivalent to minimizing ? (w). If the minimum of t occurs at a weight for which the network function is the regression (condition 2), then minimizing t is equivalent to minimizing the cost function with the intuitive regularizer ? H 5. 3.1 Infinitesimal Transformations Above we enumerated the conditions under which the correlation term in ?R vanishes exactly for unrestricted transformations. If the transformations are analytic in the paranleters 0', then by restricting ourselves to small transformations (those close to the identity) we can-show how the correlation term approximately vanishes for unbiased models. To implement this, we assume that p( a) is sharply peaked up about the origin so that large transformations are unlikely. 51? the data is to be fit optimally, with enough freedom left over to satisfy the invariance hint, then there must be several weight values (perhaps a continuum of such values) for which the network function satisfies (7). That is, the problem must be under-specified. If this is the case, then the interesting part weight space is just the subset on which (7) is satisfied. On this subset the correlation term in (6) vanishes and the regularizer assumes the intuitive form. From Data Distributions to Regularization in Invariant Learning 227 t We obtain an approximation to the cost function by expanding the integrands in (6) in power series about 0 = 0 and retaining terms to second order. This leaves t = c+ JJ JJJ dxdo p(x) p(o) (Oi :~ L=o :!" f dt dx do p(tlx)p(x)p(o) [t-f(x;w)] x -2 [ ( Og"l 0?- ? OOi 0=0 0 2 gIl 1 2' +-0?0? J OOi (0) I0=0 )(ax"of) - (8) where x P and gP denote the pth components of x and g, OJ denotes the ith component of the transformation parameter vector, repeated Greek and Roman indices are summed over, and all derivatives are evaluated at 0 o. Note that we have used the fact that Lie group transformations are analytic in the parameter vector o to derive the expansion. = Finally we introduce two assumptions on the distribution p(o). First 0 is assumed to be zero mean. This corresponds, in the linear approximation, to a distribution of distortions whose mean is the identity transformation. Second, we assume that the components of 0 are uncorrelated so that the covariance matrix is diagonal with elements ul, i = 1 ... k. 6 With these assumptions, the cost function for the distortion-enhanced ensemble simplifies to t = J p( - L (T~ JJ c+~ (Tr ~ .=1 dx x) ( ~g"l va. a=O : f vx" ) 2 k dx dt p(tlx) p(X) { (f(x; w) - t ) .=1 X [ ~:; 10=0 ( :~ ) + :~ L=o :~ L=o (ox~2?x" )]} This last expression provides a simple bridge between the the methods of adding transformed examples to the data, and the alternative of adding a regularizer to the cost function: The coefficient of the regularization term in the latter approach is equal to the variance of the transformation parameters in the former approach. 6Note that the transformed patterns may be correlated in parts of the pattern space. For example the results of applying the shearing and rotation operations to an infinite vertical line are indistinguishable. In general, there may be regions of the pattern space for which the action of several different group elements are indistinguishable; that is x' = g(x; a) = g(x; (3). However this does not imply that a and (3 are statistically correlated. 228 3.1.1 Todd Leen Unbiased Models For unbiased models the regularizer in E( w) assumes a particularly simple form. Suppose the network function is rich enough to form an unbiased estimate of the least squares regression on t for the un distorted data ensemble. That is, there exists a weight value Wo such that f(x;wo) =J (10) dt tp(tlx) == E[tlx] This is the global minimum for the original error ?( w). The arguments of section 3 apply here as well. However we can go further. Even if there is only a single, isolated weight value for which (10) is satisfied, then to O( 0- 2 ) the correlation term in the regularizer vanishes. To see this note that by the implicit function theorem the modified cost function (9) has its global minimum at the new weight 7 (11) At this weight, the network function is no longer the regression on t, but rather f(x;wo) = E[tlx] + 0(0- 2 ) (12) ? Substituting (12) into (9), we find that the minimum of (9) is, to 0(0- 2 ), at the same weight as the minimum of t = ? + L.k .=1 o-~ JdX p(x) [ oglJ 0 Q'j I Q=O of (x, w) ] 2 oxlJ To 0(0- 2 ), minimizing (13) is equivalent to minimizing (9). So we regard effective cost function. (13) t as the The regularization term in (13) is proportional to the average square of the gradient of the network function along the direction in the input space generated by the lineal' part of g. The quantity inside the square brackets is just the linear part of [f (g( X; Q')) - f (x)] from (6). The magnitude of the regularization term is just the variance of the distribution of distortion parameters. This is precisely the form of the regularizer given by Simard et al. in their tangent prop algorithm (Simard et aI, 1992). This derivation shows the equivalence (to 0(0"2)) between the tangent prop regularizer and the alternative of modifying the input distribution. Furthermore, we see that with this equivalence, the constant fixing the strengt.h of the regularization term is simply the variance of the distortions introduced into the original training set. We should stress that the equivalence between the regularizer, and the distortionenhanced ensemble in (13) only holds to 0(0- 2 ). If one allows the variance of the 7We assume that the Hessian of ? is nonsingular at woo From Data Distributions to Regularization in Invariant Learning 229 distortion parameters u 2 to become arbitrarily large in an effort to mock up an arbitrarily large regularization term, then the equivalence expressed in (13) breaks down since terms of order O( ( 4 ) can no longer be neglected. In addition, if the transformations are to be kept small so that the linearization holds (e.g. by restricting the density on a to have support on a small neighborhood of zero), then the variance will bounded above. 3.1.2 Smoothing Regularizers In the previous sections we showed the equivalence between modifying the input distribution and adding a regularizer to the cost function. We derived this equivalence to illuminate mechanisms for obtaining invariant pattern recognition. The technique for dealing with infinitesimal transformations in section ?3.1 was used by Bishop (1994) to show the equivalence between added input noise and smoothing regularizers. Bishop's results, though they preceded our own, are a special case of the results presented here. Suppose the group 9 is restricted to translations by random vectors g( X; a) X + a where a is spherically distributed with variance u!. Then the regularizer in (13) is = (14) This regularizer penalizes large magnitude gradients in the network function and is, as pointed out by Bishop, one of the class of generalized Tikhonov regularizers. 4 Summary We have shown that enhancing the input ensemble by adding examples transformed under a group x -? g(x;a), while maintaining the target values, is equivalent to adding a regularizer to the original cost function. For unbiased models the regulatizer reduces to the intuitive form that penalizes the mean squared difference between the network output for transformed and untransformed inputs - i.e. the error in satisfying the desired invariance. In general the regularizer includes a term that measures correlations between the error in fitting the data, and the error in satisfying the desired inva.riance. For infinitesimal transformations, the regularizer is equivalent (up to terms linear in the variance of the transformation parameters) to the tangent prop form given by Simard et a1. (1992), with regularization coefficient equal to the variance of the transformation parameters. In the special case that the group transformations are limited to random translations of the input, the regularizer reduces to a standard smoothing regularizer. \Ve gave conditions under which enhancing the input ensemble and adding the intuitive regularizer ?H are equivalent. However tins equivalence is only with regard to the optimal weight. We have not compared the training dynamics for the two approaches. In particular, it is quite possible that the full regularizer ?H + ?c exhibits different training dynamics from the intuitive form ?H. For the approach in which data are added to the input ensemble, one can easily construct datasets and distributions p( a) that either increase the condition number of the Hessian, or decrease it. Finally, it may be that the intuitive regularizer can have either detrimental or positive effects on the Hessian as well. 230 ToddLeen Acknowledgments I thank Lodewyk Wessels, Misha Pavel, Eric Wan, Steve Rehfuss, Genevieve Orr and Patrice Simard for stimulating and helpful discussions, and the reviewers for helpful comments. I am grateful to my father for what he gave to me in life, and for the presence of his spirit after his recent passing. This work was supported by EPRI under grant RP8015-2, AFOSR under grant FF4962-93-1-0253, and ONR under grant N00014-91-J-1482. References Yasar S. Abu-Mostafa. A method for learning from hints. In S. Hanson, J. Cowan, and C. Giles, editors, Advances in Neural Information Processing Systems, vol. 5, pages 73-80. Morgan Kaufmann, 1993. Chris M. Bishop. Training with noise is equivalent to Tikhonov regularization. To appear in Neural Computation, 1994. C.L. Giles, R.D. Griffin, and T. Maxwell. Encoding geometric invariances in higherorder neural networks. In D.Z.Anderson, editor, Neural Information Processing Systems, pages 301-309. American Institute of Physics, 1988. Y. Le Cun, B. Boser, J.S. Denker, D. Henderson, R.E. Howard, W. Hubbard, and L.D. Jackel. Handwritten digit recognition with a back-propagation network. In Advances in Neural Information Processing Systems, vol. 2, pages 396-404. Morgan Kaufmann Publishers, 1990. Patrice Simard, Bernard Victorri, Yann Le Cun, and John Denker. Tangent prop a formalism for specifying selected invariances in an adaptive network. In John E. Moody, Steven J. Hanson, and Richard P. Lippmann, editors, Advances in Neural Information Processing Systems 4, pages 895-903. Morgan Kaufmann, 1992. D.H. Sattinger and O.L. Weaver. Lie Groups and Algebras with Applications to Physics, Geometry and Mechanics. Springer-Verlag, 1986.
925 |@word covariance:1 pavel:1 tr:1 carry:1 series:1 dx:4 must:2 john:2 analytic:3 leaf:2 selected:1 plane:1 ith:1 provides:2 node:2 cse:1 lx:1 along:1 become:1 fitting:2 inside:1 introduce:2 shearing:3 indeed:1 expected:1 mock:1 mechanic:1 becomes:1 notation:1 insure:1 bounded:1 lowest:1 what:1 transformation:23 ought:1 ooi:2 exactly:2 grant:3 appear:1 positive:1 engineering:1 todd:3 encoding:1 untransformed:1 approximately:1 might:1 plus:3 equivalence:8 specifying:1 limited:1 graduate:1 statistically:1 averaged:1 acknowledgment:1 lecun:1 implement:1 digit:1 procedure:2 dictate:1 close:1 applying:1 equivalent:10 reviewer:1 straightforward:1 go:1 simplicity:1 his:2 target:8 enhanced:7 suppose:2 origin:1 element:6 recognition:4 satisfying:4 particularly:1 steven:1 region:1 decrease:1 vanishes:5 ideally:1 dynamic:2 neglected:1 carrying:1 grateful:1 algebra:1 eric:1 completely:1 easily:1 various:1 regularizer:25 derivation:1 describe:1 effective:1 neighborhood:1 whose:1 quite:1 distortion:11 gp:1 patrice:2 subtracting:1 remainder:1 achieve:1 intuitive:6 dirac:1 wired:1 derive:1 fixing:1 minor:1 direction:1 greek:1 modifying:2 pea:1 vx:1 summation:1 enumerated:1 extension:1 hold:2 considered:1 mostafa:2 substituting:3 vary:1 continuum:1 jackel:1 bridge:2 hubbard:1 modified:1 rather:2 og:1 ax:1 derived:1 rehfuss:1 sense:1 am:1 helpful:2 i0:1 unlikely:1 transformed:15 denoted:1 retaining:1 smoothing:3 integration:1 summed:1 special:2 equal:4 construct:1 integrands:1 alter:1 peaked:1 develops:1 hint:7 roman:1 richard:1 ve:1 geometry:1 ourselves:1 tlx:14 freedom:2 genevieve:1 henderson:1 introduces:1 bracket:2 misha:1 regularizers:3 old:1 penalizes:6 desired:5 isolated:1 formalism:1 giles:3 tp:1 cost:20 subset:2 expects:1 father:1 optimally:1 my:1 density:9 physic:2 invertible:1 enhance:1 moody:1 squared:3 satisfied:4 wan:1 rp8015:1 american:1 simard:6 derivative:1 suggesting:1 orr:1 includes:1 coefficient:3 oregon:2 satisfy:1 break:1 linked:1 minimize:1 square:5 oi:1 variance:10 kaufmann:3 ensemble:15 correspond:1 yield:1 nonsingular:1 raw:1 handwritten:1 multiplying:1 errol:1 sharing:2 infinitesimal:4 obvious:1 knowledge:1 back:1 steve:1 higher:1 dt:5 maxwell:1 specify:1 leen:3 evaluated:1 ox:1 though:1 anderson:1 furthermore:1 just:3 implicit:1 correlation:9 propagation:1 perhaps:1 effect:2 unbiased:6 former:1 regularization:16 spherically:1 ogi:1 indistinguishable:2 generalized:1 stress:1 image:2 rotation:4 preceded:1 he:1 approximates:1 ai:1 rd:1 trivially:1 pointed:1 longer:2 add:1 own:1 showed:1 recent:1 tikhonov:2 n00014:1 verlag:1 onr:1 arbitrarily:2 life:1 seen:1 minimum:6 additional:1 unrestricted:1 impose:1 morgan:3 relates:1 full:1 reduces:4 a1:1 va:1 regression:4 enhancing:3 sometimes:1 achieved:1 addition:1 want:3 victorri:1 walker:1 leaving:1 publisher:1 comment:1 cowan:1 spirit:1 presence:1 enough:2 fit:1 gave:2 simplifies:1 shift:1 six:1 expression:1 ul:1 effort:1 penalty:1 wo:3 hessian:3 passing:1 jj:3 action:1 gil:1 delta:1 vol:2 abu:2 group:14 four:1 kept:1 sum:1 parameterized:1 distorted:3 jdx:1 yann:1 griffin:1 scaling:2 correspondence:2 quadratic:1 strength:1 precisely:2 sharply:1 constraint:1 lineal:1 argument:1 jjj:1 department:1 according:2 remain:2 riance:1 character:1 intimately:1 cun:2 intuitively:3 invariant:7 restricted:1 equation:1 mechanism:1 operation:2 apply:2 denker:2 alternative:2 original:8 assumes:2 denotes:1 include:2 beaverton:1 maintaining:2 unchanged:2 added:2 quantity:1 occurs:3 diagonal:1 illuminate:1 exhibit:2 gradient:2 detrimental:1 thank:1 higherorder:1 me:1 chris:1 trivial:1 index:1 minimizing:6 vertical:1 datasets:1 howard:1 precise:1 arbitrary:1 introduced:2 specified:1 connection:1 hanson:2 learned:1 boser:1 pattern:5 oj:1 power:1 eh:1 regularized:1 weaver:1 technology:1 imply:1 carried:1 woo:1 embodied:1 geometric:1 l2:1 tangent:4 afosr:1 interesting:1 proportional:1 degree:1 editor:3 uncorrelated:1 translation:4 summary:1 gl:1 last:1 supported:1 institute:2 distributed:2 regard:2 rich:1 t5:1 adaptive:1 pth:1 ec:1 lippmann:1 dealing:1 global:2 assumed:2 alternatively:1 un:1 expanding:2 obtaining:1 expansion:1 da:1 noise:2 repeated:1 vve:1 wish:1 lie:4 governed:1 vanish:2 tin:1 rk:1 theorem:1 down:1 bishop:4 showing:1 er:2 exists:1 restricting:2 adding:13 magnitude:2 linearization:1 simply:1 expressed:1 springer:1 corresponds:1 satisfies:2 prop:4 stimulating:1 conditional:2 identity:3 change:5 hard:1 infinite:1 bernard:1 invariance:17 support:1 latter:1 incorporate:1 correlated:2
7,163
926
On the Computational Complexity of Networks of Spiking Neurons (Extended Abstract) Wolfgang Maass Institute for Theoretical Computer Science Technische Universitaet Graz A-80lO Graz, Austria e-mail: maass@igi.tu-graz.ac.at Abstract We investigate the computational power of a formal model for networks of spiking neurons, both for the assumption of an unlimited timing precision, and for the case of a limited timing precision. We also prove upper and lower bounds for the number of examples that are needed to train such networks. 1 Introduction and Basic Definitions There exists substantial evidence that timing phenomena such as temporal differences between spikes and frequencies of oscillating subsystems are integral parts of various information processing mechanisms in biological neural systems (for a survey and references see e.g. Abeles, 1991; Churchland and Sejnowski, 1992; Aertsen, 1993). Furthermore simulations of a variety of specific mathematical models for networks of spiking neurons have shown that temporal coding offers interesting possibilities for solving classical benchmark-problems such as associative memory, binding, and pattern segmentation (for an overview see Gerstner et al., 1992). Some aspects of these models have also been studied analytically, but almost nothing is known about their computational complexity (see Judd and Aihara, 1993, for some first results in this direction). In this article we introduce a simple formal model SNN for networks of spiking neurons that allows us to model the most important timing phenomena of neural nets (including synaptic modulation), and we prove upper and lower bounds for its computational power and learning complexity. Further 184 Wolfgang Maass details to the results reported in this article may be found in Maass, 1994a,1994b, 1994c. An SNN N consists of - a finite directed graph {V, E} (we refer to the elements of V as "neurons" and to the elements of E as "synapses") - a subset Yin S; V of input neurons - a subset Vout S; V of output neurons Definition of a Spiking Neuron Network (SNN): - for each neuron v E V - Yin a threshold-function 9 v : R+ {where R+ := {x E R : x ~ O}) - for each synapse {u, v} E E a response-function weight- function Wu,v : R+ -+ R . ?u,v : -+ R+ R U {oo} -+ R and a We assume that the firing of the input neurons v E Yin is determined from outside of N, i.e. the sets Fv S; R+ of firing times ("spike trains") for the neurons v E Yin are given as the input of N. Furthermore we assume that a set T S; R+ of potential firing times 7iiiSfjeen fixed. For a neuron v E V - Yin one defines its set Fv of firing times recursively. The first element of Fv is inf{t E T : Pv(t) ~ 0 v (O)} , and for any s E Fv the next larger element of Fv is inf{t E T : t > sand Pv(t) ~ 0 v (t - s)} ,where the potential function Pv : R+ -+ R is defined by Pv(t) := 0 + L L wu,v(s) . ?u,v(t - s) u : {u, v} E EsE Fu : s < t The firing times ("spike trains") Fv of the output neurons v E Vout that result in this way are interpreted as the output of N. Regarding the set T of potential firing times we consider in this article the case T = R+ (SNN with continuous time) and the case T = {i? JJ : i E N} for some JJ with 1/ JJ E N (SNN with discrete time). We assume that for each SNN N there exists a bound TN E R with TN > 0 such that 0 v(x) = 00 for all x E (0, TN) and all v E V - Yin (TN may be interpreted as the minimum of all "refractory periods" Tref of neurons in N). Furthermore we assume that all "input spike trains" Fv with v E Yin satisfy IFv n [0, t]l < 00 for all t E R+. On the basis of these assumptions one can also in the continuous case easily show that the firing times are well-defined for all v E V - Yin (and occur in distances of at least TN)' Input- and Output-Conventions: For simulations between SNN's and Turing machines we assume that the SNN either gets an input (or produces an output) from {O, 1}* in the form of a spike-train (i.e. one bit per unit of time), or encoded into the phase-difference of just two spikes. Real-valued input or output for an SNN is always encoded into the phase-difference of two spikes. Remarks a) In models for biological neural systems one assumes that if x time-units have On the Computational Complexity of Networks of Spiking Neurons /85 passed since its last firing, the current threshold 0 11 (z) of a neuron v is "infinite" for z < TreJ (where TreJ = refractory period of neuron v), and then approaches quite rapidly from above some constant value. A neuron v "fires" (i.e. it sends an "action potential" or "spike" along its axon) when its current membrane potential PII (t) at the axon hillock exceeds its current threshold 0 11 . PII (t) is the sum of various postsynaptic potentials W U ,II(S). t: U ,II(t - s). Each of these terms describes an excitatory (EPSP) or inhibitory (IPSP) postsynaptic potential at the axon hillock of neuron v at time t, as a result of a spike that had been generated by a "presynaptic" neuron u at time s, and which has been transmitted through a synapse between both neurons. Recordings of an EPSP typically show a function that has a constant value c (c = resting membrane potential; e.g. c = -70m V) for some initial time-interval (reflecting the axonal and synaptic transmission time), then rises to a peak-value, and finally drops back to the same constant value c. An IPSP tends to have the negative shape of an EPSP. For the sake of mathematical simplicity we assume in the SNN-model that the constant initial and final value of all response-functions t: U ,1I is equal to 0 (in other words: t: U ,1I models the difference between a postsynaptic potential and the resting membrane potential c). Different presynaptic neurons u generate postsynaptic potentials of different sizes at the axon hillock of a neuron v, depending on the size, location and current state of the synapse (or synapses) between u an? v. This effect is modelled by the weight-factors W U ,II(S). The precise shapes of threshold-, response-, and weight-functions vary among different biological neural systems, and even within the same system. Fortunately one can prove significant upper bounds for the computational complexity of SNN's N without any assumptions about the specific shapes of these functions of N. Instead, we only assume that they are of a reasonably simple mathematical structure. b) In order to prove lower bounds for the computational complexity of an SNN N one is forced to make more specific assumptions about these functions . All lower bound results that are reported in this article require only some rather weak basic assumptions about the response- and threshold-functions. They mainly require that EPSP's have some (arbitrarily short) segment where they increase linearly, and some (arbitrarily short) segment where they decrease linearly (for details see Maass, 1994a, 1994b). c) Although the model SNN is apparently more "realistic" than all models for biological neural nets whose computational complexity has previously been analyzed, it deliberately sacrifices a large number of more intricate biological details for the sake of mathematical tractability. Our model is closely related to those of (Buhmann and Schulten, 1986), and (Gerstner, 1991, 1992). Similarly as in (Buhmann and Schulten, 1986) we consider here only the deterministic case. d) The model SNN is also suitable for investigating algorithms that involve synaptic modulation at various time-scales. Hence one can investigate within this framework not only the complexity of algorithms for supervised and unsupervised learning, but also the potential computational power of rapid weight-changes within the course of a computation. In the theorems of this paper we allow that the value of a weight W U,II(S) at a firing time s E Fu is defined by an algebraic computation tree (see van Leeuwen, 1990) in terms of its value at previous firing times s' E Fu with s' < s, some preceding firing times s < s of arbitrary other neurons, and arbitrary realvalued parameters. In this way WU,II(S) can be defined by different rational functions /86 Wolfgang Maass of the abovementioned arguments, depending on the numerical relationship between these arguments (which can be evaluated by comparing first the relative size of arbitrary rational functions of these arguments). As a simple special case one can for example increase wu ?tI (perhaps up to some specified saturation-value) as long as neurons u and v fire coherently, and decrease w u ?tI otherwise. For the sake of simplicity in the statements of our results we assume in this extended abstract that the algebraic computation tree for each weight w U ? tI involves only 0(1) tests and rational functions of degree 0(1) that depend only on 0(1) of the abovementioned arguments. Furthermore we assume in Theorems 3, 4 and 5 that either each weight is an arbitrary time-invariant real, or that each current weight is rounded off to bit-length poly(1ogpN') in binary representation, and does not depend on the times of firings that occured longer than time 0(1) ago. Furthermore we assume in Theorems 3 and 5 that the parameters in the algebraic computation tree are rationals of bit-length O(1ogpN'). e) It is well-known that the Vapnik-Chervonenkis dimension {"VC-dimension"} of a neural net N (and the pseudo-dimension for the case of a neural net N with realvalued output, with some suitable fixed norm for measuring the error) can be used to bound the number of examples that are needed to train N (see Haussler, 1992). Obviously these notions have to be defined differently for a network with timedependent weights. We propose to define the VC-dimension (pseudo-dimension)of an SNN N with time-dependent weights as the VC-dimension (pseudo-dimension) of the class of all functions that can be computed by N with different assignments of values to the real-valued (or rational-valued) parameters of N that are involved in the definitions of the piecewise rational response-, threshold-, and weight-functions of N. In a biological neural system N these parameters might for example reflect the concentrations of certain chemical substances that are known to modulate the behavior of N. f) The focus in the investigation of computations in biological neural systems differs in two essential aspects from that of classical computational complexity theory. First, one is not only interested in single computations of a neural net for unrelated inputs z, but also in its ability to process an interrelated sequence ?(z( i), y( i)} )ieN of inputs and outputs, which may for example include an initial training sequence for learning or associative memory. Secondly, exact timing of computations is allimportant in biological neural nets, and many tasks have to be solved within a specific number of steps. Therefore an analysis in terms of the notion of a real-time computation and real-time simulation appears to be more adequate for models of biological neural nets than the more traditional analysis via complexity classes. One says that a sequence ?(z(i),y(i)})ieN is processed in real-time by a machine M, if for every i E N the machine M outputs y( i) within a constant number c of computation steps after having received input z(i). One says that M' simulates M in real-time (with delay factor ~), if every sequence that is processed in real-time by M (with some constant c), can also be processed in real-time by M' (with a constant ~ . c). For SNN's M we count each spike in M as a computation step. These definitions imply that a real-time simulation of M by M' is a special case of a linear-time simulation, and hence that any problem that can be solved by M with a certain time complexity ten), can be solved by M' with time complexity O(t(n? On the Computational Complexity of Networks of Spiking Neurons 187 (see Maass, 1994a, 1994b, for details). 2 Networks of Spiking Neurons with Continuous Time Theorem 1: If the response- and threshold-functions of the neurons satisfy some rather weak basic assumptions (see Maass, 1994a, 1994b), then one can build from such neurons for any given dEN an SNN NTM(d) of finite size with rational delays that can simulate with a suitable assignment of rational values from [0, 1] to its weights any Turing machine with at most d tapes in real-time. Furthermore NTM(2) can compute any function F : {0,1}* -- {0,1}* with a suitable assignment of real values from [0,"1] to its weights. The fixed SNN NTM(d) of Theorem 1 can simulate Turing machines whose tape content is much larger than the size of NTM (d), by encoding such tape content into the phase-difference between two oscillators. The proof of Theorem 1 transforms arbitrary computations of Turing machines into operations on such phase-differences. The last part of Theorem 1 implies that the VC-dimension of some finite SNN's is infinite. In contrast to that the following result shows that one can give finite bounds for the VC-dimension of those SNN's that only use a bounded numbers of spikes in their computation. Furthermore the last part of the claim of Theorem 2 implies that their VC-dimension may in fact grow linearly with the number S of spikes that occur in a computation. Theorem 2: The VC-dimension and pseudo-dimension of any SNN N with piecewise linear response- and threshold-functions, arbitrary real-valued parameters and time-dependent weights (as specified in section 1) can be bounded (even for realvalued inputs and outputs) by D(IEI . WI . S(log IVI + log S? if N uses in each computation at most S spikes. Furthermore one can construct SNN's (with any response- and threshold-functions that satisfy our basic assumptions, with fixed rational parameters and rational timeinvariant weights) whose VC-dimension is for computations with up to S spikes as large as O(IEI . S). We refer to Maass, 1994a, 1994c, for upper bounds on the computational power of SNN's with continuous time. 3 Networks of Spiking Neurons with Discrete Time In this section we consider the case where all firing times of neurons in N are multiples of some J.l with 1/ J.l EN. We restrict our attention to the biologically plausible case where there exists some tN ~ 1 such that for all z > tN all response functions ?U,II(Z) have the value and all threshold functions ell(z) have some arbitrary constant value. If tN is chosen minimal with this property, we refer to PN := rtN/J.ll as the timing-precision ofN. Obviously for PN = 1 the SNN is equivalent to a "non-spiking" neural net that consists of linear threshold gates, whereas a SNN with continuous time may be viewed as the opposite extremal case for PN -- 00. ? 188 Wolfgang Maass The following result provides a significant upper bound for the computational power of an SNN with discrete time, even in the presence of arbitrary real-valued parameters and weights. Its proof is technically rather involved. Theorem 3: Assume that N is an SNN with timing-precision PJII, piecewise polynomial response- and piecewise rational threshold-functions with arbitrary real-valued parameters, and weight-functions as specified in section 1. Then one can simulate N for boolean valued inputs in real-time by a Turing machine with poly(lVl, logpJII,log l/TJII) states and poly(lVl, logpJII, tJII/TJII) tape-cells. On the other hand any Turing machine with q states that uses at most s tapecells can be simulated in real-time by an SNN N with any response- and thresholdfunctions that satisfy our basic assumptions, with rational parameters and timeinvariant rational weights, with O(q) neurons, logpJII = O(s), and tJII/TJII = 0(1). The next result shows that the VC-dimension of any SNN with discrete time is finite, and grows proportionally to logpJII. The proof of its lower bound combines a new explicit construction with that of Maass, 1993. Theorem 4: Assume that the SNN N has the same properties as in Theorem 3. Then the VC-dimension and the pseudo-dimension of N (for arbitrary real valued inputs) can be bounded by O(IEI?IVI?logpJII), independently of the number of spikes in its computations. Furthermore one can construct SNN's N of this type with any response- and threshold-functions that satisfy our basic assumptions, with rational parameters and time-invariant rational weights, so that N has (already for boolean inputs) a VCdimension of at least O(IEI(logpJII + log IE!?. 4 Relationships to other Computational Models We consider here the relationship between SNN's with discrete time and recurrent analog neural nets. In the latter no "spikes" or other non-trivial timing-phenomena occur, but the output of a gate consists of the "analog" value of some squashingor activation function that is applied to the weighted sum of its inputs. See e.g. (Siegelmann and Sontag, 1992) or (Maass, 1993) for recent results about the computational power of such models. We consider in this section a perhaps more "realistic" version of such modelsN, where the output of each gate is rounded off to an integer multiple of some ~ (with a EN). We refer to a as the number of activation levels of N. It is an interesting open problem whether such analog neural nets (with gate-outputs interpreted as firing rates) or networks of spiking neurons provide a more adequate computational model for biological neural systems. Theorem 5 shows that in spite of their quite different structure the computational power of these two models is in fact closely related. On the side the following theorem also exhibits a new subclass of deterministic finite automata (DFA's) which turns out to be of particular interest in the context of neural nets. We say that a DFA M is a sparse DFA of size s if M can be realized by a Turing machine with s states and space-bound s (such that each step of M corresponds to one step of the Turing machine). Note that a sparse DFA may have exponentially in s many states, but that only poly(s) bits are needed to describe its On the Computational Complexity of Networks of Spiking Neurons 189 transition function. Sparse DFA's are relatively easy to construct, and hence are very useful for demonstrating (via Theorem 5) that a specific task can be carried out on a "spiking" neural net with a realistic timing precision (respectively on an analog neural net with a realistic number of activation levels). Theorem 5: The following classes of machines have closely related computational power in the sense that there is a polynomial p such that each computational model from any of these classes can be simulated in real-time (with delay-factor ~ p(s?) by some computational model from any other class (with the size-parameter s replaced by p(s?): ? sparse DFA's of size s ? SNN's with 0(1) neurons and timing precision 2 3 ? recurrent analog neural nets that consist of O( 1) gates with piecewise rational activation functions with 23 activation levels, and parameters and weights of bit-length $ s ? neural nets that consist of s linear threshold gates (with recurrencies) with arbitrary real weights. The result of Theorem 5 is remarkably stable since it holds no matter whether one considers just SNN's N with 0(1) neurons that employ very simple fixed piecewise linear response- and threshold-functions with parameters of bit-length 0(1) (with tN/TN = 0(1) and time-invariant weights of bit-length $ s), or if one considers SNN's N with s neurons with arbitrary piecewise polynomial response- and piecewise rational threshold-functions with arbitrary real-valued parameters, tN/TN ~ s, and time-dependent weights (as specified in section 1). 5 Conclusion We have introduced a simple formal model SNN for networks of spiking neurons, and have shown that significant bounds for its computational power and sample complexity can be derived from rather weak assumptions about the mathematical structure of its response-, threshold-, and weight-functions. Furthermore we have established quantitative relationships between the computational power of a model for networks of spiking neurons with a limited timing precision (i.e. SNN's with discrete time) and a quite realistic version of recurrent analog neural nets (with a bounded number of activation levels). The simulations which provide the proof of this result create an interesting link between computations with spike-coding (in an SNN) and computations with frequency-coding (in analog neural nets). We also have established such relationships for the case of SNN's with continuous time (see Maass 1994a, 1994b, 1994c), but space does not permit to report these results in this article. The Theorems 1 and 5 of this article establish the existence of mechanisms for simulating arbitrary Turing machines (and hence any common computational model) on an SNN. As a consequence one can now demonstrate that a concrete task (such as binding, pattern-matching, associative memory) can be carried out on an SNN by simply showing that some arbitrary common computational model can carry out that task. Furthermore one can bound the required timing-precision of the SNN in terms of the space needed on a Turing machine. 190 Wolfgang Maass Since we have based our investigations on the rather refined notion of a real-time simulation, our results provide information not only about the possibility to implement computations, but also adaptive behavior on networks of spiking neurons. Acknowledgement I would like to thank Wulfram Gerstner for helpful discussions. References M. Abeles. (1991) Corticonics: Neural Circuits of the Cerebral Cortex. Cambridge University Press. A. Aertsen. ed. (1993) Brain Theory: Spatio-Temporal Aspects of Brain Function. Elsevier. J. Buhmann, K. Schulten. (1986) Associative recognition and storage in a model network of physiological neurons. Bioi. Cybern. 54: 319-335. P. S. Churchland, T. J. Sejnowski. (1992) The Computational Brain. MIT-Press. W. Gerstner. (1991) Associative memory in a network of "biological" neurons. Advances in Neural Information Processing Systems, vol. 3, Morgan Kaufmann: 84-90. W. Gerstner, R. Ritz, J. L. van Hemmen. (1992) A biologically motivated and analytically soluble model of collective oscillations in the cortex. Bioi. Cybern. 68: 363-374. D. Haussler. (1992) Decision theoretic generalizations of the PAC model for neural nets and other learning applications. Inf and Comput. 95: 129-161. K. T. Judd, K. Aihara. (1993) Pulse propagation networks: A neural network model that uses temporal coding by action potentials. Neural Networks 6: 203-215. J. van Leeuwen, ed. (1990) Handbook of Theoretical Computer Science, vol. A: Algorithms and Complexity. MIT-Press. W. Maass. (1993) Bounds for the computational power and learning complexity of analog neural nets. Proc. 25th Annual ACM Symposium on the Theory of Computing, 335-344. W. Maass. (1994a) On the computational complexity of networks of spiking neurons (extended abstract). TR 393 from May 1994 of the Institutes for Information Processing Graz (for a more detailed version see the file maass.spiking.ps.Z in the neuroprose archive). W. Maass. (1994b) Lower bounds for the computational power of networks of spiking neurons. Neural Computation, to appear. W. Maass. (1994c) Analog computations on networks of spiking neurons (extended abstract). Submitted for publication. H. T. Siegelmann, E. D. Sontag. (1992) On the computational power of neural nets. Proc. 5th ACM- Workshop on Computational Learning Theory, 440-449.
926 |@word version:3 polynomial:3 norm:1 open:1 simulation:7 pulse:1 tr:1 recursively:1 carry:1 initial:3 chervonenkis:1 current:5 comparing:1 activation:6 numerical:1 realistic:5 shape:3 drop:1 short:2 provides:1 location:1 mathematical:5 along:1 symposium:1 prove:4 consists:3 combine:1 introduce:1 sacrifice:1 intricate:1 rapid:1 behavior:2 brain:3 snn:42 unrelated:1 bounded:4 circuit:1 interpreted:3 temporal:4 pseudo:5 every:2 quantitative:1 ti:3 subclass:1 unit:2 appear:1 timing:12 tends:1 consequence:1 encoding:1 firing:14 modulation:2 might:1 studied:1 limited:2 directed:1 timedependent:1 implement:1 differs:1 matching:1 word:1 spite:1 get:1 subsystem:1 storage:1 context:1 cybern:2 equivalent:1 deterministic:2 attention:1 independently:1 automaton:1 survey:1 simplicity:2 haussler:2 ritz:1 notion:3 construction:1 exact:1 us:3 element:4 recognition:1 ien:2 solved:3 graz:4 decrease:2 substantial:1 complexity:18 depend:2 solving:1 segment:2 churchland:2 technically:1 basis:1 easily:1 differently:1 various:3 train:6 forced:1 describe:1 sejnowski:2 outside:1 refined:1 quite:3 encoded:2 larger:2 valued:9 whose:3 say:3 plausible:1 otherwise:1 ability:1 ifv:1 final:1 associative:5 obviously:2 sequence:4 net:20 propose:1 epsp:4 tu:1 rapidly:1 transmission:1 p:1 oscillating:1 produce:1 lvl:2 oo:1 depending:2 ac:1 vcdimension:1 recurrent:3 received:1 involves:1 implies:2 pii:2 convention:1 direction:1 closely:3 vc:10 sand:1 require:2 generalization:1 investigation:2 biological:11 secondly:1 hold:1 claim:1 vary:1 proc:2 extremal:1 create:1 weighted:1 mit:2 always:1 rather:5 pn:3 publication:1 derived:1 focus:1 mainly:1 contrast:1 sense:1 helpful:1 elsevier:1 dependent:3 typically:1 interested:1 among:1 special:2 ell:1 equal:1 construct:3 having:1 corticonics:1 unsupervised:1 report:1 piecewise:8 employ:1 replaced:1 phase:4 fire:2 interest:1 investigate:2 possibility:2 analyzed:1 integral:1 fu:3 tree:3 theoretical:2 leeuwen:2 minimal:1 boolean:2 measuring:1 assignment:3 tractability:1 technische:1 subset:2 delay:3 reported:2 hillock:3 abele:2 peak:1 ie:1 off:2 rounded:2 concrete:1 reflect:1 soluble:1 potential:13 iei:4 coding:4 matter:1 satisfy:5 igi:1 wolfgang:5 apparently:1 kaufmann:1 vout:2 modelled:1 weak:3 ago:1 submitted:1 synapsis:2 synaptic:3 ed:2 definition:4 trej:2 frequency:2 involved:2 proof:4 rational:17 austria:1 occured:1 segmentation:1 reflecting:1 back:1 appears:1 supervised:1 response:15 synapse:3 evaluated:1 furthermore:11 just:2 hand:1 propagation:1 defines:1 perhaps:2 grows:1 effect:1 deliberately:1 analytically:2 hence:4 chemical:1 maass:19 ll:1 theoretic:1 demonstrate:1 tn:12 common:2 spiking:20 overview:1 refractory:2 exponentially:1 cerebral:1 analog:9 resting:2 refer:4 significant:3 cambridge:1 similarly:1 had:1 stable:1 longer:1 cortex:2 recent:1 inf:3 certain:2 binary:1 arbitrarily:2 transmitted:1 minimum:1 fortunately:1 morgan:1 preceding:1 period:2 ntm:4 ii:6 multiple:2 tjii:5 exceeds:1 offer:1 long:1 basic:6 cell:1 whereas:1 remarkably:1 interval:1 grow:1 sends:1 ivi:2 file:1 archive:1 recording:1 simulates:1 integer:1 axonal:1 presence:1 easy:1 variety:1 restrict:1 opposite:1 regarding:1 whether:2 motivated:1 ese:1 passed:1 algebraic:3 sontag:2 tape:4 action:2 remark:1 jj:3 adequate:2 useful:1 dfa:6 detailed:1 proportionally:1 involve:1 transforms:1 ten:1 processed:3 generate:1 inhibitory:1 per:1 discrete:6 vol:2 threshold:17 demonstrating:1 graph:1 sum:2 turing:10 almost:1 wu:4 oscillation:1 decision:1 bit:7 bound:16 annual:1 occur:3 unlimited:1 sake:3 aspect:3 simulate:3 argument:4 relatively:1 tref:1 membrane:3 describes:1 postsynaptic:4 wi:1 biologically:2 aihara:2 den:1 invariant:3 neuroprose:1 previously:1 turn:1 count:1 mechanism:2 needed:4 operation:1 permit:1 simulating:1 gate:6 existence:1 assumes:1 include:1 siegelmann:2 build:1 establish:1 classical:2 already:1 coherently:1 spike:17 realized:1 concentration:1 aertsen:2 abovementioned:2 traditional:1 exhibit:1 distance:1 link:1 thank:1 simulated:2 mail:1 presynaptic:2 considers:2 trivial:1 length:5 relationship:5 statement:1 negative:1 rise:1 collective:1 upper:5 neuron:45 benchmark:1 finite:6 extended:4 precise:1 arbitrary:15 introduced:1 required:1 specified:4 fv:7 established:2 pattern:2 saturation:1 including:1 memory:4 power:13 suitable:4 buhmann:3 imply:1 realvalued:3 rtn:1 carried:2 acknowledgement:1 relative:1 interesting:3 degree:1 article:6 lo:1 excitatory:1 course:1 last:3 formal:3 allow:1 side:1 institute:2 sparse:4 van:3 judd:2 dimension:16 transition:1 adaptive:1 universitaet:1 investigating:1 handbook:1 ipsp:2 spatio:1 continuous:6 reasonably:1 gerstner:5 poly:4 linearly:3 nothing:1 en:2 hemmen:1 axon:4 precision:8 pv:4 schulten:3 explicit:1 comput:1 theorem:18 ofn:1 specific:5 substance:1 showing:1 pac:1 physiological:1 evidence:1 exists:3 essential:1 consist:2 vapnik:1 workshop:1 yin:8 interrelated:1 simply:1 binding:2 corresponds:1 acm:2 bioi:2 modulate:1 viewed:1 oscillator:1 content:2 change:1 wulfram:1 determined:1 infinite:2 timeinvariant:2 latter:1 phenomenon:3
7,164
927
Optimal Movement Primitives Terence D. Sanger Jet Propulsion Laboratory MS 303-310 4800 Oak Grove Drive Pasadena, CA 91109 (818) 354-9127 tds@ai.mit .edu Abstract The theory of Optimal Unsupervised Motor Learning shows how a network can discover a reduced-order controller for an unknown nonlinear system by representing only the most significant modes. Here, I extend the theory to apply to command sequences, so that the most significant components discovered by the network correspond to motion "primitives". Combinations of these primitives can be used to produce a wide variety of different movements. I demonstrate applications to human handwriting decomposition and synthesis, as well as to the analysis of electrophysiological experiments on movements resulting from stimulation of the frog spinal cord. 1 INTRODUCTION There is much debate within the neuroscience community concerning the internal representation of movement, and current neurophysiological investigations are aimed at uncovering these representations. In this paper, I propose a different approach that attempts to define the optimal internal representation in terms of "movement primitives" , and I compare this representation with the observed behavior. In this way, we can make strong predictions about internal signal processing. Deviations from the predictions can indicate biological constraints or alternative goals that cause the biological system to be suboptimal. The concept of a motion primitive is not as well defined as that of a sensory primitive 1024 Terence Sanger u p y z Figure 1: Unsupervised Motor Learning: The plant P takes inputs u and produces outputs y. The sensory map C produces intermediate variables z, which are mapped onto the correct command inputs by the motor network N. within the visual system, for example. There is no direct equivalent to the "receptive field" concept that has allowed interpretation of sensory recordings. In this paper, I will propose an internal model that involves both motor receptive fields and a set of movement primitives which are combined using a weighted sum to produce a large class of movements. In this way, a small number of well-designed primitives can generate the full range of desired behaviors. I have previously developed the concept of "optimal unsupervised motor learning" to investigate optimal internal representations for instantaneous motor commands. The optimal representations adaptively discover a reduced-order linearizing controller for an unknown nonlinear plant. The theorems give the optimal solution in general, and can be applied to special cases for which both linear and nonlinear adaptive algorithms exist (Sanger 1994b). In order to apply the theory to complete movements it needs to be extended slightly, since in general movements exist within an infinite-dimensional task space rather than a finite-dimensional control space. The goal is to derive a small number of primitives that optimally encode the full set of observed movements. Generation of the internal movement primitives then becomes a data-compression problem, and I will choose primitives that minimize the resultant mean-squared error. 2 OPTIMAL UNSUPERVISED MOTOR LEARNING Optimal Unsupervised Motor Learning is based on three principles: 1. Dimensionality Reduction 2. Accurate Reduced-order Control 3. Minimum Sensory error Consider the system shown in figure 1. At time t, the plant P takes motor inputs u and produces sensory outputs y. A sensory mapping C transforms the raw sensory data y to an intermediate representation z. A motor mapping takes desired values of z and computes the appropriate command u such that CPu = z. Note that the Optimal Movement Primitives 1025 loop in the figure is not a feedback-control loop, but is intended to indicate the flow of information. With this diagram in mind, we can write the three principles as: 1. dim[z] < dim[y] 2. GPNz=z 3. IIPNGy - yll is minimized We can prove the following theorems (Sanger 1994b): Theorem 1: For all G there exists an N such that G P N z = z. If G is linear and p- 1 is linear, then N is linear. Theorem 2: For any G, define an invertible map C such that GC-l =_1 on range[G]. Then liP NGy - yll is minimized when G is chosen such that Ily - G-1GII is minimized . If G and P are linear and the singular value decomposition of P is given by LT SR, then the optimal maps are G = Land N = RT S-I. For the discussion of movement, the linear case will be the most important since in the nonlinear case we can use unsupervised motor learning to perform dimensionality reduction and linearization of the plant at each time t. The movement problem then becomes an infinite-dimensional linear problem. Previously, I have developed two iterative algorithms for computing the singular value decomposition from input/output samples (Sanger 1994a). The algorithms are called the "Double Generalized Hebbian Algorithm" (DGHA) and the "Orthogonal Asymmetric Encoder" (OAE). DGHA is given by 8G 8N T 'Y(zyT - LT[zzT]G) 'Y(zu T - LT[zzT]N T ) while OAE is described by: 'Y(iyT - LT[ZiT]G) 'Y(Gy - LT[GGT]z)uT where LT[ ] is an operator that sets the above diagonal elements of its matrix argument to zero, y = Pu, z = Gy, z = NT u, and 'Y is a learning rate constant. Both algorithms cause G to converge to the matrix of left singular vectors of P, and N to converge to the matrix of right singular vectors of P (multiplied by a diagonal matrix for DGHA) . DGHA is used in the examples below. 3 MOVEMENT In order to extend the above discussion to allow adaptive discovery of movement primitives, we now consider the plant P to be a mapping from command sequences u(t) to sensory sequences y(t). We will assume that the plant has been feedback linearized (perhaps by unsupervised motor learning). We also assume that the sensory network G is constrained to be linear. In this case, the optimal motor network N will also be linear. The intermediate variables z will be represented by a vector. The sensory mapping consists of a set of sensory "receptive fields" gi(t) 1026 Terence Sanger A .. Motor Map I Sensory Map n1(t)zl +- Zl +- ---lEt): \ : nn(~)~ ~ ~n +- Figure 2: Extension of unsupervised motor learning to the case of trajectories . Plant input and output are time-sequences u(t) and y(t). The sensory and motor maps now consist of sensory primitives gi(t) and motor primitives ni(t). such that Zi = J gj(t)y(t)dt =< gily > and the motor mapping consists of a set of "motor primitives" ni(t) such that u(t) = L ni(t)zi i as in figure 2. If the plant is equal to the identity (complete feedback linearization), then gi(t) = ni(t). In this case, the optimal sensory-motor primitives are given by the eigenfunctions of the autocorrelation function of y(t) . If the autocorrelation is stationary, then the infinite-window eigenfunctions will be sinusoids. Note that the optimal primitives depend both on the plant P as well as the statistical distribution of outputs y(t). In practice, both u(t) and y(t) are sampled at discrete time-points {tic} over a finite time-window, so that the plant input and output is in actuality a long vector. Since the plant is linear, the optimal solution is given by the singular value decomposition, and either the DGHA or OAE algorithms can be used directly. The resulting sensory primitives map the sensory information y(t) onto the finite-dimensional z, which is usually a significant data compression. The motor primitives map Z onto the sequence u(t), and the resulting y(t) = P[u(t)] will be a linear projection of y(t) onto the space spanned by the set {Pni(t)}. 4 EXAMPLE 1: HANDWRITING As a simple illustration, I examine the case of human handwriting. We can consider the plant to be the identity mapping from pen position to pen position, and the Optimal Movement Primitives 1027 1. 5. 2. 6. 3. 7. 4. 8. Figure 3: Movement primitives for sampled human handwriting. 1028 Terence Sanger human to be taking desired sensory values of pen position and converting them into motor commands to move the pen. The sensory statistics then reflect the set of trajectories used in producing handwritten letters. An optimal reduced-order control system can be designed based on the observed statistics, and its performance can be compared to human performance. For this example, I chose sampled data from 87 different examples of lower-case letters written by a single person, and represented as horizontal and vertical pen position at each point in time. Blocks of 128 sequential points were used for training, and 8 internal variables Zi were used for each of the two components of pen position. The training set consisted of 5000 randomly chosen samples. Since the plant is the identity, the sensory and motor primitives are the same, and these are shown as "strokes" in figure 3. Linear combinations of these strokes can be used to generate pen paths for drawing lowercase letters. This is shown in figure 4, where the word "hello" (not present in the training set) is written and projected using increasing numbers of intermediate variables Zi. The bottom of figure 4 shows the sequence of values of Zi that was used (horizontal component only). Good reproduction of the test word was achieved with 5 movement primitives. A total of 7 128-point segments was projected , and these were recombined using smooth 50% overlap . Each segment was encoded by 5 coefficients for each of the horizontal and vertical components, giving a total of 70 coefficients to represent 1792 data points (896 horizontal and vertical components) , for a compression ratio of 25 :1. 5 EXAMPLE 2: FROG SPINAL CORD The second example models some interesting and unexplained neurophysiological results from microstimulation of the frog spinal cord. (Bizzi et al. 1991) measured the pattern of forces produced by the frog hindlimb at various positions in the workspace during stimulation of spinal interneurons. The resulting force-fields often have a stable" equilibrium point" , and in some cases this equilibrium point follows a smooth closed trajectory during tonic stimulation of the interneuron. However, only a small number of different force field shapes have been found, and an even smaller number of different trajectory types . A hypothesis to explain this result is that larger classes of different trajectories can be formed by combining the patterns produced by these cells. This hypothesis can be modelled using the optimal movement primitives described above . Figure 5a shows a simulation of the frog leg. To train the network, random smooth planar movements were made for 5000 time points. The plant output was considered to be 32 successive cartesian endpoint positions , and the plant input was the timevarying force vector field. Two hidden units Z were used . In figure 5b we see an example of the two equilibrium point trajectories (movement primitives) that were learned by DG HA. Linear combinations of these trajectories account for over 96% of the variance of the training data, and they can approximate a large class of smooth movements . Note that many other pairs of orthogonal trajectories can accomplish this, and different trials often produced different orthogonal trajectory shapes. 1029 Optimal Movement Primitives Original 5.~ 1. 2 3. 8.~~ Coefficients Figure 4: Projection of test-word "hello" using increasing numbers of intermediate variables Zi. 1030 Terence Sanger WorkSpace o. b. Figure 5: a. Simulation of frog leg configuration. b. An example of learned optimal movement primitives. 6 CONCLUSION The examples are not meant to provide detailed models of internal processing for human or frog motor control. Rather, they are intended to illustrate the concept of optimal primitives and perhaps guide the search for neurophysiological and psychophysical correlates of these primitives. The first example shows that generation of the lower-case alphabet can be accomplished with approximately 10 coefficients per letter, and that this covers a considerable range of variability in character production . The second example demonstrates that an adaptive algorithm allows the possibility for the frog spinal cord to control movement using a very small number of internal variables. Optimal unsupervised motor learning thus provides a descriptive model for the generation of a large class of movements using a compressed internal description. A set of fixed movement primitives can be combined linearly to produce the necessary motor commands, and the optimal choice of these primitives assures that the error in the resulting movement will be minimized. References Bizzi E., Mussa-Ivaldi F. A., Giszter S., 1991, Computations underlying the execution of movement: A biological perspective, Science, 253:287-29l. Sanger T. D., 1994a, Two algorithms for iterative computation of the singular value decomposition from input/output samples, In Touretzky D., ed., Advances in Neural Information Processing 6, Morgan Kaufmann, San Mateo, CA, in press. Sanger T. D., 1994b, Optimal unsupervised motor learning, IEEE Trans. Neural Networks, in press.
927 |@word trial:1 oae:3 compression:3 simulation:2 linearized:1 decomposition:5 reduction:2 ivaldi:1 configuration:1 current:1 nt:1 written:2 shape:2 motor:27 designed:2 stationary:1 provides:1 successive:1 oak:1 direct:1 prove:1 consists:2 autocorrelation:2 behavior:2 examine:1 cpu:1 window:2 increasing:2 becomes:2 discover:2 underlying:1 tic:1 developed:2 demonstrates:1 control:6 zl:2 unit:1 producing:1 path:1 approximately:1 chose:1 frog:8 mateo:1 range:3 practice:1 block:1 projection:2 word:3 onto:4 operator:1 equivalent:1 map:8 primitive:32 microstimulation:1 spanned:1 zzt:2 dgha:5 hypothesis:2 element:1 asymmetric:1 observed:3 bottom:1 cord:4 movement:30 depend:1 segment:2 represented:2 various:1 alphabet:1 train:1 encoded:1 larger:1 drawing:1 compressed:1 encoder:1 statistic:2 gi:3 sequence:6 descriptive:1 propose:2 loop:2 combining:1 description:1 double:1 produce:6 derive:1 illustrate:1 measured:1 strong:1 zit:1 involves:1 indicate:2 zyt:1 correct:1 human:6 investigation:1 recombined:1 biological:3 extension:1 considered:1 equilibrium:3 mapping:6 bizzi:2 unexplained:1 weighted:1 mit:1 hello:2 rather:2 command:7 timevarying:1 encode:1 ily:1 dim:2 lowercase:1 nn:1 pasadena:1 hidden:1 uncovering:1 constrained:1 special:1 field:6 equal:1 tds:1 unsupervised:10 minimized:4 randomly:1 dg:1 intended:2 mussa:1 n1:1 attempt:1 interneurons:1 investigate:1 possibility:1 accurate:1 grove:1 necessary:1 orthogonal:3 desired:3 cover:1 deviation:1 optimally:1 accomplish:1 combined:2 adaptively:1 person:1 workspace:2 terence:5 invertible:1 synthesis:1 squared:1 reflect:1 choose:1 account:1 gy:2 coefficient:4 closed:1 minimize:1 formed:1 ni:4 variance:1 kaufmann:1 correspond:1 modelled:1 raw:1 handwritten:1 produced:3 trajectory:9 drive:1 stroke:2 explain:1 touretzky:1 ed:1 resultant:1 handwriting:4 sampled:3 ut:1 dimensionality:2 electrophysiological:1 dt:1 planar:1 horizontal:4 nonlinear:4 mode:1 perhaps:2 concept:4 consisted:1 sinusoid:1 laboratory:1 during:2 m:1 linearizing:1 generalized:1 complete:2 demonstrate:1 motion:2 instantaneous:1 stimulation:3 spinal:5 endpoint:1 extend:2 interpretation:1 significant:3 ai:1 stable:1 gj:1 pu:1 perspective:1 accomplished:1 morgan:1 minimum:1 converting:1 converge:2 signal:1 full:2 hebbian:1 smooth:4 jet:1 long:1 ggt:1 concerning:1 prediction:2 controller:2 represent:1 achieved:1 cell:1 diagram:1 singular:6 sr:1 eigenfunctions:2 recording:1 gii:1 flow:1 intermediate:5 variety:1 zi:6 suboptimal:1 actuality:1 cause:2 detailed:1 aimed:1 transforms:1 hindlimb:1 reduced:4 generate:2 exist:2 neuroscience:1 per:1 write:1 discrete:1 sum:1 letter:4 constraint:1 argument:1 combination:3 smaller:1 slightly:1 character:1 leg:2 previously:2 assures:1 mind:1 multiplied:1 apply:2 appropriate:1 alternative:1 original:1 sanger:10 giving:1 psychophysical:1 move:1 receptive:3 rt:1 diagonal:2 mapped:1 propulsion:1 illustration:1 iyt:1 ratio:1 debate:1 unknown:2 perform:1 vertical:3 finite:3 extended:1 tonic:1 variability:1 discovered:1 gc:1 community:1 pair:1 yll:2 learned:2 trans:1 below:1 usually:1 pattern:2 overlap:1 force:4 representing:1 discovery:1 plant:15 generation:3 interesting:1 principle:2 land:1 production:1 guide:1 allow:1 wide:1 pni:1 taking:1 feedback:3 computes:1 sensory:20 made:1 adaptive:3 projected:2 san:1 correlate:1 approximate:1 search:1 iterative:2 pen:7 lip:1 ca:2 linearly:1 allowed:1 position:7 theorem:4 zu:1 reproduction:1 exists:1 consist:1 sequential:1 linearization:2 execution:1 cartesian:1 interneuron:1 lt:6 neurophysiological:3 visual:1 goal:2 identity:3 considerable:1 infinite:3 giszter:1 called:1 total:2 internal:10 meant:1
7,165
928
Reinforcement Learning Predicts the Site of Plasticity for Auditory Remapping in the Barn Owl Alexandre Pougett Cedric Deffayett Terrence J. Sejnowskit cedric@salk.edu terry@salk.edu alex@salk .edu tHoward Hughes Medical Institute The Salk Institute La Jolla, CA 92037 Department of Biology University of California, San Diego and tEcole Normale Superieure 45 rue d'Ulm 75005 Paris, France Abstract The auditory system of the barn owl contains several spatial maps . In young barn owls raised with optical prisms over their eyes, these auditory maps are shifted to stay in register with the visual map, suggesting that the visual input imposes a frame of reference on the auditory maps. However, the optic tectum, the first site of convergence of visual with auditory information, is not the site of plasticity for the shift of the auditory maps; the plasticity occurs instead in the inferior colliculus, which contains an auditory map and projects into the optic tectum . We explored a model of the owl remapping in which a global reinforcement signal whose delivery is controlled by visual foveation. A hebb learning rule gated by reinforcement learned to appropriately adjust auditory maps. In addition, reinforcement learning preferentially adjusted the weights in the inferior colliculus, as in the owl brain, even though the weights were allowed to change throughout the auditory system. This observation raises the possibility that the site of learning does not have to be genetically specified, but could be determined by how the learning procedure interacts with the network architecture . 126 Alexandre Pouget, Cedric Deffayet, Te"ence J. Sejnowski c,an c:::======:::::? ? ~_m Visual System Optic Tectum Inferior Colllc-ulus External nucleua .- t Forebrain Field L U~ a~ t Ovold.H. Nucleull ?"bala:m.ic Relay Inferior Colltculu. Cenlnll Nucleus C1ec) t Cochlea Figure 1: Schematic view of the auditory pathways in the barn owl. 1 Introduction The barn owl relies primarily on sounds to localize prey [6] with an accuracy vastly superior to that of humans. Figure 1A illustrates some of the nuclei involved in processing auditory signals. The barn owl determines the location of sound sources by comparing the time and amplitude differences of the sound wave between the two ears. These two cues are combined together for the first time in the shell and core of the inferior colliculus (ICc) which is shown at the bottom of the diagram . Cells in the ICc are frequency tuned and subject to spatial aliasing. This prevents them from unambiguously encoding the position of objects. The first unambiguous auditory map is found at the next stage, in the external capsule of the inferior colliculus (ICx) which itself projects to the optic tectum (OT). The OT is the first subforebrain structure which contains a multimodal spatial map in which cells typically have spatially congruent visual and auditory receptive fields. In addition, these subforebrain auditory pathways send one major collateral toward the forebrain via a thalamic relay. These collaterals originate in the ICc and are thought to convey the spatial location of objects to the forebrain [3]. Within the forebrain, two major structures have been involved in auditory processing: the archistriatum and field L. The archistriatum sends a projection to both the inferior colliculus and the optic tectum . Knudsen and Knudsen (1985) have shown that these auditory maps can adapt to systematic changes in the sensory input. Furthermore, the adaptation appears to be under the control of visual input, which imposes a frame of reference on the incoming auditory signals. In owls raised with optical prisms, which introduce a systematic shift in part of the visual field, the visual map in the optic tectum was identical to that found in control animals, but the auditory map in the ICx was shifted by the amount of visual shift introduced by the prisms. This plasticity ensures that the visual and auditory maps stay in spatial register during growth Reinforcement Learning Predicts the Site of Plasticity for Auditory Remapping 127 and other perturbations to sensory mismatch. Since vision instructs audition, one might expect the auditory map to shift in the optic tectum, the first site of visual-auditory convergence. Surprisingly, Brainard and Knudsen (1993b) observed that the synaptic changes took place between the ICc and the ICx, one synapse before the site of convergence. These observations raise two important questions: First, how does the animal knows how to adapt the weights in the ICx in the absence of a visual teaching signal? Second, why does the change take place at this particular location and not in the aT where a teaching signal would be readily available? In a previous model [7], this shift was simulated using backpropagation to broadcast the error back through the layers and by constraining the weights changes to the projection from the ICc to ICx. There is, however, no evidence for a feedback projection between from the aT to the ICx that could transmit the error signal; nor is there evidence to exclude plasticity at other synapses in these pathways. In this paper, we suggest an alternative approach in which vision guides the remapping of auditory maps by controlling the delivery of a scalar reinforcement signal. This learning proceeds by generating random actions and increasing the probability of actions that are consistently reinforced [1, 5] . In addition, we show that reinforcement learning correctly predicts the site of learning in the barn owl, namely at the ICx-ICc synapse, whereas backpropagation [8] does not favor this location when plasticity is allowed at every synapse. This raises a general issue: the site of synaptic adjustment might be imposed by the combination of the architecture and learning rule, without having to restrict plasticity to a particular synapse. 2 2.1 Methods Network Architecture The network architecture of the model based on the barn owl auditory system, shown in figure 2A, contains two parallel pathways. The input layer was an 8x21 map corresponding to the ICc in which units responded to frequency and interaural phase differences. These responses were pooled together to create auditory spatial maps at subsequent stages in both pathways. The rest of the network contained a series of similar auditory maps, which were connected topographically by receptive fields 13 units wide. We did not distinguish between field L and the archistriatum in the forebrain pathways and simply used two auditory maps, both called FBr. We used multiplicative (sigma-pi) units in the aT whose activities were determined according to: Yi = L,. w~Br yfBr WfkBr yfc:c (1) j The multiplicative interaction between ICx and FBr activities was an important assumption of our model. It forced the ICx and FBr to agree on a particular position before the aT was activated. As a result, if the ICx-aT synapses were modified during learning, the ICx-FBr synapses had to be changed accordingly. 128 Alexandre Pouget, Cedric Deffayet, Terrence J. Sejnowski Figure 2: Schematic diagram of weights (white blocks) in the barn owl auditory system. A) Diagram of the initial weights in the network. B) Pattern of weights after training with reinforcement learning on a prism-induced shift offour units. The remapping took place within the ICx and FBr. C) Pattern of weights after training with backpropagation. This time the ICx-OT and FBr-OT weights changed. Weights were clipped between 5.0 and 0.01, except for the FBr-ICx connections whose values were allowed to vary between 8.0 and 0.01. The minimum values were set to 0.01 instead of zero to prevent getting trapped in unstable local minima which are often associated with weights values of zero. The strong coupling between FBr and ICx was another important assumption of the model whose consequence will be discussed in the last section. Examples were generated by simply activating one unit in the ICc while keeping the others to zero, thereby simulating the pattern of activity that would be triggered by a single localized auditory stimulus. In all simulations, we modeled a prism-induced shift of four units. 2.2 Reinforcement learning We used stochastic units and trained the network using reinforcement learning [1]. The weighted sum of the inputs, neti, passed through a sigmoid, f(x) , is interpreted as the probability, Pi, that the unit will be active: Pi = f(neti) * 0.99 + 0.01 (2) were the output of the unit Yi was: ._{a y, - with probability 1 - Pi 1 with probability Pi (3) Reinforcement Learning Predicts the Site of Plasticity for Auditory Remapping 129 Because of the form of the equation for Pi, all units in the network had a small probability (0.01) of being spontaneously active in the absence of any inputs. This is what allowed the network to perform a stochastic search in action space to find which actions were consistently associated with positive reinforcement. We ensured that at most one unit was active per trial by using a winner-take-all competition in each layer. Adjustable weights in the network were updated after each training examples with hebb-like rule gated by reinforcement: (4) A trial consisted in choosing a random target location for auditory input (ICc) and the output of the OT was used to generate a head movement . The reinforcement , r , was then set to 1 for head movements resulting in the foveation of the stimulus and to -0.05 otherwise. 2.3 Backpropagation For the backpropagation network , we used deterministic units with sigmoid activation functions in which the output of a unit was given by: (5) where neti is the weighted sum of the inputs as before. The chain rule was used to compute the partial derivatives of the squared error, E , with respect to each weights and the weights were updated after each training example according to: (6) The target vectors were similar to the input vectors, namely only one OT units was required to be activated for a given pattern, but at a position displaced by 4 units compared to the input. 3 3.1 Results Learning site with reinforcement In a first set of simulation we kept the ICc-ICx and ICc-FBr weights fixed. Plasticity was allowed at these site in later simulations. Figure 2A shows the initial set of weights before learning starts. The central diagonal lines in the weight diagrams illustrate the fact that each unit receives only one non-zero weight from the unit in the layer below at the same location. 130 Alexandre Pouget, Cedric Deffayet, Terrence J. Sejnowski There are two solutions to the remapping: either the weights change within the ICx and FBr, or from the ICx and the FBr to the ~T. As shown in figure 2B , reinforcement learning converged to the first solution. In contrast, the weights between the other layers were unaltered, even though they were allowed to change. To prove that the network could have actually learned the second solution, we trained a network in which the ICc-ICx weights were kept fixed . As we expected, the network shifted its maps simultaneously in both sets of weights converging onto the OT, and the resulting weights were similar to the ones illustrated in figure 2C. However, to reach this solution , three times as many training examples were needed. The reason why learning in the ICx and FBr were favored can be attributed to probabilistic nature of reinforcement learning. If the probability of finding one solution is p, the probability of finding it twice independently is p2. Learning in the ICx and FBR is not independent because of the strong connection from the FBr to the ICx. When the remapping is learned in the FBR this connection automatically remapped the activities in the ICx which in turn allows the ICx-ICx weights to remap appropriately. In the OT on the other hand, the multiplicative connection between the ICx and FBr weights prevent a cooperation between this two sets of weights. Consequently, they have to change independently, a process which took much more training. 3.2 Learning at the ICc-ICx and ICc-FBr synapses The aliasing and sharp frequency tuning in the response of ICc neurons greatly slows down learning at the ICc-ICx and ICc-FBr synapses. We found that when these synapses were free to change, the remapping still took place within the ICx or FBr (figure 3). 3.3 Learning site with backpropagation In contrast to reinforcement learning, backpropagation adjusted the weights in two locations: between the ICx and the OT and between the Fbr and OT (figure 2C). This is the consequence of the tendency of the backpropagation algorithm to first change the weights closest to where the error is injected. 3.4 Temporal evolution of weights Whether we used reinforcement or supervised learning, the map shifted in a very similar way. There was a simultaneous decrease of the original set of weights with a simultaneous increase of the new weights, such that both sets of weights coexisted half way through learning. This indicates that the map shifted directly from the original setting to the new configuration without going through intermediate shifts. This temporal evolution of the weights is consistent the findings of Brainard and Knudsen (1993a) who found that during the intermediate phase of the remapping, cells in the inferior colli cuI us typically have two receptive fields. More recent work however indicates that for some cells the remapping is more continuous(Brainard and Knudsen , personal communication) , a behavior that was not reproduced by either of the learning rule. Reinforcement Learning Predicts the Site of Plasticity for Auditory Remapping 131 Figure 3: Even when the ICc-ICx weights are free to change, the network update the weights in the ICx first. A separate weight matrix is shown for each isofrequency map from the ICc to ICx. The final weight matrices were predominantly diagonal; in contrast, the weight matrix in ICx was shifted. 4 Discussion Our simulations suggest a biologically plausible mechanism by which vision can guide the remapping of auditory spatial maps in the owl's brain. Unlike previous approaches, which relied on visual signals as an explicit teacher in the optic tectum [7], our model uses a global reinforcement signal whose delivery is controlled by the foveal representation of the visual system. Other global reinforcement signals would work as well. For example, a part of the forebrain might compare auditory and visual patterns and report spatial mismatch between the two. This signal could be easily incorporated in our network and would also remap the auditory map in the inferior colli cuI us. Our model demonstrates that the site of synaptic plasticity can be constrained by the interaction between reinforcement learning and the network architecture. Reinforcement learning converged to the most probably solution through stochastic search. In the network, the strong lateral coupling between ICx and FBr and the multiplicative interaction in the OT favored a solution in which the remapping took place simultaneously in the ICx and FBr. A similar mechanism may be at work in the barn owl's brain. Colaterals from FBr to ICx are known to exist, but the multiplicative interaction has not been reported in the barn owl optic tectum. Learning mechanisms may also limit synaptic plasticity. NMDA receptors have been reported in the ICx, but they might not be expressed at other synapses. There may, however, be other mechanisms for plasticity. The site of remapping in our model was somewhat different from the existing observations. We found that the change took place within the ICx whereas Brainard and Knudsen [3] report that it is between the ICc and the ICx. A close examination of their data (figure 11 in [3]) reveals that cells at the bottom of ICx were not 132 Alexandre Pouget, Cedric Deffayet, Terrence J. Sejnowski remapped, as predicted by our model, but at the same time, there is little anatomical or physiological evidence for a functional and hierarchical organization within the ICx. Additional recordings are need to resolve this issue. We conclude that for the barn owl's brain, as well as for our model, synaptic plasticity within ICx was favored over changes between ICc and ICx. This supports the hypothesis that reinforcement learning is used for remapping in the barn owl auditory system. Acknowledgments We thank Eric Knudsen and Michael Brainard for helpful discussions on plasticity in the barn owl auditory system and the results of unpublished experiments. Peter Dayan and P. Read Montague helped with useful insights on the biological basis of reinforcement learning in the early stages of this project. References [1] A.G. Barto and M.1. Jordan. Gradient following without backpropagation in layered networks. Proc. IEEE Int. Conf. Neural Networks, 2:629-636, 1987. [2] M.S. Brainard and E.1. Knudsen. Dynamics of the visual calibration of the map of interaural time difference in the barn owl's optic tectum. In Society For Neuroscience Abstracts, volume 19, page 369.8, 1993. [3] M.S. Brainard and E.!. Knudsen. Experience-dependent plasticity in the inferior colliculus: a site for visual calibration of the neural representation of auditory space in the barn owl. The journal of Neuroscience, 13:4589-4608, 1993. [4] E. Knudsen and P. Knudsen. Vision guides the adjustment of auditory localization in the young barn owls. Science, 230:545-548, 1985. [5] P.R. Montague, P. Dayan, S.J. Nowlan, A. Pouget, and T.J. Sejnowski. Using aperiodic reinforcement for directed self-organization during development. In S.J. Hanson, J.D. Cowan, and C.L. Giles, editors, Advances in Neural Information Processing Systems, volume 5. Morgan-Kaufmann, San Mateo, CA, 1993. [6] R.S. Payne. Acoustic location of prey by barn owls (tyto alba). Journal of Experimental Biology, 54:535-573, 1970. [7] D.J. Rosen, D.E. Rumelhart, and E.I. Knudsen. A connectionist model of the owl's sound localization system. In Advances in Neural Information Processing Systems, volume 6. Morgan-Kaufmann, San Mateo, CA, 1994. [8] D.E. Rumelhart, G.E. Hinton, and R.J . Williams. Learning internal representations by error propagation. In D. E. Rumelhart, J. L. McClelland, and the PDP Research Group, editors, Parallel Distributed Processing, volume 1, chapter 8, pages 318-362. MIT Press, Cambridge, MA, 1986.
928 |@word trial:2 unaltered:1 simulation:4 thereby:1 ulus:1 initial:2 configuration:1 contains:4 series:1 foveal:1 tuned:1 existing:1 comparing:1 nowlan:1 activation:1 readily:1 subsequent:1 plasticity:17 update:1 cue:1 half:1 accordingly:1 core:1 location:8 instructs:1 prove:1 pathway:6 interaural:2 introduce:1 expected:1 behavior:1 nor:1 aliasing:2 brain:4 automatically:1 resolve:1 little:1 increasing:1 project:3 remapping:16 what:1 tyto:1 interpreted:1 yfc:1 offour:1 finding:3 temporal:2 every:1 growth:1 ensured:1 demonstrates:1 control:2 unit:17 medical:1 before:4 positive:1 local:1 limit:1 consequence:2 receptor:1 encoding:1 might:4 twice:1 mateo:2 directed:1 acknowledgment:1 spontaneously:1 hughes:1 block:1 backpropagation:9 procedure:1 coexisted:1 thought:1 projection:3 suggest:2 onto:1 close:1 layered:1 map:26 deterministic:1 imposed:1 send:1 williams:1 independently:2 pouget:5 rule:5 insight:1 sejnowskit:1 transmit:1 updated:2 diego:1 tectum:10 controlling:1 ulm:1 target:2 us:1 hypothesis:1 rumelhart:3 predicts:5 bottom:2 observed:1 ensures:1 connected:1 movement:2 decrease:1 colli:2 dynamic:1 personal:1 trained:2 raise:3 topographically:1 localization:2 eric:1 basis:1 multimodal:1 easily:1 montague:2 chapter:1 forced:1 sejnowski:5 choosing:1 whose:5 plausible:1 otherwise:1 favor:1 itself:1 final:1 reproduced:1 triggered:1 took:6 interaction:4 adaptation:1 payne:1 competition:1 getting:1 convergence:3 congruent:1 generating:1 object:2 brainard:7 coupling:2 illustrate:1 p2:1 strong:3 predicted:1 aperiodic:1 stochastic:3 human:1 owl:23 activating:1 biological:1 adjusted:2 barn:18 ic:1 major:2 vary:1 early:1 relay:2 proc:1 create:1 weighted:2 mit:1 modified:1 normale:1 barto:1 consistently:2 indicates:2 greatly:1 contrast:3 helpful:1 dayan:2 dependent:1 typically:2 going:1 france:1 issue:2 favored:3 development:1 animal:2 spatial:8 raised:2 constrained:1 field:7 having:1 biology:2 identical:1 rosen:1 others:1 stimulus:2 report:2 connectionist:1 primarily:1 simultaneously:2 phase:2 organization:2 possibility:1 adjust:1 activated:2 chain:1 partial:1 experience:1 collateral:2 giles:1 ence:1 reported:2 teacher:1 combined:1 stay:2 systematic:2 terrence:4 probabilistic:1 michael:1 together:2 vastly:1 squared:1 ear:1 central:1 broadcast:1 external:2 conf:1 audition:1 derivative:1 suggesting:1 exclude:1 pooled:1 int:1 alba:1 register:2 multiplicative:5 view:1 later:1 helped:1 wave:1 start:1 thalamic:1 parallel:2 relied:1 accuracy:1 responded:1 kaufmann:2 who:1 reinforced:1 archistriatum:3 converged:2 simultaneous:2 synapsis:7 reach:1 synaptic:5 isofrequency:1 frequency:3 involved:2 associated:2 attributed:1 auditory:39 nmda:1 amplitude:1 actually:1 back:1 appears:1 alexandre:5 supervised:1 unambiguously:1 response:2 synapse:4 though:2 furthermore:1 stage:3 hand:1 receives:1 propagation:1 consisted:1 evolution:2 spatially:1 read:1 illustrated:1 white:1 during:4 self:1 inferior:10 unambiguous:1 predominantly:1 superior:1 sigmoid:2 functional:1 winner:1 volume:4 discussed:1 cambridge:1 tuning:1 teaching:2 had:2 calibration:2 closest:1 recent:1 jolla:1 prism:5 yi:2 morgan:2 minimum:2 additional:1 somewhat:1 signal:11 sound:4 adapt:2 controlled:2 schematic:2 converging:1 vision:4 cochlea:1 cell:5 addition:3 whereas:2 diagram:4 source:1 sends:1 appropriately:2 ot:11 rest:1 unlike:1 probably:1 subject:1 induced:2 recording:1 cowan:1 jordan:1 constraining:1 intermediate:2 architecture:5 restrict:1 br:1 shift:8 whether:1 passed:1 remap:2 peter:1 action:4 useful:1 amount:1 mcclelland:1 generate:1 exist:1 shifted:6 trapped:1 neuroscience:2 correctly:1 per:1 anatomical:1 group:1 four:1 localize:1 prevent:2 prey:2 kept:2 sum:2 colliculus:6 injected:1 place:6 throughout:1 clipped:1 delivery:3 layer:5 distinguish:1 bala:1 activity:4 optic:10 alex:1 optical:2 department:1 according:2 combination:1 cui:2 biologically:1 equation:1 agree:1 forebrain:6 turn:1 mechanism:4 neti:3 needed:1 know:1 available:1 hierarchical:1 simulating:1 alternative:1 original:2 x21:1 society:1 question:1 occurs:1 receptive:3 interacts:1 diagonal:2 subforebrain:2 gradient:1 separate:1 thank:1 simulated:1 lateral:1 originate:1 unstable:1 toward:1 reason:1 modeled:1 preferentially:1 sigma:1 slows:1 adjustable:1 gated:2 perform:1 observation:3 displaced:1 neuron:1 knudsen:12 hinton:1 communication:1 head:2 incorporated:1 frame:2 pdp:1 perturbation:1 sharp:1 introduced:1 namely:2 paris:1 specified:1 required:1 connection:4 unpublished:1 hanson:1 california:1 acoustic:1 learned:3 icx:44 proceeds:1 below:1 pattern:5 mismatch:2 remapped:2 genetically:1 deffayet:4 terry:1 examination:1 eye:1 icc:21 cedric:6 expect:1 localized:1 nucleus:2 consistent:1 imposes:2 editor:2 pi:6 changed:2 cooperation:1 surprisingly:1 last:1 keeping:1 free:2 guide:3 institute:2 wide:1 distributed:1 feedback:1 sensory:2 reinforcement:27 san:3 global:3 active:3 incoming:1 reveals:1 conclude:1 search:2 continuous:1 why:2 nature:1 capsule:1 ca:3 rue:1 did:1 allowed:6 convey:1 site:17 hebb:2 salk:4 position:3 explicit:1 young:2 down:1 explored:1 physiological:1 evidence:3 fbr:23 te:1 illustrates:1 simply:2 visual:18 prevents:1 expressed:1 adjustment:2 contained:1 scalar:1 determines:1 relies:1 ma:1 shell:1 consequently:1 absence:2 change:13 determined:2 foveation:2 except:1 called:1 tendency:1 la:1 experimental:1 internal:1 support:1 superieure:1
7,166
929
Hierarchical Mixtures of Experts Methodology Applied to Continuous Speech Recognition Ying Zhao, Richard Schwartz, Jason Sroka*: John Makhoul BBN System and Technologies 70 Fawcett Street Cambridge MA 02138 Abstract In this paper, we incorporate the Hierarchical Mixtures of Experts (HME) method of probability estimation, developed by Jordan [1], into an HMMbased continuous speech recognition system. The resulting system can be thought of as a continuous-density HMM system, but instead of using gaussian mixtures, the HME system employs a large set of hierarchically organized but relatively small neural networks to perform the probability density estimation. The hierarchical structure is reminiscent of a decision tree except for two important differences: each "expert" or neural net performs a "soft" decision rather than a hard decision, and, unlike ordinary decision trees, the parameters of all the neural nets in the HME are automatically trainable using the EM algorithm. We report results on the ARPA 5,OOO-word and 4O,OOO-word Wall Street Journal corpus using HME models. 1 Introduction Recent research has shown that a continuous-density HMM (CD-HMM) system can outperform a more constrained tied-mixture HMM system for large-vocabulary continuous speech recognition (CSR) when a large amount of training data is available [2]. In other work, the utility of decision trees has been demonstrated in classification problems by using the "divide and conquer" paradigm effectively, where a problem is divided into a hierarchical set of simpler problems. We present here a new CD-HMM system which **MIT, Cambridge MA 02139 860 Ying Zhao, Richard Schwartz, Jason Sroka, John Makhoul has similar properties and possesses the same advantages as decision trees, but has the additional important advantage of having automatically trainable "soft" decision boundaries. 2 Hierarchical Mixtures of Experts The method of Hierarchical Mixtures of Experts (HME) developed recently by Jordan [1] breaks a large scale task into many small ones by partitioning the input space into a nested set of regions, then building a simple but specific model (local expert) in each region. The idea behind this method follows the principle of divide-and-conquer which has been utilized in certain approaches to classification problems, such as decision trees. In the decision tree approach, at each level of the tree, the data are divided explicitly into regions. In contrast, the HME model makes use of "soft" splits of the data, i.e., instead of the data being explicitly divided into regions, the data may lie simultaneously in multiple regions with certain probabilities. Therefore, the variance-increasing effect of lopping off distant data in the decision tree can be ameliorated. Furthermore, the "hard" boundaries in the decision tree are fixed once a decision is made, while the "soft" boundaries in the HME are parameterized with generalized sigmoidal functions, which can be adjusted automatically using the Expectation-Maximization (EM) algorithm during the splitting. Now we describe how to apply the HME methodology to the CSR problem. For each state of a phonetic HMM, a separate HME is used to estimate the likelihood. The actual HME first computes a posterior probability P(llz, s), the probability of phoneme class I, given the input feature vector z and state s. That probability is then divided by the a priori probability of the phone class I at state s. A one-level HME performs the following computation: c P(llz, s) =L P(llci, z, s)P(cilz, s) (1) i=l where I = 1, ,.. , L indicates phoneme class, Ci represents a local region in the input space, and C is the number of regions. P(cilz, s) can be viewed as a gating network, while P(lICi, z, s) can be viewed as a local expert classifier (expert network) in the region [1]. In a two-level HME, each region Ci is divided in turn into C subregions. The term P(IICi, z, s) is then computed in a similar manner to equation (1), and so on. If in some of these subregions there are no data available, we back off to the parent network. c, 3 TECHNICAL DETAILS As in Jordan's paper, we use a generalized sigmoidal function to parameterize P(cilz) as follows: (2) where z can be the direct input (in a one-layer neural net) or the hidden layer vector (in a two-layer neural net), and v,, i = 1, .. " C are weights which need to train. Similarly, the local phoneme classifier in region Ci, P(llc" z), can be parameterized with a generalized Mixtures of Experts Applied to Continuous Speech Recognition 861 sigmoidal function also: (3) where 8;i,j = 1, ... , L are weights. The whole system consists of two set of parameters: Vi, i 1, ... , C and 8;i' j 1, ... , L, = {8;i' Vi}. All parameters are estimated by using the EM algorithm. = e = The EM is an iterative approach to maximum likelihood estimation. Each iteration of an EM algorithm is composed of two steps: an Expectation (E) step and a Maximization (M) step. The M step involves the maximization of a likelihood function that is redefined in each iteration by the E step. Using the parameterizations in (2) and (3), we obtain the following iterative procedure for computing parameters = {Vi, 8;i}: . .. I'lze Vi(0) an d 8(0) 1. lDltIa ;i f or 1.. = I , ... , C ,}' = 1, ... , L . 2. E-step: In each iteration n, for each data pair (z(t), l(t?, t = 1, ... ,N, compute e zi(tin ) = P(cilz(t), l(t), e(n~ = P(Ci Iz(t), v~n?p(l(t)lci' z(t), 8~~~,i) (4) where i = 1, ... , C. Zi(t)<n) represents the probability of the data t lying in the region i, given the current parameter estimation e(n) . It will be used as a weight for this data in the region i in the M-step. The idea of "soft" splitting reflects that these weights are probabilities between 0 and 1, instead of a "hard"decision 0 or 1. 3. M-step: (5) (6) 4. Iterate until 8;i' Vi converge. The first maximization means fitting a generalized sigmoidal model (3) using the labeled data (z(t), l(t? and weighting Zi(t)<n). The second one means fitting a generalized sigmoidal model (2) using inputs z(t) and outputs Zi(t)<n). The criterion for fitting is the cross-entropy. Typically, the fitting can be solved by the Newton-Raphson method. However, it is quite expensive. Viewing this type of fitting as a multi-class classification task, we developed a technique to invert a generalized sigmoidal function more efficiently, which will be described in the following. A common method in a multi-class classification is to divide the problem into many 2-c1ass classifications. However, this method results in a positive and negative training unbalance usually. To avoid the positive and negative training unbalance, the following technique can be used to solve multi-class posterior probabilities simultaneously. Suppose we have a labeled data set, (z(t), l(t?, t = 1, ... , N, where l(t) E {I, ... , L} is the label for t-th data. We use a generalized sigmoidal function to model the posterior 862 Ying Zhao, Richard Schwartz, Jason Sroka, John Makhoul probability P(llz), where 1 = I, ... , L as follows: e9'f'z P,(z) = P(llz) = (7) 9T L:k e "z Obviously, since these probabilities sum up to one, we have L-I PL(Z) L P,(z). =1 - (8) '=1 Now, a training sample z(t) with a class label let) can be interpreted as: P,(z(t? = { 1 = let) 1 =/l(t) 0.9 1:':1 If we define (9) P,(z) T 9, z (10) = log PL(z) equation (10) implies that (11) for 1 = I, ... , L with 9Lz = O. This expression is the generalized sigmoidal function in (7). This means, we can train parameters in (7) to satisfy Equation (10) from the data. Using a least squares criterion, the objective is 9T z(t) - . ,,[ mm L..J t P'(Z(t?] 2 PL(Z(t? (12) log - - - for 1 = I, ... , L - 1. Denote a data matrix as z(l) z(2) x= zeN) A least squares solution to (12) is 9, = (loga)(XT X)-I [L L z(t) - '(t)=l Z(t)] (13) '(t)=L for 1 = I, ... , L, where a =9(L - 1). Substituting (13) into (11), we get a P,(z) = ZT(X T X)-l ~ L.JI(I~I L:k azT(XT X)-l L: z(t) z(t) (14) 1(1)=" Equation (13) and (14) are very easy to compute. Basically, we only have to accumulate the matrix XT X and sum z(t) into different classes 1 = I, ... , L. We can obtain probabilities P,(z) by a single inversion of matrix XT X after a pass through the training data. Mixtures of Experts Applied to Continuous Speech Recognition 863 4 Relation to Other Work The work reported here is very different from our previous work utilizing neural nets for CSR. There, a single segmental neural network (SNN) is used to model a complete phonetic segment [3]. Here, each HME estimates the probability density for each state of a phonetic HMM. The work here is more similar to that by Cohen et al. [4], the major difference being that in [4], a single very large neural net is used to perform the probability density modeling. The training of such a large network requires the use of a specialized parallel processing machine, so that the training can be done in a reasonable amount of time. By using the HME method and dividing the problem into many smaller problems, we are able to perform the needed training computation on regular workstations. Most of the previous work on CD-HMM work has utilized mixtures of gaussians to estimate the probability densities of an HMM. Since a ' multilayer feedforward neural network is a universal continuous function approximator, we decided to explore the use of neural nets as an alternative approach for continuous density estimation. 5 Experimental Results HMM SNN HMM+SNN HME HME+HMM Prior-modified HME + HMM Word Error Rate 7.8 8.5 7.1 7.6 6.8 6.2 Table 1: Error Rates for the ARPA WSJ 5K Development Test, Trigram Grammar HMM HME+ HMM Word Error Rate 9.5 8.7 Table 2: Error Rates for the ARPA WSJ 40K Test Set, Trigram Grammar In our initial application of the HME method to large-vocabulary CSR, we used phonetic context-independent HMEs to estimate the likelihoods at each state of 5-state HMMs. We implemented a two-level HME, with the input space divided into 46 regions, and each of those regions is further divided into 46 subregions. The initial divisions were accomplished by supervised training, with each division trained to one of the 46 phonemes in the system. All gating and local expert networks in the HME had identical structures - a two-layer generalized sigmoidal network. The whole HME system was implemented within an N-best paradigm [3], where the recognized sequence was obtained as a result of a rescoring of an N-best list obtained from our baseline BYBLOS system (tied-mixture HMM) with a statistical trigram grammar. 864 Ying Zhao, Richard Schwartz, Jason Sroka, John Makhoul We then built a context-dependent HME system based on the structure of the contextindependent HME models described above. For each state, the whole training data was divided into 46 parts according to its left or right context. Then for each context, a separate HME model was built for that context. To be computationally feasible, we used only one-level HMEs here. We first experimented using a left-context and right-context model. We tested the HME implementation on the ARPA 5,OOO-word Wall Street Journal corpus (WSJl, H2 dev set). We report the word error rates on the same test set for a number of different systems. Table 1 shows the word error rates for i) the baseline HMM system; ii) the segment-based neural net system (SNN) iii) the hybrid SNNIHMM system iv) a HME system alone. v) a HME system combined with HMM; vi) a HME +HMM system with modified priors. From Table 1, The performance of the baseline tied-mixture HMM is 7.8%. The performance of the SNN system (8.5%) is comparable to the HMM alone. We see that the performance of a HME (7.6%) is as good as the HMM system, which is better than the SNN system. When combined with the baseline HMM system, the HME and SNN both improve performance over the HMM alone about 10% from 7.8% to 6.8% and from 7.8% to 7.1% respectively. We found out that the improvement could be made larger for a hybrid HMElHMM by adjusting the context-dependent priors with the context-independent priors, and then smooth context-dependent models with a context-independent model. More specifically, in a context-dependent HME model, we usually estimate the posterior probability phoneme I, P(llc, z, s), given left or right context c and the acoustic input z in a particular state s. Because the samples may be sparse for many of context models, it is necessary to regularize (smooth) context-dependent models with a contextindependent model, where there is much more data available. However, since the two models have different priors: P(llc, s) in a context-dependent model and P(lls) in a context-independent model, a simple interpolation between the two models which is c,.)P(' c.) . d d od I d P(ll z, s ) = P(xP(x , .)P(' .) P(ll c, z, s ) = P(x ,P(x c,.) 10 a context- epen ent mean .) in a context-independent model is inconsistent. To scale the context-dependent priors P(llc, s) with a context-independent prior P(lls), we weighted each input data point z with the weight for a prior adjusting. After this modification, a context-dependent :c.i'c:;) HME actually estimates P(z ~~:~('I'). It combines better with a context-independent model. For the same experiment we showed in Table 1, the word error for the HME (with HMM) droped from 6.8% to 6.2% when priors were modified. For this 5,OOO-word development set, we got a total of about 20% word error reduction over the tied-mixture HMM system using a HME-based neural network system. We then switched our experiment domain from a 5,OOO-word to 40,OOO-word the test set. During this year, the BYBLOS system has been improVed from a tied-mixture system to a continuous density system. We also switched to using this new continuous density BYBLOS in our hybrid HMElHMM system. The language model used here was a 40,OOO-word trigram grammar. The result is shown in Table 2. From Table 2, we see that there is about a 10% word error rate reduction over the continuous density HMM system by combining a context-dependent HME system. Compared with the 20% improvement over the tied-mixture system we made for the 5,OOO-word development set, the improvement over the continuous density system in this 40,OOO-word Mixtures of Experts Applied to Continuous Speech Recognition 865 development is less. This may be due to the big improvement of the HMM system itself. 6 CONCLUSIONS The method of hierarchical mixtures of experts can be used as a continous density estimator to speech recognition. Experimental results showed that estimations from this approach are consistent with the estimations from the HMM system. The frame-based neural net system using hierarchical mixtures of experts improves the performance of both the state-of-the-art tied mixture HMM system and the continuous density HMM system. The HME system itself has the same performance as the state-of-the-art tied mixture HME system. 7 Acknowledgments This work was funded by the Advanced Research Projects Agency of the Department of Defense. References [1] Michael Jordan, "Hierarchical Mixtures of Experts and the EM Algorithm," Neural Computation, 1994, in press. [2] D. Pallett, J. Fiscus, W. Fisher, J. Garofolo, B. Lund, and M. Pryzbocki, "1993 Benchmark Tests for the ARPA Spoken Language Program," Proc. ARPA Human Language Technology Workshop, Plainsboro, NJ, Morgan Kaufman Publishers, 1994. [3] G. Zavaliagkos, Y. Zhao, R. Schwartz and J. Makhoul, "A Hybrid Neural Net System for State-of-the-Art Continuous Speech Recognition," in Advances in Neural Information Processing Systems 5, S. J. Hanson, J. D. Cowan and C. L. Giles, eds., Morgan Kaufmann Publishers, 1993. [4] M. Cohen, H. Franco, N. Morgan, D. Rumelhart and V. Abrash, "Context-Dependent Multiple Distribution Phonetic Modeling with MLPS," in Advances in Neural Information Processing Systems 5, S. J. Hanson, 1. D. Cowan and C. L. Giles, eds., Morgan Kaufmann Publishers, 1993.
929 |@word inversion:1 reduction:2 c1ass:1 initial:2 current:1 od:1 reminiscent:1 john:4 distant:1 alone:3 rescoring:1 parameterizations:1 sigmoidal:9 simpler:1 direct:1 consists:1 fitting:5 combine:1 manner:1 multi:3 automatically:3 actual:1 snn:7 increasing:1 project:1 kaufman:1 interpreted:1 developed:3 spoken:1 nj:1 classifier:2 schwartz:5 partitioning:1 positive:2 local:5 interpolation:1 garofolo:1 hmms:1 decided:1 acknowledgment:1 procedure:1 universal:1 thought:1 got:1 word:16 regular:1 get:1 context:25 demonstrated:1 splitting:2 estimator:1 utilizing:1 regularize:1 suppose:1 rumelhart:1 recognition:8 expensive:1 utilized:2 labeled:2 solved:1 parameterize:1 region:14 fiscus:1 agency:1 trained:1 segment:2 division:2 train:2 describe:1 quite:1 larger:1 solve:1 grammar:4 azt:1 itself:2 obviously:1 advantage:2 sequence:1 net:10 combining:1 ent:1 parent:1 wsj:2 dividing:1 implemented:2 involves:1 implies:1 human:1 viewing:1 wall:2 adjusted:1 pl:3 mm:1 lying:1 substituting:1 major:1 trigram:4 estimation:7 proc:1 label:2 reflects:1 weighted:1 iici:1 mit:1 gaussian:1 modified:3 rather:1 avoid:1 improvement:4 likelihood:4 indicates:1 contrast:1 baseline:4 dependent:10 typically:1 hmes:2 hidden:1 relation:1 classification:5 priori:1 development:4 constrained:1 art:3 once:1 having:1 identical:1 represents:2 report:2 richard:4 employ:1 lls:2 composed:1 simultaneously:2 mixture:20 behind:1 necessary:1 tree:9 iv:1 divide:3 arpa:6 soft:5 modeling:2 giles:2 dev:1 maximization:4 ordinary:1 reported:1 combined:2 density:13 off:2 michael:1 zen:1 e9:1 expert:15 zhao:5 hme:38 satisfy:1 explicitly:2 vi:6 break:1 jason:4 parallel:1 mlps:1 square:2 variance:1 phoneme:5 efficiently:1 kaufmann:2 basically:1 lci:1 ed:2 workstation:1 adjusting:2 improves:1 organized:1 actually:1 back:1 supervised:1 methodology:2 improved:1 ooo:9 done:1 furthermore:1 until:1 building:1 effect:1 ll:2 during:2 abrash:1 criterion:2 generalized:9 complete:1 performs:2 recently:1 common:1 specialized:1 ji:1 cohen:2 accumulate:1 cambridge:2 similarly:1 language:3 had:1 sroka:4 funded:1 segmental:1 posterior:4 recent:1 showed:2 loga:1 phone:1 phonetic:5 certain:2 accomplished:1 morgan:4 additional:1 recognized:1 converge:1 paradigm:2 ii:1 multiple:2 smooth:2 technical:1 cross:1 raphson:1 divided:8 multilayer:1 expectation:2 iteration:3 fawcett:1 invert:1 publisher:3 unlike:1 posse:1 cowan:2 inconsistent:1 jordan:4 feedforward:1 split:1 easy:1 iii:1 iterate:1 zi:4 idea:2 pallett:1 byblos:3 expression:1 utility:1 defense:1 speech:8 amount:2 subregions:3 outperform:1 estimated:1 iz:1 sum:2 year:1 parameterized:2 reasonable:1 decision:13 comparable:1 layer:4 franco:1 relatively:1 department:1 according:1 makhoul:5 smaller:1 em:6 modification:1 computationally:1 equation:4 turn:1 needed:1 available:3 gaussians:1 apply:1 hierarchical:9 alternative:1 newton:1 unbalance:2 conquer:2 objective:1 separate:2 street:3 hmm:31 ying:4 negative:2 implementation:1 zt:1 redefined:1 perform:3 zavaliagkos:1 benchmark:1 frame:1 csr:4 pair:1 trainable:2 continous:1 hanson:2 acoustic:1 able:1 usually:2 lund:1 program:1 built:2 hybrid:4 advanced:1 improve:1 technology:2 prior:9 approximator:1 h2:1 switched:2 xp:1 consistent:1 principle:1 llz:4 pryzbocki:1 cd:3 sparse:1 boundary:3 vocabulary:2 llc:4 hmmbased:1 computes:1 made:3 lz:1 ameliorated:1 corpus:2 continuous:16 iterative:2 table:7 domain:1 hierarchically:1 whole:3 big:1 contextindependent:2 lie:1 tied:8 weighting:1 tin:1 specific:1 xt:4 gating:2 list:1 experimented:1 workshop:1 effectively:1 bbn:1 ci:4 entropy:1 explore:1 llci:1 nested:1 ma:2 viewed:2 fisher:1 feasible:1 hard:3 specifically:1 except:1 total:1 pas:1 experimental:2 incorporate:1 tested:1
7,167
93
477 A COMPUTATIONA.LLY ROBUST ANATOlVIICAL MODEL FOR RETIN.AL DIRECTIONAL SELECTI\l ITY Norberto M. Grzywacz Center BioI. Inf. Processing MIT, E25-201 Cambridge, MA 02139 Franklin R. Amthor Dept. Psychol. Univ. Alabama Birmingham Birmingham, AL 35294 ABSTRACT We analyze a mathematical model for retinal directionally selective cells based on recent electrophysiological data, and show that its computation of motion direction is robust against noise and speed. INTROduCTION Directionally selective retinal ganglion cells discriminate direction of visual motion relatively independently of speed (Amthor and Grzywacz, 1988a) and with high contrast sensitivity (Grzywacz, Amthor, and Mistler, 1989). These cells respond well to motion in the "preferred" direction, but respond poorly to motion in the opposite, null, direction. There is an increasing amount of experimental work devoted to these cells. Three findings are particularly relevant for this paper: 1- An inhibitory process asymmetric to every point of the receptive field underlies the directional selectivity of ON-OFF ganglion cells of the rabbit retina (Barlow and Levick, 1965). This distributed inhibition allows small motions anywhere in the receptive field center to elicit directionally selective responses. 2- The dendritic tree of directionally selective ganglion cells is highly branched and most of its dendrites are very fine (Amthor, Oyster and Takahashi, 1984; Amthor, Takahashi, and Oyster, 1988). 3The distributions of excitatory and inhibitory synapses along these cells' dendritic tree appear to overlap. (Famiglietti, 1985). Our own recent experiments elucidated some ofthe spatiotemporal properties of these cells' receptive field. In contrast to excitation , which is transient with stimulus, the inhibition is sustained and might arise from sustained amacrine cells (Amthor and Grzywacz, 1988a). Its spatial distribution is wide, extending to the borders of the receptive field center (Grzywacz and Amthor, 1988). Finally, the inhibition seems to be mediated by a high-gain shunting, not hyperpolarizing, synapse, that is, a synapse whose reversal potential is close to cell's resting potential (Amthor and Grzywacz, 1989). 478 Grzywacz and Amthor In spite of this large amount of experimental work, theoretical efforts to put these pieces of evidence into a single framework have been virtually inexistent. We propose a directional selectivity model based on our recent data on the inhibition's spatiotemporal and nonlinear properties. This model, which is an elaboration of the model of Torre and Poggio (1978), seems to account for several phenomena related to retinal directional <;eledivity. THE ]\IIODEL Figure 1 illustrates the new model for retinal directional selectivity. In this modd, a stimulus moving in the null direction progressively activates receptive field regions linked to synapses feeding progressively more distal dendrites Of the ganglion cells . Every point in the receptive field center activates adjacent excitatory a~d inhibitory synapses. The inhibitory synapses are assumed to cause shunting inhibition. ("'"e also formulated a pre-ganglionic version of this model, which however, is outside the scope of this paper) . .-- NULL FIGURE 1. The new model for retinal directional selectivity. This model is different than that of Poggio and Koch (1987), where the motion axis is represented as a sequence of activation of different dendrites. Furthermore, in their model, the inhibitory synapses must be closer to the soma than the excitatory ones . (However, our model is similar a model proposed, and argued against, elsewhere (Koc,h, Poggio, and Torre, 1982). An advantage of our model is that it accounts for the large inhibitory areas t.o most points of the receptive field (Grzywacz and Amthor, 1988). Also, in the new model, the distributions of the excitatory and inhibitory synapses overlap along the dendritic tree, as suggested (Famiglietti, 1985). Finally, the dendritic tree of ONOFF directionally selective ganglion cells (inset- Amthor, Oyster, and Takahashi, A Computationally Robust Anatomical Model 1984) is consistent with our model. The tree's fine dendrites favor the multiplicity of directional selectivity and help to deal with noise (see below). In this paper, we make calculations with an unidimensional version of the model dealing with motions in the preferred and null directions. Its receptive field maps into one dendrite of the cell. Set the origin of coordinates of the receptive field to be the point where a dot moving in the null direction enters the receptive field. Let S be the size of the receptive field. Next, set the origin of coordinates in the dendri te to be the soma and let L be the length of the dendrite. The model postulates that a point z in the receptive field activates excitatory and inhibitory synapses in pain t z = zL/ S of the dendrite. The voltages in the presynaptic sites are assumed to be linearly related to the stimulus, [(z,t), that is, there are functions fe{t) and li(t) such that the excitatory, {3e(t), and inhibitory, {3i(t), presynaptic voltages of the synapses to position ;r in the dendrite are . J = e, . l, where "*" stands for convolution. We assume that the integral of Ie is zero, (the excitation is transient), and that the integral of Ii is positive. (In practice, gamma distribution functions for Ii and the derivatives of such functions for Ie were used in this paper's simulations.) The model postulates that the excitatory, ge, and inhibitory, gi, postsynaptic conductances are rectified functions of the presynaptic voltages. In some examples, we use the hyperbolic tangent as the rectification function: where Ij and T; are constants. In other examples, we use the rectification functions described in Grzywacz and Koch (1987), and their model of ON-OFF rectifications. For the postsynaptic site, we assume, without loss of generality, zero reversal potential and neglect the effect of capacitors (justified by the slow time-courses of excitation and inhibition). Also, we assume that the inhibitory synapse leads to shunting inhibition, that is, its conductance is not in series with a battery. Let Ee be the voltage of the excitatory battery. In general, the voltage, V, in different positions of the dendrite is described by the cable equation: d2~~:, t) = Ra (-Ee!}e (z, t) + V (z, t) (ge (z, t) + 9j (z, t) + 9,.)), where Ra is the axoplasm resistance per unit length, g,. is the resting membrane conductance, and the tilde indicates that in this equation the conductances are given per unit length. 479 480 Grzywacz and Amthor For the calculations illustrated in this paper, the stimuli are always delivered to the receptive field through two narrow slits. Thus, these stimuli activate synapses in two discrete positions of a dendrite. In this paper, we only show results for square wave and sinusoidal modulations of light, however, we also performed calculations for more general motions. The synaptic sites are small so that their resting conductances are negligible, and we assume that outside these sites the excitatory and inhibitory conductances are zero. In this case, the equation for outside the synapses is: d2U dy2 = U, = where we defined A 1/(Ra y,,)1/2 (the length constant), U = V/Ee, and y The boundary conditions used are = = Z/A. = where L L/ A, and where if R, is the soma's input resistance, then p R, /(RaA) (the dendritic-to-soma cond uctance ratio). The first condition means that currents do not flow through the tips of the dendrites. Finally, label by 1 the synapse proximal to the soma, and by 2 the distal one; the boundary conditions at the synapses are 1I~lIj = lim II> 111 1I<lIj lim U 1I~lIj U, j = 1,2, j = 1,2, (I ) where 7'e = geRa).. and 7'i = 9iRa)... It can be shown that the relative inhibitory strength for motions in the preferred direction decreases with L and increases with p. Thus, to favor conditions for multiplicity of direction selectivity in the receptive field, we perform calculations with L -+ 00 and p = 1. The strengths of the excitatory syna.pses are set such that their contribution to somatic voltage in the absence of inhibition is in variant with position. Finally, we ensure that the excitatory synapses never sat urate. Under these conditions, one can show that the voltage in the soma is: U (0) = {27'e,2 + (7'e,2 + 7'i,2 + 2) 7'e,d e2~y - (7'?,2 + 7'j,2) 7'e,1 , ((7'i,l +2)7'i,2 + 27'j,1 +4)e 26Y - 7'i,17'i,2 (2) where 6y is the distance between the synapses. A fina.l quantity, which is used in this paper is the directional selectivity index It. Let Up and Un be the total responses to the second slit in the apparent motion in the preferred and null directions respectively. {Alternatively, for the sinusoidal A Computationally Robust Anatomical Model motion, these quantities are the respective average responses .) \Ve follow Grzywacz and Koch (198i) and define (3) RESULTS This section presents the results of calcuhtions ba."ed on the modd. \Ve address: the multiple computations of directional seir->divity in the {.?ells? receptive fields; the robustness of these computations again~t noise: I h" robustness of these computations against speed. Figure 2 plots the degree of directional selectivity for apparent lIlotions activating two synapses as function of the synapses' distan,'e in a dendrite (computed from Equations 2 and 3). 1. ~--------------------------- 32 .".. 16 ...c:: -.... -.., ... B >~ '" .. .5 u ~ 2 c: a '" u "- 0.5 c'" .0 .00 l!~IG URE 1 .5 1.0 . 50 Dendritic Distance (>.1 2. Locality of lnkraction betwt't"'n 2.0 synap"'t'~ a(I1\-'(1l,'" hv apparent mo- tions. It can be shown that the critical parameter controlling whether a certain synaptic distance produces a criterion directional selectivity is rj (Equation 1). As the parameter rj increases, the criterion distance decreases , Thus , "ince in retinal directionally selective cells the inhibition has high gain (Amthor and Grzywa<."z, 1989) and the dendrites are fine (Amthor, Oyster and Takahashi, 1984; Amthor, Takahashi, and Oyster, 1988), then rj is high, and motions anywhere in receptive field should elicit directionally selective responses (Barlow and Levick, 1965). In other words, the model's receptive field computes motion direction multiple times. Next, we show that the high inhibitory gain and the cells' fine dendrites help to deal with noise, and thus . may explain the high contrast sensitivity (0.5% contrast- 481 482 Grzywacz and Amthor Grzywacz, Amthor, and Mistler, 1989) of the cells' directional selectivity. This conclusion is illustrated in Figure 3's multiple plots. .. - -Looo .. ~ 1.0 - OUTPUT NOISE J ~I -"11 INIIIIT DAY 1IfI\/1 NOIS? VCCJ1ATDA'f 1IfI\/1 NOI. !. ~1 , ".., ......a -NlgII , .. .50 .. N ,, ... oo-===--e:;..........::.-....;::.,---==--..::.. 00 10 ? .0 Response ?. 00 .0 10 .00 Response 4.0 . 1.0 Aupan .. FIGURE 3. The model's computatifm of direction is robust against additive noise in the output, and in the excitatory and inhibitory inputs. To generate this figure, we used Equation 2 assuming that a Gaussian noise is added to the cell's output, excitatory input, or inhibitory input. {In the latter case, we used an approximation that assumes small standard deviation for the inhibitory input's noise.) Once again the critical parameter is the ri defined in Equation 1. The larger this parameter is, the better the model deals with noise. In the case of output noise, an increase of the parameter separatt's the preferred and null mean responses. For noise in the excitatory input, a parameter increase not only separates the means, but also reduces the standard deviation: Shunting inhibition SHUnTS down the noise. Finally, the most dramatic improvement occurs when the noise is in the inhibitory input. (In all these plots, the parameter increase is always by a factor of three.) Since for retinal direc tionally selective ganglion cells, ri is high (high inhibitory gain and fine dendrites), we conclude that the cells' mechanism are particularly well suited to deal with noise. For sinusoidal motions, the directional selectivity is robust for a large range of temporal frequencies provided that the frequencies are sufficiently low (Figure 4). (Nevertheless, the cell's preferred direction responses may be sharply tuned to either temporal frequency or speed- Amthor and Grzywacz, 1988). A Computationally Robust Anatomical Model I -HI", rl - 'llHl rl I ! , \ I I I I I I i 2.00 ~ , I iis :: " I -- 20.0 200. --------------- ,i i i is 2.00 20 .0 FrtQlltncy ~I IE-roo FIGURE 4. Directional selectivity is robust against speed modulation. To generate this curve, we subtracted the average respons~ to a isolated flickering slit from the preferred and null average responses (from Equation 2). This robustness is due to the invariance with speed for low speeds of the relative temporal phase shift between inhibition and excitation. Since the excitation has band-pass characteristics, it leads the stimulus by a constant phase. On the other hand, the inhibition is delayed and advanced in the preferred and null directions respectively, due to the asymmdric spatial integration. The phase shifts due to this integration are also speed invariant. CONCLUSIONS We propose a new model for retinal directional selectivity. The shunting inhibition of ganglion cells (Torre and Poggio, 1978), which is temporally sustained, is the main biophysical mechanism of the model. It postulates that for null direction motion, the stimulus activates regions of the receptive field that are linked to excitatory and inhibitory synapses, which are progressively farther away from the soma. This models accounts for: 1- the distribution of inhibition around points of the receptive field (Grzywacz and Amthor, 1988); 2- the apparent full overlap of the distribution of excitatory and inhibitory synapses along the dendritic trees of directionally selective ganglion cells (Famiglietti, 1985); 3- the multiplicity of directionally selective regions (Barlow and Levick, 1965); 4- the high contrast sensitivity of the cells' directional selectivity (Grzywacz, Amthor, and Mistler, 1989); 5- the relative in variance of directional selectivity on stimulus speed (Amthor and Grzywacz, 1988). Two lessons of our model to neural network modeling are: Threshold is not the only neural mechanism, and the basic computational unit may not be a neuron 483 484 Grzywacz and Amthor but a piece of membrane (Grzywacz and Poggio, 1989). In our model, nonlinear interactions are relatively confined to specific dendritic tree branches (Torre and Poggio, 1978). This allows local computations by which single cells might generate receptive fields with multiple directionally selective regions, as observed by Barlow and Levick (1965). Such local computations could not occur if the inhibition only worked through a reduction in spike rate by somatic hyperpolarization. Thus, most neural network models may be biologically irrelevant, since they are built upon a too simple model of the neuron. The properties of a network depend strongly on its basic elements. Therefore, to understand the computations of biological networks. it may be essential to first understand the basic biophysical mechanisms of information processing before developing complex networks. ACKNOWLEDGMENTS We thank Lyle Borg-Graham and Tomaso Poggio for helpful discussions. Also, we thank Consuelita Correa for help with the figures. N .M.G. was supported by grant BNS-8809528 from the National Science Foundation, by the Sloan Foundation, and by a grant to Tomaso Poggio and Ellen Hildreth from the Office of Naval Research, Cognitive and Neural Systems Division. F.R.A. was supported by grants from the National Institute of Health (EY05070) and the Sloan Foundation. REFERENCES Amthor & Grzywacz (1988) Invest. Ophthalmol. Vi". Sci. 29:225. Amthor & Grzywacz (1989) Retinal Directional Selectivity Is Accounted for by Shunting Inhibition. Submitted for Publication. Amthor, Oyster & Takahashi (1984) Brain Res. 298:187. Amthor, Takahashi & Oyster (1989) J. Compo Neurol. In Press. Barlow & Levick (1965) J. Physiol. 178:477. Famiglietti (1985) Neuro"ci. Abst. 11:337. Grzywacz & Amthor (1988) Neurosci. Ab"t. 14:603. Grzywacz, Amthor & Mistler (1989) Applicability of Quadratic and Threshold Models to Motion Discrimination in the Rabbit Retina. Submitted for Publication. Grzywacz & Koch (1987) Synapse 1:417. Grzywacz & Poggio (1989) In An Introduction to Neural and Electronic Networks. Zornetzer, Davis & Lau, Eds. Academic Press, Orlando, USA. In Press. Koch, Poggio & Torre (1982) Philos . Tran". R. Soc. B 298:227. Poggio & Koch (1987) Sci. Am. 256:46. Torre & Poggio (1978) Proc. R. Soc. B 202:409.
93 |@word version:2 seems:2 d2:1 simulation:1 dramatic:1 reduction:1 series:1 tuned:1 franklin:1 current:1 activation:1 must:1 physiol:1 additive:1 hyperpolarizing:1 plot:3 progressively:3 discrimination:1 farther:1 compo:1 mathematical:1 along:3 borg:1 sustained:3 ra:3 tomaso:2 brain:1 increasing:1 provided:1 null:10 finding:1 temporal:3 every:2 zl:1 unit:3 grant:3 appear:1 positive:1 negligible:1 before:1 local:2 ure:1 modulation:2 might:2 shunt:1 range:1 acknowledgment:1 oyster:7 practice:1 lyle:1 area:1 elicit:2 hyperbolic:1 pre:1 word:1 spite:1 close:1 put:1 map:1 center:4 independently:1 rabbit:2 ellen:1 ity:1 coordinate:2 grzywacz:25 controlling:1 origin:2 element:1 particularly:2 asymmetric:1 observed:1 enters:1 hv:1 region:4 decrease:2 noi:1 battery:2 depend:1 upon:1 division:1 represented:1 univ:1 activate:1 outside:3 whose:1 apparent:4 larger:1 roo:1 favor:2 gi:1 delivered:1 directionally:10 sequence:1 advantage:1 biophysical:2 abst:1 propose:2 tran:1 interaction:1 relevant:1 poorly:1 invest:1 extending:1 produce:1 help:3 tions:1 oo:1 ij:1 soc:2 direction:15 torre:6 transient:2 argued:1 respons:1 feeding:1 activating:1 orlando:1 dendritic:8 biological:1 koch:6 sufficiently:1 around:1 scope:1 mo:1 inexistent:1 proc:1 birmingham:2 label:1 ophthalmol:1 mit:1 activates:4 always:2 gaussian:1 amacrine:1 voltage:7 office:1 publication:2 ira:1 naval:1 improvement:1 indicates:1 ps:1 contrast:5 am:1 helpful:1 dy2:1 selective:11 i1:1 spatial:2 integration:2 ell:1 field:20 once:1 never:1 stimulus:8 retina:2 gamma:1 ve:2 national:2 delayed:1 phase:3 ab:1 conductance:6 highly:1 light:1 devoted:1 integral:2 closer:1 poggio:12 respective:1 tree:7 re:1 isolated:1 theoretical:1 modeling:1 applicability:1 deviation:2 too:1 spatiotemporal:2 proximal:1 sensitivity:3 ie:3 off:2 tip:1 e25:1 again:2 postulate:3 cognitive:1 derivative:1 li:1 takahashi:7 potential:3 account:3 sinusoidal:3 retinal:9 sloan:2 vi:1 piece:2 performed:1 analyze:1 linked:2 wave:1 contribution:1 square:1 variance:1 characteristic:1 ofthe:1 lesson:1 directional:18 rectified:1 submitted:2 explain:1 synapsis:17 koc:1 synaptic:2 ed:2 against:5 frequency:3 e2:1 gain:4 lim:2 electrophysiological:1 levick:5 day:1 follow:1 response:9 synapse:5 strongly:1 generality:1 furthermore:1 anywhere:2 hand:1 nonlinear:2 hildreth:1 usa:1 effect:1 barlow:5 illustrated:2 deal:4 distal:2 adjacent:1 davis:1 excitation:5 criterion:2 correa:1 motion:16 ince:1 rl:2 hyperpolarization:1 resting:3 cambridge:1 philos:1 tionally:1 dot:1 moving:2 inhibition:16 own:1 recent:3 inf:1 irrelevant:1 selectivity:16 certain:1 ganglionic:1 ii:4 branch:1 multiple:4 full:1 rj:3 reduces:1 academic:1 calculation:4 elaboration:1 shunting:6 underlies:1 variant:1 basic:3 neuro:1 lly:1 confined:1 cell:24 justified:1 fine:5 virtually:1 flow:1 capacitor:1 ee:3 ifi:2 opposite:1 unidimensional:1 shift:2 whether:1 effort:1 modd:2 resistance:2 cause:1 amount:2 band:1 generate:3 inhibitory:21 per:2 anatomical:3 discrete:1 soma:7 nevertheless:1 threshold:2 branched:1 respond:2 electronic:1 graham:1 hi:1 syna:1 quadratic:1 elucidated:1 strength:2 occur:1 sharply:1 worked:1 ri:2 speed:9 bns:1 relatively:2 developing:1 membrane:2 postsynaptic:2 cable:1 biologically:1 lau:1 invariant:1 multiplicity:3 rectification:3 computationally:3 equation:8 mechanism:4 ge:2 reversal:2 away:1 nois:1 subtracted:1 robustness:3 assumes:1 ensure:1 neglect:1 added:1 quantity:2 occurs:1 spike:1 receptive:20 pain:1 distance:4 separate:1 thank:2 sci:2 presynaptic:3 assuming:1 length:4 index:1 ratio:1 fe:1 ba:1 perform:1 convolution:1 neuron:2 tilde:1 somatic:2 synap:1 alabama:1 narrow:1 address:1 suggested:1 below:1 built:1 overlap:3 critical:2 advanced:1 temporally:1 axis:1 psychol:1 mediated:1 health:1 lij:3 tangent:1 relative:3 loss:1 foundation:3 degree:1 consistent:1 excitatory:16 elsewhere:1 course:1 looo:1 raa:1 supported:2 accounted:1 understand:2 institute:1 wide:1 distributed:1 boundary:2 curve:1 stand:1 computes:1 ig:1 preferred:8 dealing:1 sat:1 assumed:2 conclude:1 alternatively:1 zornetzer:1 un:1 robust:8 dendrite:15 complex:1 main:1 linearly:1 neurosci:1 border:1 noise:14 arise:1 site:4 slow:1 position:4 down:1 specific:1 inset:1 neurol:1 evidence:1 essential:1 ci:1 te:1 illustrates:1 direc:1 locality:1 suited:1 ganglion:8 visual:1 ma:1 bioi:1 formulated:1 flickering:1 absence:1 total:1 discriminate:1 slit:3 experimental:2 invariance:1 pas:1 cond:1 latter:1 dept:1 phenomenon:1
7,168
930
FINANCIAL APPLICATIONS OF LEARNING FROM HINTS Yaser s. Abu-Mostafa California Institute of Technology and NeuroDollars, Inc. e-mail: yaser@caltech.edu Abstract The basic paradigm for learning in neural networks is 'learning from examples' where a training set of input-output examples is used to teach the network the target function. Learning from hints is a generalization of learning from examples where additional information about the target function can be incorporated in the same learning process. Such information can come from common sense rules or special expertise. In financial market applications where the training data is very noisy, the use of such hints can have a decisive advantage. We demonstrate the use of hints in foreign-exchange trading of the U.S. Dollar versus the British Pound, the German Mark, the Japanese Yen, and the Swiss Franc, over a period of 32 months. We explain the general method of learning from hints and how it can be applied to other markets. The learning model for this method is not restricted to neural networks. 1 INTRODUCTION When a neural network learns its target function from examples (training data), it knows nothing about the function except what it sees in the data. In financial market applications, it is typical to have limited amount of relevant training data, with high noise levels in the data. The information content of such data is modest, and while the learning process can try to make the most of what it has, it cannot create new information on its own. This poses a fundamental limitation on the 412 Yaser S. Abu-Mostafa learning approach, not only for neural networks, but for all other models as well. It is not uncommon to see simple rules such as the moving average outperforming an elaborate learning-from-examples system. Learning from hints (Abu-Mostafa, 1990, 1993) is a value-added feature to learning from examples that boosts the information content in the data. The method allows us to use prior knowledge about the target function, that comes from common sense or expertise, along with the training data in the same learning process. Different types of hints that may be available in a given application can be used simultaneously. In this paper, we give experimental evidence of the impact of hints on learning performance, and explain the method in some detail to enable the readers to try their own hints in different markets. Even simple hints can result in significant improvement in the learning performance. Figure 1 shows the learning performance for foreign exchange (FX) trading with and without the symmetry hint (see section 3), using only the closing price history. The plots are the Annualized Percentage Returns (cumulative daily, unleveraged, transaction cost included), for a sliding one-year test window in the period from April 1988 to November 1990, averaged over the four major FX markets with more than 150 runs per currency. The error bar in the upper left corner is 3 standard deviations long (based on 253 trading days, assuming independence between different runs). The plots establish a statistically significant differential in performance due to the use of hints. This differential holds for all four currencies. 10 r---------~--------_r--------~--------~r_--------~ average: without hint -;-I" average: with hint;..':':.--, I 8 " ~ 30 /"_,,, " 6 / ,-' " " " ,-'" iI'?.' 4 J ,-" .. , ..I , .....r';' _",'''.-~''''''''.'-'' ,~_" .." 2 .. ,,-" _A' ,-,,-, ..'" ,.1 o '.~:<.~.~........................................... -2 ______ ________L -________ 100 150 200 250 Test Day Number ~--------~--------~-- o 50 ~ ~ Figure 1: Learning performance with and without hint Since the goal of hints is to add information to the training data, the differential in performance is likely to be less dramatic if we start out with more informative training data. Similarly, an additional hint may not have a pronounced effect if Financial Applications of Learning from Hints 413 we have already used a few hints in the same application. There is a saturation in performance in any market that reflects how well the future can be forecast from the past. (Believers in the Efficient Market Hypothesis consider this saturation to be at zero performance). Hints will not make us forecast a market better than whatever that saturation level may be. They will, however, enable us to approach that level through learning. This paper is organized as follows. Section 2 characterizes the notion of very noisy data by defining the '50% performance range'. We argue that the need for extra information in financial market applications is more pronounced than in other pattern recognition applications. In section 3, we discuss our method for learning from hints. We give examples of different types of hints, and explain how to represent hints to the learning process. Section 4 gives result details on the use of the symmetry hint in the four major FX markets. Section 5 provides experimental evidence that it is indeed the information content of the hint, rather than the incidental regularization effect, that results in the performance differential that we observe. 2 FINANCIAL DATA This section provides a characterization of very noisy data that applies to the financial markets. For a broad treatment of neural-network applications to the financial markets, the reader is referred to (Abu-Mostafa et al, 1994). Other Information Input X Input X -- -- Target Ouput MARKET NEURAL NETWORK -... -- y Forecast y Figure 2: Illustration of the nature of noise in financial markets Consider the market as a system that takes in a lot of information (fundamentals, news events, rumors, who bought what when, etc.) and produces an output y (say up/down price movement for simplicity) . A model, e.g., a neural network, attempts 414 Yaser S. Abu-Mostafa to simulate the market (figure 2), but it takes an input x which is only a small subset of the information. The 'other information' cannot be modeled and plays the role of noise as far as x is concerned. The network cannot determine the target output y based on x alone, so it approximates it with its output y. It is typical that this approximation will be correct only slightly more than half the time. What makes us consider x 'very noisy' is that y and y agree only! + f. of the time (50% performance range). This is in contrast to the typical pattern recognition application, such as optical character recognition, where y and y agree 1 - f. of the time (100% performance range). It is not the poor performance per se that poses a problem in the 50% range, but rather the additional difficulty of learning in this range. Here is why. In the 50% range, a performance of ! + f. is good, while a performance of ! f. is disastrous. During learning, we need to distinguish between good and ~ad hypotheses based on a limited set of N examples. The problem with the 50% range is that the number of bad hypotheses that look good on N points is huge. This is in contrast to the 100% range where a good performance is as high as 1 - f.. The number of bad hypotheses that look good here is limited. Therefore, one can have much more confidence in a hypothesis that was learned in the 100% range than one learned in the 50% range. It is not uncommon to see a random trading policy making good money for a few weeks, but it is very unlikely that a random character recognition system will read a paragraph correctly. Of course this problem would diminish if we used a very large set of examples, because the law of large numbers would make it less and less likely that y and y can agree! + f. of the time just by 'coincidence'. However, financial data has the other problem of non-stationarity. Because of the continuous evolution in the markets, old data may represent patterns of behavior that no longer hold. Thus, the relevant data for training purposes is limited to fairly recent times. Put together, noise and non-stationarity mean that the training data will not contain enough information for the network to learn the function. More information is needed, and hints can be the means of providing it. 3 HINTS In this section, we give examples of different types of hints and discuss how to represent them to the learning process. We describe a simple way to use hints that allows the reader to try the method with minimal effort. For a more detailed treatment, please see (Abu-Mostafa, 1993). As far as our method is concerned, a hint is any property that the target function is known to have. For instance, consider the symmetry hint in FX markets as it applies to the U.S. Dollar versus the German Mark (figure 3). This simple hint asserts that if a pattern in the price history implies a certain move in the market, then this implication holds whether you are looking at the market from the U.S. Dollar viewpoint or the German Mark viewpoint. Formally, in terms of normalized prices, the hint translates to invariance under inversion of these prices. Is the symmetry hint valid? The ultimate test for this is how the learning performance is affected by the introduction of the hint. The formulation of hints is an art. 415 Financial Applications of Learning from Hints We use our experience, common sense, and analysis of the market to come up with a list of what we believe to be valid properties of this market. We then represent these hints in a canonical form as we will see shortly, and proceed to incorporate them in the learning process. The improvement in performance will only be as good as the hints we put in. ? ? u.s. GERMAN DOLLAR MARK Figure 3: Illustration of the symmetry hint in FX markets The canonical representation of hints is a more systematic task. The first step in representing a hint is to choose a way of generating 'virtual examples' of the hint. For illustration, suppose that the hint asserts that the target function y is an odd function of the input. An example of this hint would have the form y( -x) = -y(x) for a particular input x. One can generate as many virtual examples as needed by picking different inputs. After a hint is represented by virtual examples, it is ready to be incorporated in the learning process along with the examples of the target function itself. Notice that an example of the function is learned by minimizing an error measure, say (y(x) - y(x))2, as a way of ultimately enforcing the condition y(x) = y(x). In the same way, a virtual example of the oddness hint can be learned by minimizing (y(x) + y(-x))2 as a way of ultimately enforcing the condition y(-x) = -y(x). This involves inputting both x and -x to the network and minimizing the difference between the two outputs. It is easy to show that this can be done using backpropagation (Rumelhart et al, 1986) twice. The generation of a virtual example of the hint does not require knowing the value of the target function; neither y(x) nor y( -x) is needed to compute the error for the oddness hint. In fact, x and -x can be artificial inputs. The fact that we do not need the value of the target function is crucial, since it was the limited resource of examples for which we know the value of the target function that got us interested in hints in the first place. On the other hand, for some hints, we can take the examples of the target function that we have, and employ the hint to duplicate these examples. For instance, an example y(x) = 1 can be used to infer a second example y( -x) = -1 using the oddness hint. Representing the hint by duplicate examples is an easy way to try simple hints using the same software that we use for learning from examples. Yaser S. Abu-Mostafa 416 Let us illustrate how to represent two common types of hints. Perhaps the most common type is the invariance hint. This hint asserts that i)(x) = i)(x / ) for certain pairs x, x'. For instance, "i) is shift-invariant" is formalized by the pairs x, x' that are shifted versions of each other. To represent the invariance hint, an invariant pair (x, x') is picked as a virtual example. The error associated with this example is (y(x) - y(x /?2. Another related type of hint is the monotonicity hint. The hint asserts for certain pairs x, x' that i)(x) :5 i)(x / ). For instance, "i) is monotonically non decreasing in x" is formalized by the pairs x, x' such that x < x'. One application where the monotonicity hint occurs is the extension of personal credit. If person A is identical to person B except that A makes less money than B, then the approved credit line for A cannot exceed that of B. To represent the monotonicity hint, a monotonic pair (X,X/) is picked as a virtual example. The error associated with this example is given by (y(x) - Y(X /?2 if y(x) > y(x / ) and zero if y(x) :5 y(x'). 4 FX TRADING We applied the symmetry hint in the four FX markets of the U.S. Dollar versus the British Pound, the German Mark, the Japanese Yen, and the Swiss Franc. In each case, only the closing prices for the preceding 21 days were used for inputs. The objective (fitness) function we chose was the total return on the training set, and we used simple filtering methods on the inputs and outputs of the networks. In each run, the training set consisted of 500 days, and the test was done on the following 253 days. All four currencies show an improved performance when the symmetry hint is used. Roughly speaking, we are in the market half the time, each trade takes 4 days, the hit rate is close to 50%, and the A.P.R. without hint is 5% and with hint is 10% (the returns are annualized, unleveraged, and include the transaction cost; spread and average slippage). Notice that having the return as the objective function resulted in a fairly good return with a modest hit rate. 5 CROSS CHECKS In this final section, we report more experimental results aimed at validating our claim that the information content of the hint is the reason behind the improved performance. Why is this debatable? A hint plays an incidental role as a constraint on the neural network during learning, since it restricts the solutions the network may settle in. Because overfitting is a common problem in learning from examples, any restriction may improve the out-of-sample performance by reducing overfitting (Akaike, 1969, Moody, 1992). This is the idea behind regularization. To isolate the informative role from the regularizing role of the symmetry hint, we ran two experiments. In the first experiment, we used an uninformative hint, or 'noise' hint, which provides a random target output for the same inputs used in the examples of the symmetry hint. Figure 4 contrasts the performance of the noise hint with that of the real symmetry hint, averaged over the four currencies. Notice that the performance with the noise hint is close to that without any hint (figure 1), which is consistent with the notion of uninformative hint. The regularization effect seems to be negligible. 417 Financial Applications of Learning from Hints 10 ~--------~~--------~----------~----------~-----------n 8 6 4 2 o -2 ~ __________ o ~ ________ 50 ~ __________ 100 ~ __________ ~ __________-u 200 1 50 250 Test Day Number Figure 4: Performance of the real hint versus a noise hint 10 .------------r------------.------------r----------~r_----------~ with fa 1 s e ):l+tlt--:::=':with JJi!trI' hint ---- . I 3a ,,-.,,-?? 5 '~ ~ I ~ " +> Q) IX .. "'"~~.~.~.~.~.~.~.~~~~~~~.~.~.~ ~ ~-.~ -.-.?.-.- ---.-'-.~-.-.~'.~.?~-------.-.-.-.-. c .?. 0 .. . . - J .. - .. .. ....... ... ........... ..................................... . Q) "" f\1 +> C Q) tl ~ .,. -5 Q) "0 Q) .....-< N f\1 -10 " C ~ -15 -20 ~-- o ________ ~ 50 __________ ~ __________- L__________ 100 150 ~ __________--u 200 Test Day Number Figure 5: Performance of the real hint versus a false hint 2 50 418 Yaser S. Abu-Mostafa In the second experiment, we used a harmful hint, or 'false' hint, in place of the symmetry hint. The hint takes the same examples used in the symmetry hint and asserts antisymmetry instead. Figure 5 contrasts the performance of the false hint with that of the real symmetry hint. As we see, the false hint had a detrimental effect on the performance. This is consistent with the hypothesis that the symmetry hint is valid, since its negation results in worse performance than no hint at all. Notice that the transaction cost is taken into consideration in all of these plots, which works as a negative bias and amplifies the losses of bad trading policies. CONCLUSION We have explained learning from hints, a systematic method for combining rules and data in the same learning process, and reported experimental results of a statistically significant improvement in performance in the four major FX markets that resulted from using a simple symmetry hint. We have described different types of hints and simple ways of using them in learning, to enable the readers to try their own hints in different markets. Acknowledgements I would like to acknowledge Dr. Amir Atiya for his valuable input. I am grateful to Dr. Ayman Abu-Mostafa for his expert remarks. References Abu-Mostafa, Y. S. (1990), Learning from hints in neural networks, Journal of Complexity 6, pp. 192-198. Abu-Mostafa, Y. S. (1993), A method for learning from hints, Advances in Neural Information Processing Systems 5, S. Hanson et al (eds), pp. 73-80, MorganKaufmann. Abu-Mostafa, Y. S. et al (eds) (1994), Proceedings of Neural Networks in the Capital Markets, Pasadena, California, November 1994. Akaike, H. (l969), Fitting autoregressive models for prediction, Ann. Inst. Stat. Math. 21, pp. 243-247. Moody, J. (1992), The effective number of parameters: An analysis of generalization and regularization in nonlinear learning systems, in Advances in Neural Information Processing Systems 4, J. Moody et al (eds), pp. 847-854, Morgan Kaufmann. Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986), Learning internal representations by error propagation, Parallel Distributed Processing 1, D. Rumelhart et al, pp. 318-362, MIT Press. Weigend, A., Rumelhart, D., and Huberman, B. (1991), Generalization by weight elimination with application to forecasting, in Advances in Neural Information Processing Systems 3, R. Lippmann et al (eds), pp. 875-882, Morgan Kaufmann.
930 |@word version:1 inversion:1 seems:1 approved:1 dramatic:1 past:1 informative:2 plot:3 alone:1 half:2 amir:1 characterization:1 math:1 provides:3 along:2 differential:4 ouput:1 fitting:1 paragraph:1 indeed:1 market:28 roughly:1 behavior:1 nor:1 decreasing:1 window:1 what:5 inputting:1 hit:2 whatever:1 negligible:1 annualized:2 twice:1 chose:1 limited:5 range:10 statistically:2 averaged:2 swiss:2 backpropagation:1 got:1 confidence:1 cannot:4 close:2 put:2 restriction:1 williams:1 simplicity:1 formalized:2 rule:3 financial:12 his:2 notion:2 fx:8 target:14 play:2 suppose:1 akaike:2 hypothesis:6 rumelhart:4 recognition:4 slippage:1 role:4 coincidence:1 news:1 movement:1 trade:1 valuable:1 ran:1 complexity:1 personal:1 ultimately:2 grateful:1 represented:1 rumor:1 describe:1 effective:1 artificial:1 say:2 noisy:4 itself:1 final:1 advantage:1 debatable:1 relevant:2 combining:1 asserts:5 pronounced:2 amplifies:1 produce:1 generating:1 illustrate:1 stat:1 pose:2 odd:1 involves:1 come:3 trading:6 implies:1 correct:1 enable:3 settle:1 virtual:7 elimination:1 exchange:2 require:1 generalization:3 extension:1 hold:3 diminish:1 credit:2 week:1 claim:1 mostafa:12 major:3 purpose:1 create:1 reflects:1 mit:1 rather:2 r_:2 improvement:3 check:1 contrast:4 dollar:5 sense:3 am:1 inst:1 foreign:2 unlikely:1 pasadena:1 interested:1 art:1 special:1 fairly:2 having:1 identical:1 broad:1 look:2 future:1 report:1 hint:101 few:2 franc:2 employ:1 duplicate:2 simultaneously:1 resulted:2 fitness:1 negation:1 attempt:1 stationarity:2 huge:1 uncommon:2 behind:2 implication:1 daily:1 experience:1 modest:2 old:1 harmful:1 minimal:1 instance:4 cost:3 deviation:1 subset:1 reported:1 person:2 fundamental:2 systematic:2 picking:1 together:1 moody:3 choose:1 dr:2 worse:1 corner:1 expert:1 return:5 inc:1 ad:1 decisive:1 try:5 lot:1 picked:2 characterizes:1 start:1 parallel:1 yen:2 kaufmann:2 who:1 expertise:2 history:2 explain:3 ed:4 pp:6 associated:2 treatment:2 knowledge:1 organized:1 day:8 improved:2 april:1 formulation:1 done:2 pound:2 just:1 hand:1 nonlinear:1 propagation:1 perhaps:1 believe:1 effect:4 contain:1 normalized:1 consisted:1 evolution:1 regularization:4 read:1 during:2 please:1 demonstrate:1 consideration:1 common:6 approximates:1 significant:3 similarly:1 closing:2 had:1 moving:1 longer:1 money:2 etc:1 add:1 own:3 recent:1 certain:3 outperforming:1 caltech:1 morgan:2 additional:3 preceding:1 determine:1 paradigm:1 period:2 monotonically:1 ii:1 sliding:1 currency:4 infer:1 cross:1 long:1 impact:1 prediction:1 basic:1 represent:7 uninformative:2 crucial:1 extra:1 tri:1 isolate:1 validating:1 bought:1 exceed:1 enough:1 concerned:2 easy:2 independence:1 idea:1 knowing:1 translates:1 shift:1 whether:1 ultimate:1 effort:1 forecasting:1 yaser:6 proceed:1 speaking:1 remark:1 se:1 detailed:1 aimed:1 amount:1 atiya:1 generate:1 percentage:1 restricts:1 canonical:2 notice:4 shifted:1 per:2 correctly:1 affected:1 abu:12 four:7 capital:1 neither:1 year:1 weigend:1 run:3 you:1 place:2 reader:4 ayman:1 distinguish:1 jji:1 constraint:1 software:1 simulate:1 optical:1 poor:1 slightly:1 character:2 making:1 explained:1 restricted:1 invariant:2 taken:1 resource:1 agree:3 discus:2 german:5 needed:3 know:2 available:1 observe:1 shortly:1 include:1 establish:1 move:1 objective:2 added:1 already:1 occurs:1 fa:1 detrimental:1 mail:1 argue:1 reason:1 enforcing:2 assuming:1 modeled:1 illustration:3 providing:1 minimizing:3 disastrous:1 teach:1 negative:1 incidental:2 policy:2 upper:1 acknowledge:1 oddness:3 november:2 defining:1 hinton:1 incorporated:2 looking:1 antisymmetry:1 pair:6 hanson:1 california:2 learned:4 boost:1 bar:1 pattern:4 saturation:3 event:1 difficulty:1 representing:2 improve:1 technology:1 ready:1 prior:1 acknowledgement:1 law:1 loss:1 generation:1 limitation:1 filtering:1 versus:5 consistent:2 viewpoint:2 course:1 bias:1 institute:1 distributed:1 valid:3 cumulative:1 autoregressive:1 far:2 tlt:1 transaction:3 lippmann:1 monotonicity:3 overfitting:2 continuous:1 why:2 nature:1 learn:1 symmetry:15 japanese:2 spread:1 noise:8 nothing:1 referred:1 tl:1 elaborate:1 learns:1 ix:1 british:2 down:1 bad:3 list:1 evidence:2 false:4 forecast:3 likely:2 applies:2 monotonic:1 month:1 goal:1 ann:1 price:6 content:4 included:1 typical:3 except:2 reducing:1 huberman:1 total:1 invariance:3 experimental:4 formally:1 internal:1 mark:5 morgankaufmann:1 incorporate:1 regularizing:1
7,169
931
Asymptotics of Gradient-based Neural Network 'fraining Algorithms Sayandev Mukherjee Terrence L. Fine saymukh~ee.comell.edu tlfine~ee.comell.edu School of Electrical Engineering Cornell University Ithaca, NY 14853 School of Electrical Engineering Cornell University Ithaca, NY 14853 Abstract We study the asymptotic properties of the sequence of iterates of weight-vector estimates obtained by training a multilayer feed forward neural network with a basic gradient-descent method using a fixed learning constant and no batch-processing. In the onedimensional case, an exact analysis establishes the existence of a limiting distribution that is not Gaussian in general. For the general case and small learning constant, a linearization approximation permits the application of results from the theory of random matrices to again establish the existence of a limiting distribution. We study the first few moments of this distribution to compare and contrast the results of our analysis with those of techniques of stochastic approximation. 1 INTRODUCTION The wide applicability of neural networks to problems in pattern classification and signal processing has been due to the development of efficient gradient-descent algorithms for the supervised training of multilayer feedforward neural networks with differentiable node functions. A basic version uses a fixed learning constant and updates all weights after each training input is presented (on-line mode) rather than after the entire training set has been presented (batch mode) . The properties of this algorithm as exhibited by the sequence of iterates are not yet well-understood. There are at present two major approaches. Sayandev Mukherjee, Terrence L. Fine 336 Stochastic approximation techniques (Bucklew,Kurtz,Sethares, 1993; Finnoff, 1993; Kuan,Hornik, 1991; White, 1989) study the limiting behavior of the stochastic process that is the piecewise-constant or piecewise-linear interpolation of the sequence of weight-vector iterates (assuming infinitely many U.d. training inputs) as the learning constant approaches zero. It can be shown (Bucklew,Kurtz,Sethares, 1993; Finnoff, 1993) that as the learning constant tends to zero, the fluctuation between the paths and their limit, suitably normalized, tends to a Gaussian diffusion process. Leen and Moody (1993) and Orr and Leen (1993) have considered the Markov process formed by the sequence of iterates (again, assuming infinitely many Li.d. training inputs) for a fixed nonzero learning constant. This approach has the merit of dealing with the nonzero learning constant case anq of linking the study of the training algorithm with the well-developed literature on Markov processes. In particular, it is possible to solve (Leen,Moody, 1993) for the asymptotic distribution of the sequence of weight-vector iterates from the Chapman-Kolmogorov equation after certain assumptions have been used to simplify it considerably. However, the assumptions are unrealistic: in particular, the assumption of detailed balance does not hold in more than one dimension. This approach also fails to establish the existence of a limiting distribution in the general case. This paper follows the method of considering the sequence of weight-vector iterates as a discrete-time continuous state-space Markov process, when the learning constant is fixed and nonzero. We shall first seek to establish the existence of an asymptotic distribution, and then examine this distribution through its first few moments. It can be proved (Mukherjee, 1994), using Foster's criteria (Tweedie, 1976) for the positive-recurrence of a Markov process, that when a single sigmoidal node with one parameter is trained using the iterative form of the basic gradient-descent training algorithm (without batch-processing), the sequence of iterates of the parameter has a limiting distribution which is in general non-Gaussian, thereby qualifying the oftstated claims in the literature (see, for example, (Bucklew,Kurtz,Sethares, 1993; Finnoff, 1993; White, 1989)). However, this method proves to be intractable in the multiple parameter case. 2 THE GENERAL CASE AND LINEARIZATION IN W n The general version of this problem for a neural network", with scalar output involves training ", with the i. i.d. training sequence {( X n' Yn )}, loss function ? (;f, y, w) = ~ [y - ",(~, w W (~ E lRd , Y E lR, w E lRm) and the gradient-descent updating equation for the estimates of the weight vector given by W n+1 = Wn -/LV'1??(;f,y,w)I(X n + ,Yn+l,W n ) = W n + /L[Yn+1 - ",(W n' X n+1)]V'J?."'(w, ;f)lw",x n +1 ? 1 As is customary in this kind of analysis, the training set is assumed infinite, so that {W n}~=O forms a homogeneous Markov process in discrete time. In our analysis, the training data is assumed to come from the model Y = ",(WO,X) + Z, Asymptomatics of Gradient-Based Neural Network Training Algorithms 337 where Z and X are independent, and Z has zero mean and variance (J'2. Hence, the unrestricted Bayes estimator of Y given X, E(YIX) = "l(wO,X), is in the class of neural network estimators, and WO is the goal of training. For convenience, we define W = w - woo Assuming that J.L is small and that after a while, successive iterates, with high probability, jitter about in a close neighborhood of the optimal value wO, we make the important assumption that (1) for some 0 < k < 1 (see Section 4) 1. Applying Taylor series expansions to "l and \1 w"l and neglecting all terms Op(J.L1+ 2 k) and higher, we obtain the following linearized form of the updating equation: (2) where Bn+1 = J.LZn+l \11?"l( W, ;?)11?0,x n+l ' An+1 = G n+1 1m - J.L(\11?"l(w,;?)I~o ,Xn+l)(\11?"l(w,;?)I~o ,Xn+lf + J.L Z n+1 \1-w\1-w"l(w,;?) Iw- o'-n+l X (3) 1m - J.L(Gn+1 - Zn+1 J n+1), (4) = (\11!!."l(w,~)I~o ,Xn+l)(\11!!."l(w,~)I~o,Xn+l f oX I n +1 = (\1w\1w"l(w,;?)lw ) - '-n+l do not depend on W n. The matrices {(A n+1, B n +1 )} form an Li.d. sequence, but An+! and Bn+1 are dependent for each n. Hence the linearized W n again forms a homogeneous Markov process in discrete time. In what follows we analyze this process in the hope that its asymptotics agree with those of the original Markov process. 3 EXISTENCE OF A LIMITING DISTRIBUTION Let A, B, G, J denote random matrices with the common distributions of the i.i.d. sequences {An}, {Bn}, {G n }, and {In} respectively, and let T : lRm ---+ lRm be the random affine transformation w~Aw+B . The following result establishes the existence of a limiting distribution of W n. Lemma 1 (Berger Thm. V, p.162) Suppose E[1og+ IIAII + log+ IIBlll ElogIIA n A n - 1 ? ? ?AtII < < where log+ X = log x V Then the following conclusions hold: , 00? o for some n o. ll.e., (VE > O)(3M.)(Vn)lP(J.L- k IlW nil $ M.) ~ 1 - E. (5) (6) 338 Sayandev Mukherjee, Terrence L. Fine 1. Unique stationary distribution: There exists a unique mndom variable WEIRm, upto distribution, that is stationary with respect to T (i. e., W is independent of T, and TW has the same distribution as W). 2. Asymptotic stationarity: We have convergence in distribution: - 'D - Wn~W. Our choice of norm is the operator norm for the matrix A, /lAII = max IA(A)I, where {A(A)} are the eigenvalues of A, and the Euclidean norm for the vector B, m IIBII 2: IB iI = 2. i=1 We first verify (5). From the inequality "Ix E IR, log+ X ::; x 2 , it is easily seen that if r] is a feedforward net where all activation functions are twice-continuously differentiable in the weights, all hidden-layer activation functions are bounded and have bounded derivatives up to order 2, and if the training sequence (Xn' Yn ) is Li.d. with finite fourth moments, then (5) holds for the Euclidean norm for Band the Frobenius norm for A, IIAII2 = 2::1 2:7=1IAijI2. Since m (max IA(A)1)2 :::; m 2: IA(AW :::; 2: 2: IAij12, (7) i=1 j=1 we see that (5) also holds for the operator norm of A. Assumption (6) forces the product An'" A1 to tend to Omxm almost surely (Berger, 1993, p.146) and therefore removes the dependence of the asymptotic distribution of {W n} on that of the initial value ll:=o. A sufficient condition for (6) is given by the following lemma. Lemma 2 Suppose lEG is positive definite (note that it is positive semidefinite by definition), and for all n, lEAn < 00. Then (6) holds for sufficiently small, positive 11? Proof: By assumption, minA(lEG) = 8 > 0 for some 8. Let Hn = ~ 2:~=1 (G i - ZiJi)' By the Strong Law of Large Numbers applied to the LLd. random matrices (Gi - ZiJi) , we have Hn --+ lEG a.s., so minA(Hn) --+ minA(lEG) a.s. (8) Applying (7) to min A(Hn), it is easily shown that the same conditions on r] and the training sequence that are sufficient for (5) also give sUPn lE[min A(Hn )J2 < 00, which in turn implies that {min A(Hn)} are uniformly integrable. Together with (8), this implies (Loeve, 1977, p.165) that min A(Hn) --+ min A(lEG) in L1. Hence there Asymptomatics of Gradient-Based Neural Network Training Algorithms 339 exists some (nonrandom) N, say, such that ElminA(HN) - minA(EG)1 ::; 8/2. Since IEminA(HN) - minA(EG)1 ::; EI minA(HN) - minA(EG)1 ::; 8/2, we therefore have E [min'\ (~ t,(Gi - ZiJi?) 1? rnin'\(EG) - 8/2 = 8 - 8/2 = 8/2> O. (9) We shall prove that (6) holds for this N (;:::: m) by showing that Elog IIANAN- I ??? Ad 2 = 2lEiog IIANAN-I'" Alii < O. For our choice of norm, we therefore want E[log(max IA(AN .. . Al)I)2] < O. From Jensen's inequality, it is sufficient to have log E[max IA(AN . . . Ad 1]2 < 0, or equivalently, (10) Now, since N is fixed, we can choose /-l small enough that N AN'" Al = 1m - /-l2:)G i - ZiJi) + Op(/-l2). i=l Hence, A(AN ... AI) =1- /-lA(l:~1 (G i I'\(AN ... AI)I' = 1 - - ZiJi? + Op (/-l2), and N is fixed, so 21''\ (t,(G, - Z,J ,?) + Op (1"), giving max I'\(AN' .. Adl' <; 1 - 21' min'\ (t,(Gi - Z,J ,?) + Op (1"), EmaxIA(AN,,?AIW::; I-N8/-l+o(/-l), (11) where we use (9) and the observation that the structure of the last Op(/-l2) term is such that its expectation (guaranteed finite by the hypothesis EAN < 00) is o (/-l2) , or o(/-l), and we also restrict /-l < 1/N 8 so that 1- 2p.Ernin'\ (t,(G, - Z,J,?) > O. From (11), it is clear that (10) holds for all sufficiently small, positive /-l (<< 1/N8) . Therefore (6) holds for n = N. We can combine these two lemmas into the following theorem. Theorem 1 Let 'fJ be a feedforward net where all activation functions are twicecontinuously differentiable in the weights, all hidden-layer activation junctions are bounded and have bounded derivatives up to order 2, and let the training sequence (Xn' Y n ) be i.i.d. with finite moments. Further, assume that lEG is positive definite. Then, for all sufficiently small, positive /-l the sequence of random vectors {W n}~l obtained from the updating equation (2) has a unique limiting distribution. Sayandev Mukherjee, Terrence L. Fine 340 We circumvent the generally intractable problem of finding the limiting distribution by calculating and investigating the behavior of its moments. 4 MOMENTS OF THE LIMITING DISTRIBUTION Let us assume that the mean and variance of the limiting distribution exist, and that Z '" N(O, (1'2). From (2) and the form of An+! and B n+!, it is easy to show that EW = 0, or EW = wo, so the optimal value Wo is the mean of the limiting distribution of the sequence of iterates {W n}' It can also be shown (Mukherjee, 1994) that EWWT = (f.L(1'2/2) 1m, yielding W = Op(yTi). This is consistent with our assumption (1) with k = 1/2. In the one-dimensional case (d = m = 1), we have EW = 0 and EW2 = ~f.L(1'2 if lE[Xn+!'1}'(WO Xn+!W i= O. Using these results, the fact that Z '" N(O, (1'2), EW = 0, the independence of Z and X, and assuming that EX B < 00, it is not difficult to compute the expressions - 3 _ lEW - f.L2(1'4E[X3'1}"'1}'(1 - f.LX2'1}'2)] , E[X2'1}'2 - f.LX4( '1}'4 + (1'2'1}"2) + f.L2 X6'1}'2( '1}'4 /3 + (1'2'1}"2)] and where E[X2'1}'2(1 - f.LX2'1}'2)2 + f.LX4'1}'4 + 3f.L2 X6'1}"2'1}'2] K(f.L) 18f.L2(1'2E[X3'1}"'1}'(1 - f.LX2'1}'2)2 K(f.L) E[X2'1}'2 - + f.L4 (1' 4x 7 '1}"3'1}'] ~f.LX4('1}14 + (1'2'1}"2(1 2 f.LX2'1}'2)2) + f.L2 X6'1}'6 - ~f.L3 XB('1}'B + 3(1'4'1}"4)], and '1}' and '1}" are evaluated at the argument wO X for'1}. From the above expressions, it is seen that if'1}O = 1/[1 + e- O ] and X has a symmetric distribution (say N(O, 1)), then EW 3 i= 0 and lEW 4 i= 3(EW2)2, implying that W is non-Gaussian in general. This result is consistent with that obtained by direct application of Foster's criterion (Mukherjee, 1994). 5 RECONCILING LINEARIZATION AND STOCHASTIC APPROXIMATION METHODS The results of stochastic approximation analysis give a Gaussian distribution for W in the limit as f.L ---. 0 (Bucklew,Kurtz,Sethares, 1993; Finnoff, 1993). However, our results establish that the Gaussian distribution result is not valid for small nonzero f.L in general. To reconcile these results, recall W = Op (yTi). Hence, if we consider Asymptomatics of Gradient-Based Neural Network Training Algorithms 341 only moments of the normalized quantity W/ fo (and neglect higher-order terms in Op(fo)), we obtain E(W / fo)3 = 0 and E(W / fo)4 = 3[E(W / fo)2j2, which suggests that the normalized quantity W/ fo is Gaussian in the limit of vanishing {L, a conclusion also reached from stochastic approximation analysis. In support of this theoretical indication that the conclusions of our analysis (based on linearization for small {L) might tally with those of stochastic approximation techniques for small values of {L, simulations were done on the simple onedimensional training case of the previous section for 8 cases: {L = 0.1,0.2,0.3,0.5, and (72 = 0.1,0.5 for each value of {L, with w O fixed at 3. For each of the 8 cases, either 5 or 10 runs were made, with lengths (for the given values of {L) of 810000, 500000, 300000, and 200000 respectively. Each run gave a pair of sequences {Wn } obtained by starting off at Wo = 0 and training the network independently twice. Each resulting sequence {Wn} was then downs amp led at a large enough rate that the true autocorrelation of the downsampled sequence was less than 0.05, followed by deleting the first 10% of the samples of this downsampled sequence, so as to remove any dependence on initial conditions that might persist. (Autocorrelation at lag unity for this Markov Chain was so high that when {L = 0.1, a decimation rate of 9000 was required.) This was done to ensure that the elements of the resulting downsampled sequences could be assumed independent for the various hypothesis tests that were to follow. (a) For each run of each case, the empirical distribution functions of the two downsampled sequences thus generated were compared by means of the Kolmogorov-Smirnov test (Bickel,Doksum, 1977) at level 0.95, with the null hypothesis being that both sequences had the same actual cumulative distribution function (assumed continuous). This test was passed with ease on all trials, thereby showing that a limiting distribution existed and was attained by such a training algorithm. (b) For each run of each case, a skewness test and a kurtosis test (Bickel,Doksum, 1977) for normality were done at level 0.95 to test for normality. The sequences generated failed both tests for the ({L, (7) pair (0.1,0.1) and passed them both for the pairs (0 .1,0.5), (0.3,0.1), (0 .5,0.1), and (0.5,0.5). For the pair (0.2,0.5), the skewness test was passed and the kurtosis test failed, and for the pairs (0.2,0.1) and (0.3,0.5), the skewness test was failed and the kurtosis test passed. (c) All trials cleared the Kolmogorov tests (Bickel,Doksum, 1977) for normality at level 0.95, both when the normal distribution was taken to have the sample mean and variance (c.)mputed on the downsampled sequence), and when the normal distribution function had the asymptotic values of mean (zero) and variance ({L(72 /2). Hence we may conclude: 1. The limiting distribution of {Wn } exists. 2. For small values of {L, the deviation from Gaussianness is so small that the Gaussian distribution may be taken as a good approximation to the limiting distribution. 342 Sayandev Mukherjee, Terrence L. Fine In other words, though stochastic approximation analysis states that W/ yfji is Gaussian only in the limit of vanishing f.J" our simulation shows that this is a good approximation for small values of f.J, as well. Acknowledgements The research reported here was partially supported by NSF Grant SBR-9413001. References Berger, Marc A. An Introduction to Probability and Stochastic Processes. SpringerVerlag, New York, 1993. Bickel, Peter, and Doksum, Kjell. Mathematical Statistics: Basic Ideas and Selected Topics. Holden-Day, San Francisco, 1977. Bucklew, J.A., Kurtz, T.G., and Sethares, W.A. "Weak Convergence and Local Stability Properties of Fixed Step Size Recursive Algorithms," IEEE Trans. Inform. Theory, vol. 39, pp. 966-978, 1993. Finnoff, W. "Diffusion Approximations for the Constant Learning Rate Backpropagation Algorithm and Resistence to Local Minima." In Giles, C.L., Hanson, S.J., and Cowan, J.D., editors, Advances in Neural Information Processing Systems 5. Morgan Kaufmann Publishers, San Mateo CA, 1993, p.459 ff. Kuan, C-M, and Hornik, K. "Convergence of Learning Algorithms with Constant Learning Rates," IEEE Trans. Neural Networks, vol. 2, pp. 484-488, 1991. Leen, T.K., and Moody, J .E. "Weight Space Probability Densities in Stochastic Learning: 1. Dynamics and Equilibria," Adv. in NIPS 5, Morgan Kaufmann Publishers, San Mateo CA, 1993, p.451 ff. Loeve, M. Probability Theory I, 4th ed. Springer-Verlag, New York, 1977. Mukherjee, Sayandev. Asymptotics of Gradient-based Neural Network Training Algorithms. M.S. thesis, Cornell University, Ithaca, NY, 1994. Orr, G.B., and Leen, T.K. "Probability densities in stochastic learning: II. Transients and Basin Hopping Times," Adv. in NIPS 5, Morgan Kaufmann Publishers, San Mateo CA, 1993, p.507 ff. Rumelhart, D.E., Hinton, G.E., and Williams, R.J. "Learning interval representations by error propagation." In D.E. Rumelhart and J.L. McClelland, editors, Parallel Distributed Processing, Ch. 8, MIT Press, Cambridge MA, 1985. Tweedie, R.L. "Criteria for Classifying General Markov Chains," Adv. Appl. Prob., vol. 8,737-771, 1976. White, H. "Some Asymptotic Results for Learning in Single Hidden-Layer Feedforward Network Models," J. Am. Stat. Assn., vol. 84, 1003-1013, 1989. PART IV REThWORCEMENTLEARMNG
931 |@word trial:2 version:2 norm:7 smirnov:1 suitably:1 seek:1 linearized:2 bn:3 simulation:2 thereby:2 n8:2 moment:7 initial:2 series:1 amp:1 cleared:1 comell:2 activation:4 yet:1 laii:1 remove:2 update:1 stationary:2 implying:1 selected:1 vanishing:2 lr:1 iterates:9 node:2 successive:1 sigmoidal:1 mathematical:1 direct:1 prove:1 lx2:4 combine:1 autocorrelation:2 behavior:2 examine:1 actual:1 considering:1 bounded:4 null:1 what:1 kind:1 skewness:3 developed:1 finding:1 transformation:1 nonrandom:1 grant:1 yn:4 kurtz:5 positive:7 engineering:2 understood:1 sbr:1 tends:2 limit:4 local:2 fluctuation:1 interpolation:1 path:1 lrm:3 might:2 twice:2 mateo:3 suggests:1 appl:1 ease:1 unique:3 recursive:1 definite:2 lf:1 x3:2 backpropagation:1 asymptotics:3 empirical:1 word:1 downsampled:5 convenience:1 close:1 operator:2 applying:2 williams:1 starting:1 independently:1 estimator:2 stability:1 sethares:5 limiting:15 suppose:2 exact:1 homogeneous:2 us:1 hypothesis:3 decimation:1 element:1 rumelhart:2 updating:3 mukherjee:9 lean:1 persist:1 electrical:2 adv:3 dynamic:1 trained:1 depend:1 easily:2 various:1 kolmogorov:3 ew2:2 neighborhood:1 lag:1 solve:1 say:2 statistic:1 gi:3 kuan:2 sequence:24 differentiable:3 eigenvalue:1 net:2 indication:1 kurtosis:3 product:1 j2:2 frobenius:1 convergence:3 stat:1 school:2 op:9 strong:1 involves:1 come:1 implies:2 stochastic:11 transient:1 hold:8 sufficiently:3 considered:1 normal:2 equilibrium:1 elog:1 claim:1 major:1 bickel:4 iw:1 establishes:2 hope:1 mit:1 gaussian:9 yix:1 rather:1 cornell:3 og:1 contrast:1 am:1 dependent:1 entire:1 holden:1 hidden:3 classification:1 development:1 chapman:1 lrd:1 piecewise:2 simplify:1 few:2 ve:1 stationarity:1 semidefinite:1 yielding:1 xb:1 chain:2 neglecting:1 tweedie:2 iv:1 taylor:1 euclidean:2 theoretical:1 giles:1 gn:1 zn:1 applicability:1 deviation:1 reported:1 aw:2 considerably:1 density:2 terrence:5 off:1 together:1 continuously:1 moody:3 again:3 thesis:1 hn:10 choose:1 derivative:2 li:3 orr:2 ad:2 analyze:1 reached:1 bayes:1 parallel:1 formed:1 ir:1 lew:2 variance:4 kaufmann:3 weak:1 doksum:4 fo:6 inform:1 ed:1 definition:1 pp:2 proof:1 adl:1 finnoff:5 proved:1 recall:1 bucklew:5 feed:1 higher:2 attained:1 supervised:1 x6:3 follow:1 day:1 leen:5 evaluated:1 ox:1 done:3 though:1 ei:1 propagation:1 mode:2 normalized:3 verify:1 true:1 hence:6 symmetric:1 nonzero:4 white:3 eg:4 ll:2 recurrence:1 criterion:3 mina:7 l1:2 fj:1 common:1 linking:1 onedimensional:2 cambridge:1 ai:2 had:2 l3:1 certain:1 verlag:1 inequality:2 integrable:1 seen:2 minimum:1 unrestricted:1 morgan:3 surely:1 signal:1 ii:2 multiple:1 a1:1 basic:4 multilayer:2 expectation:1 want:1 fine:5 interval:1 resistence:1 publisher:3 ithaca:3 exhibited:1 tend:1 cowan:1 ee:2 feedforward:4 enough:2 wn:5 easy:1 independence:1 gave:1 restrict:1 idea:1 expression:2 passed:4 wo:9 peter:1 york:2 aiw:1 generally:1 detailed:1 clear:1 band:1 mcclelland:1 assn:1 exist:1 nsf:1 discrete:3 shall:2 vol:4 diffusion:2 run:4 prob:1 jitter:1 fourth:1 almost:1 vn:1 lx4:3 layer:3 guaranteed:1 followed:1 existed:1 x2:3 loeve:2 argument:1 min:7 lld:1 unity:1 lp:1 tw:1 leg:6 taken:2 equation:4 agree:1 turn:1 merit:1 junction:1 permit:1 upto:1 batch:3 customary:1 existence:6 original:1 reconciling:1 ensure:1 hopping:1 calculating:1 neglect:1 giving:1 prof:1 establish:4 quantity:2 dependence:2 gradient:9 supn:1 topic:1 anq:1 assuming:4 length:1 berger:3 balance:1 equivalently:1 difficult:1 ilw:1 observation:1 markov:9 finite:3 descent:4 hinton:1 thm:1 pair:5 required:1 hanson:1 nip:2 trans:2 pattern:1 max:5 deleting:1 unrealistic:1 ia:5 force:1 circumvent:1 normality:3 qualifying:1 woo:1 ziji:5 alii:1 literature:2 l2:10 acknowledgement:1 asymptotic:7 law:1 loss:1 lv:1 affine:1 sufficient:3 consistent:2 basin:1 foster:2 editor:2 lzn:1 classifying:1 supported:1 last:1 wide:1 distributed:1 dimension:1 xn:8 valid:1 cumulative:1 forward:1 made:1 san:4 dealing:1 investigating:1 assumed:4 conclude:1 francisco:1 continuous:2 iterative:1 ca:3 hornik:2 ean:1 expansion:1 marc:1 reconcile:1 ff:3 ny:3 fails:1 tally:1 ib:1 lw:2 iiaii:1 ix:1 theorem:2 down:1 showing:2 jensen:1 intractable:2 exists:3 linearization:4 led:1 infinitely:2 failed:3 partially:1 scalar:1 springer:1 ch:1 ma:1 goal:1 yti:2 springerverlag:1 infinite:1 uniformly:1 lemma:4 nil:1 la:1 ew:5 l4:1 support:1 ex:1
7,170
932
Instance-Based State Identification for Reinforcement Learning R. Andrew McCallum Department of Computer Science University of Rochester Rochester, NY 14627-0226 mccallumCcs.rochester.edu Abstract This paper presents instance-based state identification, an approach to reinforcement learning and hidden state that builds disambiguating amounts of short-term memory on-line, and also learns with an order of magnitude fewer training steps than several previous approaches. Inspired by a key similarity between learning with hidden state and learning in continuous geometrical spaces, this approach uses instance-based (or "memory-based") learning, a method that has worked well in continuous spaces. 1 BACKGROUND AND RELATED WORK When a robot's next course of action depends on information that is hidden from the sensors because of problems such as occlusion, restricted range, bounded field of view and limited attention, the robot suffers from hidden state. More formally, we say a reinforcement learning agent suffers from the hidden state problem if the agent's state representation is non-Markovian with respect to actions and utility. The hidden state problem arises as a case of perceptual aliasing: the mapping between states of the world and sensations of the agent is not one-to-one [Whitehead, 1992]. If the agent's perceptual system produces the same outputs for two world states in which different actions are required, and if the agent's state representation consists only of its percepts, then the agent will fail to choose correct actions. Note that even if an agent's state representation includes some internal state beyond its 378 R. Andrew McCallum immediate percepts, the agent can still suffer from hidden state if it does not keep enough internal state to uncover the non-Markovian-ness of its environment. One solution to the hidden state problem is simply to avoid passing through the aliased states. This is the approach taken in Whitehead's Lion algorithm [Whitehead, 1992]. Whenever the agent finds a state that delivers inconsistent reward, it sets that state's utility so low that the policy will never visit it again. The success of this algorithm depends on a deterministic world and on the existence of a path to the goal that consists of only unaliased states. Other solutions do not avoid aliased states, but do as best they can given a nonMarkovian state representation [Littman, 1994; Singh et al., 1994; Jaakkola et al., 1995]. They involve either learning deterministic policies that execute incorrect actions in some aliased states, or learning stochastic policies with action choice probabilities matching the proportions of the different underlying aliased world states. These approaches do not depend on a path of unaliased states, but they have other limitations: when faced with many aliased states, a stochastic policy degenerates into random walk; when faced with potentially harmful results from incorrect actions, deterministically incorrect or probabilistically incorrect action choice may prove too dangerous; and when faced with performance-critical tasks, inefficiency that is proportional to the amount of aliasing may be unacceptable. The most robust solution to the hidden state problem is to augment the agent's state representation on-line so as to disambiguate the aliased states. State identification techniques uncover the hidden state information-that is, they make the agent's internal state space Markovian. This transformation from an imperfect state information model to a perfect state information model has been formalized in the decision and control literature, and involves adding previous percepts and actions to the definition of agent internal state [Bertsekas and Shreve, 1978]. By augmenting the agent's perception with history information-.short-term memory of past percepts, actions and rewards-the agent can distinguish perceptually aliased states, and can then reliably choose correct actions from them. Predefined, fixed memory representations such as order n Markov models (also known as constant-sized perception windows, linear traces or tapped-delay lines) are often undesirable. When the length of the window is more than needed, they exponentially increase the number of internal states for which a policy must be stored and learned; when the length of the memory is less than needed, the agent reverts to the disadvantages of undistinguished hidden state. Even if the agent designer understands the task well enough to know its maximal memory requirements, the agent is at a disadvantage with constant-sized windows because, for most tasks, different amounts of memory are needed at different steps of the task. The on-line memory creation approach has been adopted in several reinforcement learning algorithms. The Perceptual Distinctions Approach [Chrisman, 1992] and Utile Distinction Memory [McCallum, 1993] are both based on splitting states of a finite state machine by doing off-line analysis of statistics gathered over many steps. Recurrent-Q [Lin, 1993] is based on training recurrent neural networks. Indexed Memory [Teller, 1994] uses genetic programming to evolve agents that use load and store instructions on a register bank. A chief disadvantage of all these techniques is that they require a very large number of steps for training. Instance-Based State Identification for Reinforcement Learning 2 379 INSTANCE-BASED STATE IDENTIFICATION This paper advocates an alternate solution to the hidden state problem we term instance-based state identification. The approach was inspired by the successes of instance-based (also called "memory-based") methods for learning in continuous perception spaces, (i.e. [Atkeson, 1992; Moore, 1992]). The application of instance-based learning to short-term memory for hidden state is driven by the important insight that learning in continuous spaces and learning with hidden state have a crucial feature in common: they both begin learning without knowing the final granularity of the agent's state space. The former learns which regions of continuous input space can be represented uniformly and which areas must be finely divided among many states. The later learns which percepts can be represented uniformly because they uniquely identify a course of action without the need for memory, and which percepts must be divided among many states each with their own detailed history to distinguish them from other perceptually aliased world states. The first approach works with a continuous geometrical input space, the second works with a percept-act ion-reward "sequence" space, (or "history" space). Large continuous regions correspond to less-specified, small memories; small continuous regions correspond to more-specified, large memories. Furthermore, learning in continuous spaces and sequence spaces both have a lot to gain from instance-based methods. In situations where the state space granularity is unknown, it is especially useful to memorize the raw previous experiences. If the agent tries to fit experience to its current, flawed state space granularity, it is bound to lose information by attributing experience to the wrong states. Experience attributed to the wrong state turns to garbage and is wasted. When faced with an evolving state space, keeping raw previous experience is the path of least commitment, and thus the most cautious about losing information. 3 NEAREST SEQUENCE MEMORY There are many possible instance-based techniques to choose from, but we wanted to keep the first application simple. With that in mind, this initial algorithm is based on k-nearest neighbor. We call it Nearest Sequence Memory, (NSM). It bears emphasizing that this algorithm is the most straightforward, simple, almost naive combination of instance-based methods and history sequences that one could think of; there are still more sophisticated instance-based methods to try. The surprising result is that such a simple technique works as well as it does. Any application of k-nearest neighbor consists of three parts: 1) recording each experience, 2) using some distance metric to find neighbors of the current query point, and 3) extracting output values from those neighbors. We apply these three parts to action-percept-reward sequences and reinforcement learning by Q-Iearning lWatkins, 1989] as follows: 1. For each step the agent makes in the world, it records the action, percept and reward by adding a new state to a single, long chain of states. Thus, each state in the chain contains a snapshot of immediate experience; and all the experiences are laid out in a time-connected history chain. 380 R. Andrew McCallum Learning in a Geometric Space k-nearest neighbor, k = 3 ? ? ? ? ? Learning in a Sequence Space k-nearest neighbor, k =3 00014001301201 ? ? match length 3~ 0 1 " action. percept. rewanl Figure 1: A continuous space compared with a sequence space. In each case, the "query point" is indicated with a gray cross, and the three nearest neighbors are indicated with gray shadows. In a geometric space, the neighborhood metric is defined by Euclidean distance. In a sequence space, the neighborhood metric is determined by sequence match length-the number of preceding states that match the states preceding the query point. 2. When the agent is about to choose an action, it finds states considered to be similar by looking in its state chain for states with histories similar to the current situation. The longer a state's string of previous experiences matches the agent's most recent experiences, the more likely the state represents where the agent is now. 3. Using the states, the agent obtains Q-values by averaging together the expected future reward values associated with the k nearest states for each action. The agent then chooses the action with the highest Q-value. The regular Q-Iearning update rule is used to update the k states that voted for the chosen action. Choosing to represent short-term memory as a linear trace is a simple, wellestablished technique. Nearest Sequence Memory uses a linear trace to represent memory, but it differs from the fixed-sized window approaches because it provides a variable memory-length-like k-nearest neighbor, NSM can represent varying resolution in different regions of state space. 4 DETAILS OF THE ALGORITHM A more complete description of Nearest Sequence Memory, its performance and its possible improvements can be found in [McCallum, 1995]. The interaction between the agent and its environment is described by actions, percepts and rewards. There is a finite set of possible actions, A {al,a2, ... ,am }, = Instance-Based State Identification for Reinforcement Learning 381 = a finite set of possible percepts, () {Ol, 02, ... , On}, and scalar range of possible rewards, = [x, y], x, Y E~. At each time step, t, the agent executes an action, at E A, then as a result receives a new percept, Ot E (), and a reward, rt E The agent records its raw experience at time t in a "state" data point, St. Also associated with St is a slot to hold a single expected future discounted reward value, denoted q(st). This value is associated with at and no other action. n n. 1. Find the k nearest neighbor (most similar) states for each possible future action. The state currently at the end of the chain is the "query point" from which we measure all the distances. The neighborhood metric is defined by the number of preceding experience records that match the experience records preceding the "query point" state. (Here higher values of n(s;, sJ) indicate that S; and Sj are closer neighbors.) if (a;-1 = aj-I) A (0;-1 otherwise _ _)_ { 1+n(s;_1,Sj-I), n ( S" SJ 0, Considering each of the possible future actions neighbors and give them a vote, v(s;). v(S;) ={ ~: ill = OJ-I) A (r;-1 = rj_I) (1) turn, we find the k nearest if n(st, s;) is among the k maxv$jlaj=a; n(st, Sj)'s otherwise (2) 2. Determine the Q-value for each action by averaging individual the q-values from the k voting states for that action. L Qt(a;) = (v(s;)/k)q(sj) (3) V$jlaj=a; 3. Select an action by maximum Q-value, or by random exploration. According to an exploration probability, e, either let at+1 be randomly chosen from A, or (4) 4. Execute the action chosen in step 3, and record the resulting experience. Do this by creating a new "state" representing the current state of the environment, and storing the action-percept-reward triple associated with it: Increment the time counter: t ~ t + 1. Create St; record in it at, Ot, rt. The agent can limit its storage and computational load by limiting the number of instances it maintains to N (where N is some reasonably large number) . Once the agent accumulates N instances, it can discard the oldest instance each time it adds a new one. This also provides a way to handle a changing environment. 5. Update the q-values by vote. Perform the dynamic programming step using the standard Q-Iearning rule to update those states that voted for the chosen action. Note that this actually involves performing steps 1 and 2 to get the next Q-values needed for calculating the utility of the agent's current state, Ut . (Here (3 is the learning rate.) Ut = (Vsda; = at-I) q(s;) ~ max Qt(a) (5) a (1- (3v(s;))q(s;) + (3v(s;)(r; + "YUt) (6) 382 R. Andrew McCallum Performance during learning 20 Nearest Sequence Memory Perceptual Distinctions Approach - Nearest Sequence Memory . Recurrent-a - . 60 ~ 50 15 .,c. li ~ 5 Steps per Trial during learning 70 10 40 30 5 20 1000 2000 3000 lD ~: t fBI 74 4000 5000 Steps 6000 7000 10 8000 40 ~: ...._._' 60 100 80 StepslOLMm 2500 -, StejM 10 L.", 238 _ o f ...... ~:; 153 20 Trials StejM 10 LMm 1500 0 Ii?iii :: ....borof ...... I0_~. ,."I0 0 =L-.'-...:39==5=-_ _ _Steps __ .....borof ...... Figure 2: Comparing Nearest Sequence Memory with three other algorithms: Perceptual Distinction Approach, Recurrent-Q and Utile Distinction Memory. In each case, NSM learns with roughly an order of magnitude fewer steps. 5 EXPERIMENTAL RESULTS The performance of NSM is compared to three other algorithms using the tasks chosen by the other algorithms' designers. In each case, NSM learns the task with roughly an order of magnitude fewer steps. Although NSM learns good policies quickly, it does not always learn optimal policies. In section 6 we will discuss why the policies are not always optimal and how NSM could be improved. The Perceptual Distinctions Approach [Chrisman, 1992] was demonstrated in a space ship docking application with hidden state. The task was made difficult by noisy sensors and actions. Some of the sensors returned incorrect values 30% of the time. Various actions failed 70, 30 or 20% of the time, and when they failed, resulted in random states. NSM used f3 = 0.2, I = 0.9, k = 8, and N = 1000. PDA takes almost 8000 steps to learn the task. NSM learns a good policy in less than 1000 steps, although the policy is not quite optimal. Utile Distinction Memory [McCallum, 1993] was demonstrated on several local perception mazes. Unlike most reinforcement learning maze domains, the agent perceives only four bits indicating whether there is a barrier to the immediately adjacent north, east, south and west. NSM used f3 = 0.9, I = 0.9, k = 4, and N = 1000. In two of the mazes, NSM learns the task in only about 1/20th the time required by UDM; in the other two, NSM learns mazes that UDM did not solve at all . Instance-Based State Identification for Reinforcement Learning 383 Recurrent-Q [Lin, 1993] was demonstrated on a robot 2-cup retrieval task. The env,jronment is deterministic, but the task is made difficult by two nested levels of hidden state and by providing no reward until the task is completely finished. NSM used {3 = 0.9, I = 0.9, k = 4, and N = 1000. NSM learns good performance in about 15 trials , Recurrent-Q takes about 100 trials to reach equivalent performance. 6 DISCUSSION Nearest Sequence Memory offers much improved on-line performance and fewer training steps than its predecessors. Why is the improvement so dramatic? I believe the chief reason lies with the inherent advantage of instance-based methods, as described in section 2: the key idea behind Instance-Based State Identification is the recognition that recording raw experience is particularly advantageous when the agent is learning a policy over a changing state space granularity, as is the case when the agent is building short-term memory for disambiguating hidden state. If, instead of using an instance-based technique, the agent simply averages new experiences into its current, flawed state space model, the experiences will be applied to the wrong states, and cannot be reused when the agent reconfigures its state space. Furthermore, and perhaps even more detrimentally, incoming data is always interpreted in the context of the flawed state space, always biased in an inappropriate way-not simply recorded, kept uncommitted and open to easy reinterpretation in light of future data. The experimental results in this paper bode well for instance-based state identification. Nearest Sequence Memory is simple-if such a simplistic implementation works as well as it does, more sophisticated approaches may work even better. Here are some ideas for improvement: The agent should use a more sophisticated neighborhood distance metric than exact string match length. A new metric could account for distances between different percepts instead of considering only exact matches. A new metric could also handle continuous-valued inputs. Nearest Sequence Memory demonstrably solves tasks that involve noisy sensation and action, but it could perhaps handle noise even better if it used some technique for explicitly separating noise from structure. K-nearest neighbor does not explicitly discriminate between structure and noise. If the current query point has neighbors with wildly varying output values, there is no way to know if the variations are due to noise , (in which case they should all be averaged), or due to fine-grained structure of. the underlying function (in which case only the few closest should be averaged). Because NSM is built on k-nearest neighbor, it suffers from the same inability to methodically separate history differences that are significant for predicting reward and history differences that are not. I believe this is the single most important reason that NSM sometimes did not find optimal policies. Work in progress addresses the structure/noise issue by combining instance-based state identification with the structure/noise separation method from Utile Distinction Memory [McCallum, 1993]. The algorithm, called Utile Suffix Memory, uses a tree-structured representation, and is related to work with Ron, Singer and Tishby's Prediction Suffix Trees, Moore's Parti-game, Chapman and Kaelbling's 384 R. Andrew McCallum G-algorithm, and Moore's Variable Resolution Dynamic Programming. See [McCallum, 1994] for more details as well as references to this related work. Aclmowledgments This work has benefited from discussions with many colleagues, including: Dana Ballard, Andrew Moore, Jeff Schneider, and Jonas Karlsson. This material is based upon work supported by NSF under Grant no. IRI-8903582 and by NIH/PHS under Grant no. 1 R24 RR06853-02. References [Atkeson, 1992] Christopher G. Atkeson. Memory-based approaches to approximating continuous functions. In M. Casdagli and S. Eubank, editors, Nonlinear Modeling and Forecasting, pages 503-521. Addison Wesley, 1992. [Bertsekas and Shreve, 1978] Dimitri. P. Bertsekas and Steven E. Shreve. Stochastic Optimal Control. Academic Press, 1978. [Chrisman, 1992] Lonnie Chrisman. Reinforcement learning with perceptual aliasing: The perceptual distinctions approach. In Tenth Nat'l Conf. on AI, 1992. [Jaakkola et al., 1995] Tommi Jaakkola, Satinder Pal Singh, and Michael 1. Jordan. Reinforcement learning algorithm for partially observable markov decision problems. In Advances of Neural Information Processing Systems 7. Morgan Kaufmann, 1995. [Lin, 1993] Long-Ji Lin. Reinforcement Learning for Robots Using Neural Networks. PhD thesis, Carnegie Mellon, School of Computer Science, January 1993. [Littman, 1994] Michael Littman. Memoryless policies: Theoretical limitations and practical results. In Proceedings of the Third International Conference on Simulation of Adaptive Behavior: From Animals to Animats, 1994. [McCallum, 1993] R. Andrew McCallum. Overcoming incomplete perception with utile distinction memory. In The Proceedings of the Tenth International Machine Learning Conference. Morgan Kaufmann Publishers, Inc., 1993. [McCallum, 1994] R. Andrew McCallum. Utile suffix memory for reinforcement learning with hidden state. TR 549, U. of Rochester, Computer Science, 1994. [McCallum, 1995] R. Andrew McCallum. Hidden state and reinforcement learning with instance-based state identification. IEEE Trans. on Systems, Man, and Cybernetics, 1995. (In press) [Earlier version available as U. of Rochester TR 502]. [Moore, 1992] Andrew Moore. Efficient Memory-based Learning for Robot Control. PhD thesis, University of Cambridge, November 1992. [Singh et al., 1994] Satinder Pal Singh, Tommi Jaakkola, and Michael 1. Jordan. Modelfree reinforcement learning for non-markovian decision problems. In The Proceedings of the Eleventh International Machine Learning Conference, 1994. [Teller, 1994] Astro Teller. The evolution of mental models. In Kim Kinnear, editor, Advances in Genetic Programming, chapter 9. MIT Press, 1994. [Watkins, 1989] Chris Watkins. Learning from delayed rewards. PhD thesis, Cambridge University, 1989. [Whitehead, 1992] Steven Whitehead. Reinforcement Learning for the Adaptive Control of Perception and Action. PhD thesis, Department of Computer Science, University of Rochester, 1992.
932 |@word trial:4 version:1 proportion:1 advantageous:1 casdagli:1 reused:1 open:1 instruction:1 simulation:1 dramatic:1 tr:2 ld:1 initial:1 inefficiency:1 contains:1 genetic:2 past:1 current:7 comparing:1 surprising:1 must:3 wanted:1 update:4 maxv:1 fewer:4 mccallum:16 oldest:1 short:5 record:6 utile:7 mental:1 provides:2 ron:1 unacceptable:1 predecessor:1 jonas:1 incorrect:5 consists:3 prove:1 advocate:1 eleventh:1 expected:2 behavior:1 roughly:2 aliasing:3 ol:1 inspired:2 discounted:1 window:4 considering:2 inappropriate:1 perceives:1 begin:1 bounded:1 underlying:2 aliased:8 interpreted:1 string:2 transformation:1 act:1 voting:1 iearning:3 wrong:3 control:4 grant:2 bertsekas:3 local:1 limit:1 accumulates:1 path:3 limited:1 range:2 averaged:2 practical:1 differs:1 area:1 evolving:1 matching:1 regular:1 get:1 cannot:1 undesirable:1 storage:1 context:1 equivalent:1 deterministic:3 demonstrated:3 straightforward:1 attention:1 iri:1 resolution:2 formalized:1 splitting:1 immediately:1 insight:1 rule:2 parti:1 handle:3 variation:1 increment:1 limiting:1 exact:2 programming:4 losing:1 us:4 tapped:1 recognition:1 particularly:1 steven:2 region:4 connected:1 counter:1 highest:1 environment:4 pda:1 reward:14 littman:3 dynamic:2 singh:4 depend:1 reinterpretation:1 creation:1 upon:1 completely:1 represented:2 various:1 chapter:1 query:6 neighborhood:4 choosing:1 quite:1 solve:1 valued:1 say:1 otherwise:2 statistic:1 think:1 noisy:2 final:1 sequence:18 advantage:1 interaction:1 maximal:1 commitment:1 combining:1 degenerate:1 description:1 cautious:1 requirement:1 produce:1 perfect:1 andrew:10 recurrent:6 augmenting:1 nearest:21 school:1 qt:2 progress:1 solves:1 involves:2 memorize:1 shadow:1 indicate:1 tommi:2 sensation:2 correct:2 stochastic:3 exploration:2 material:1 require:1 undistinguished:1 hold:1 considered:1 mapping:1 a2:1 lose:1 currently:1 create:1 mit:1 sensor:3 always:4 avoid:2 varying:2 jaakkola:4 probabilistically:1 improvement:3 nonmarkovian:1 kim:1 am:1 suffix:3 i0:1 hidden:19 issue:1 among:3 ill:1 augment:1 denoted:1 animal:1 ness:1 field:1 once:1 never:1 f3:2 flawed:3 env:1 chapman:1 represents:1 future:5 inherent:1 few:1 randomly:1 nsm:16 resulted:1 individual:1 delayed:1 occlusion:1 karlsson:1 light:1 behind:1 chain:5 predefined:1 closer:1 experience:17 indexed:1 tree:2 harmful:1 euclidean:1 walk:1 incomplete:1 theoretical:1 instance:23 modeling:1 earlier:1 markovian:4 disadvantage:3 kaelbling:1 delay:1 too:1 tishby:1 pal:2 stored:1 chooses:1 st:6 international:3 off:1 michael:3 together:1 quickly:1 again:1 yut:1 recorded:1 thesis:4 choose:4 conf:1 creating:1 dimitri:1 li:1 account:1 includes:1 north:1 inc:1 register:1 explicitly:2 depends:2 later:1 view:1 lot:1 try:2 doing:1 maintains:1 rochester:6 voted:2 kaufmann:2 percept:15 gathered:1 identify:1 correspond:2 identification:12 raw:4 eubank:1 cybernetics:1 executes:1 history:8 reach:1 suffers:3 whenever:1 definition:1 colleague:1 associated:4 attributed:1 gain:1 ut:2 uncover:2 sophisticated:3 actually:1 understands:1 wesley:1 higher:1 improved:2 execute:2 wildly:1 furthermore:2 shreve:3 until:1 receives:1 christopher:1 nonlinear:1 aj:1 gray:2 perhaps:2 indicated:2 believe:2 building:1 evolution:1 former:1 memoryless:1 moore:6 adjacent:1 during:2 game:1 uniquely:1 modelfree:1 complete:1 delivers:1 geometrical:2 nih:1 common:1 ji:1 exponentially:1 significant:1 mellon:1 cup:1 cambridge:2 ai:1 robot:5 similarity:1 longer:1 add:1 closest:1 own:1 recent:1 driven:1 discard:1 ship:1 store:1 success:2 morgan:2 preceding:4 schneider:1 determine:1 lmm:1 ii:1 match:7 academic:1 cross:1 long:2 lin:4 retrieval:1 divided:2 offer:1 visit:1 prediction:1 simplistic:1 metric:7 represent:3 sometimes:1 ion:1 background:1 fine:1 crucial:1 publisher:1 ot:2 biased:1 finely:1 unlike:1 south:1 recording:2 inconsistent:1 jordan:2 call:1 extracting:1 granularity:4 iii:1 enough:2 easy:1 fit:1 lonnie:1 imperfect:1 idea:2 knowing:1 whether:1 utility:3 forecasting:1 suffer:1 returned:1 passing:1 action:35 garbage:1 useful:1 detailed:1 involve:2 amount:3 uncommitted:1 ph:1 demonstrably:1 wellestablished:1 nsf:1 designer:2 per:1 carnegie:1 key:2 four:1 changing:2 tenth:2 kept:1 wasted:1 almost:2 laid:1 separation:1 decision:3 bit:1 bound:1 distinguish:2 r24:1 dangerous:1 worked:1 performing:1 department:2 structured:1 according:1 alternate:1 combination:1 bode:1 restricted:1 taken:1 turn:2 discus:1 fail:1 needed:4 know:2 mind:1 singer:1 addison:1 end:1 whitehead:5 adopted:1 available:1 apply:1 fbi:1 existence:1 calculating:1 build:1 especially:1 approximating:1 rt:2 distance:5 separate:1 separating:1 chris:1 astro:1 reason:2 length:6 providing:1 difficult:2 potentially:1 trace:3 implementation:1 reliably:1 policy:13 unknown:1 perform:1 animats:1 snapshot:1 markov:2 finite:3 november:1 january:1 immediate:2 situation:2 looking:1 overcoming:1 required:2 specified:2 learned:1 distinction:10 chrisman:4 trans:1 address:1 beyond:1 lion:1 perception:6 reverts:1 built:1 oj:1 memory:37 max:1 including:1 critical:1 reconfigures:1 predicting:1 representing:1 finished:1 naive:1 faced:4 docking:1 literature:1 teller:3 geometric:2 evolve:1 bear:1 limitation:2 proportional:1 dana:1 triple:1 agent:38 editor:2 bank:1 storing:1 course:2 supported:1 keeping:1 detrimentally:1 neighbor:14 barrier:1 world:6 maze:4 made:2 reinforcement:16 adaptive:2 atkeson:3 sj:6 obtains:1 observable:1 keep:2 satinder:2 incoming:1 continuous:12 chief:2 why:2 disambiguate:1 learn:2 reasonably:1 robust:1 ballard:1 domain:1 did:2 noise:6 benefited:1 west:1 ny:1 deterministically:1 lie:1 perceptual:8 watkins:2 third:1 learns:10 grained:1 emphasizing:1 load:2 adding:2 phd:4 magnitude:3 nat:1 perceptually:2 attributing:1 simply:3 likely:1 failed:2 partially:1 scalar:1 nested:1 slot:1 goal:1 sized:3 disambiguating:2 jeff:1 man:1 determined:1 uniformly:2 averaging:2 called:2 discriminate:1 experimental:2 vote:2 east:1 indicating:1 formally:1 select:1 internal:5 arises:1 inability:1
7,171
933
Learning with Preknowledge: Clustering with Point and Graph Matching Distance Measures Steven Gold!, Anand Rangarajan 1 and Eric Mjolsness 2 Department of Computer Science Yale University New Haven, CT 06520-8285 Abstract Prior constraints are imposed upon a learning problem in the form of distance measures. Prototypical 2-D point sets and graphs are learned by clustering with point matching and graph matching distance measures. The point matching distance measure is approx. invariant under affine transformations - translation, rotation, scale and shear - and permutations. It operates between noisy images with missing and spurious points. The graph matching distance measure operates on weighted graphs and is invariant under permutations. Learning is formulated as an optimization problem . Large objectives so formulated ('" million variables) are efficiently minimized using a combination of optimization techniques - algebraic transformations, iterative projective scaling, clocked objectives, and deterministic annealing. 1 Introduction While few biologists today would subscribe to Locke's description of the nascent mind as a tabula rasa, the nature of the inherent constraints - Kant's preknowl1 E-mail address of authors: lastname-firstname@cs.yale.edu 2Department of Computer Science and Engineering, University of California at San Diego (UCSD), La Jolla, CA 92093-0114. E-mail: emj@cs.ucsd.edu 714 Steven Gold, Anand Rangarajan, Eric Mjolsness edge - that helps organize our perceptions remains much in doubt. Recently, the importance of such preknowledge for learning has been convincingly argued from a statistical framework [Geman et al., 1992]. Researchers have proposed that our brains may incorporate preknowledge in the form of distance measures [Shepard, 1989]. The neural network community has begun to explore this idea via tangent distance [Simard et al., 1993], model learning [Williams et al., 1993] and point matching distances [Gold et al., 1994]. However, only the point matching distances have been invariant under permutations. Here we extend that work by enhancing both the scope and function of those distance measures, significantly expanding the problem domains where learning may take place. We learn objects consisting of noisy 2-D point-sets or noisy weighted graphs by clustering with point matching and graph matching distance measures. The point matching measure is approx. invariant under permutations and affine transformations (separately decomposed into translation, rotation, scale and shear) and operates on point-sets with missing or spurious points. The graph matching measure is invariant under permutations. These distance measures and others like them may be constructed using Bayesian inference on a probabilistic model of the visual domain. Such models introduce a carefully designed bias into our learning, which reduces its generality outside the problem domain but increases its ability to generalize within the problem domain. (From a statistical viewpoint, outside the problem domain it increases bias while within the problem domain it decreases variance). The resulting distance measures are similar to some of those hypothesized for cognition. The distance measures and learning problem (clustering) are formulated as objective functions. Fast minimization of these objectives is achieved by a combination of optimization techniques - algebraic transformations, iterative projective scaling, clocked objectives, and deterministic annealing. Combining these techniques significantly increases the size of problems which may be solved with recurrent network architectures [Rangarajan et al., 1994]. Even on single-cpu workstations non-linear objectives with a million variables can routinely be minimized. With these methods we learn prototypical examples of 2-D points set and graphs from randomly generated experimental data. 2 2.1 Distance Measures in Unsupervised Learning An Affine Invariant Point Matching Distance Measure The first distance measure quantifies the degree of dissimilarity between two unlabeled 2-D point images, irrespective of bounded affine transformations, i.e. differences in position, orientation, scale and shear. The two images may have different numbers of points. The measure is calculated with an objective that can be used to find correspondence and pose for unlabeled feature matching in vision. Given two sets of points {Xj} and {Yk}, one can minimize the following objective to find the affine transformation and permutation which best maps Y onto X : J Epm(m, t,A) K = L: L: mjkllXj j=lk=l with constraints: Yj Ef=l mjk ~ 1 , J t - A? Ykll 2 + g(A) - a K L: L: mjk j=lk=l Yk Ef=l mjk ~ 1 , Yjk mjk 2:: O. Learning with Preknowledge 715 A is decomposed into scale, rotation, vertical shear and oblique shear components. g(A) regularizes our affine transformation - bounding the scale and shear compo- nents. m is a fuzzy correspondence matrix which matches points in one image with corresponding points in the other image. The inequality constraint on m allows for null matches - that is a given point in one image may match to no corresponding point in the other image. The a term biases the objective towards matches. Then given two sets of points {Xj} and {Yk} the distance between them is defined as: D({Xj}, {Yk}) = min (Epm(m,t, A) I constraints on m) m,t,A This measure is an example of a more general image distance measure derived in [Mjolsness, 1992]: d(z, y) = mind(z, T(y? E [0, (0) T where T is a set of transformation parameters introduced by a visual grammar. Using slack variables, and following the treatment in [Peterson and Soderberg, 1989; Yuille and Kosowsky, 1994] we employ Lagrange multipliers and an z logz barrier function to enforce the constraints with the following objective: J Epm(m, t, A) J K = L: L: mjkllXj - t - A? Ykll 2 + g(A) - a j=lk=l L: L: mjk j=lk=1 1 J+1 K+1 +-p L: L: mjk(logmjk - K 1) + j=1 k=l J K+1 j=l k=l L: J.'j (L: mjk - 1) + K J+1 k=1 j=l L: lIk(L: mjk - 1) (1) In this objective we are looking for a saddle point. (1) is minimized with respect to m, t, and A which are the correspondence matrix, translation, and affine transform, and is maximized with respect to l' and 11, the Lagrange multipliers that enforce the row and column constraints for m. The above can be used to define many different distance measures, since given the decomposition of A it is trivial to construct measures which are invariant only under some subset of the transformations (such as rotation and translation) . The regularization and a terms may also be individually adjusted in an appropriate fashion for a specific problem domain. 2.2 Weighted Graph Matching Distance Measures The following distance measure quantifies the degree of dissimilarity between two unlabeled weighted graphs. Given two graphs, represented by adjacency matrices Gab and gij, one can minimize the objective below to find the permutation which best maps G onto g: ' Egm(m) A I B a=l i=l b=1 = L: L:(L: Gabmbi - J L: m jgji)2 a ;=1 Steven Gold, Anand Rangarajan, Eric Mjolsness 716 with constraints: 'Va 2::=1 mai = 1 , 'Vi 2::=1 mai = 1 , 'Vai mai ;::: O. These constraints are enforced in the same fashion as in (1). An algebraic fixed-point transformation and self-amplification term further transform the objective to: A I z: 1 J B Egm(m) = L L(J.'ai(L Gabmbi majDji) a=1 i=1 b=1 j=1 z: lAI +(j L a=1 i=1 A 2J.'~i - ,lTaimai z: I I + ~lT~i) A mai(logmai - 1) + L lI:a(Z: mai - 1) + Ai(L mai - 1) a=1 i=1 i=1 a=1 (2) In this objective we are also looking for a saddle point. A second, functionally equivalent, graph matching objective is also used clustering problem: A B I III J Egm/(m) = LZ:LZ:mai m bj(Gab-Dji)2 a=lb=li=lj=l with constraints: 'Va 2::=1 mai 2.3 =1 , the (3) 'Vi 2::=1 mai = 1 , 'Vai mai ;::: O. The Clustering Objective The learning problem is formulated as follows: Given a set of I objects, {Xi} find a set of A cluster centers {Ya } and match variables {Mia} defined as M. _ {I 0 la - if Xi is in Ya's cluster otherwise, such that each object is in only one cluster, and the total distance of all the objects from their respective cluster centers is minimized. To find {Ya} and {Mia} minimize the cost function, I A Eeltuter(Y,M) = LLMiaD(Xi' Ya) i=l a=l with the constraint that 'Vi 2:a Mia = 1 , 'Vai Mai ;::: O. D(Xi, Y a), the distance function, is a measure of dissimilarity between two objects. The constraints on M are enforced in a manner similar to that described for the distance measure, except that now only the rows of the matrix M need to add to one, instead of both the rows and the columns. I Eeituter(Y, M) 1 I A Z:Z:MiaD(Xi , Ya) i=l a=1 I + (j Z:Z: Mia(log Mia - 1) i=l a=1 A + Z:Ai(LMia i=l A a=1 1) (4) Learning with Pre knowledge 717 Here, the objects are point-sets or weighted graphs. If point-sets the distance measure D(Xi, Ya) is replaced by (1), if graphs it is replaced by (2) or (3). Therefore, given a set of objects, X, we construct Ecltuter and upon finding the appropriate saddle point of that objective, we will have Y, their cluster centers, and M, their cluster memberships. 3 The Algorithm The algorithm to minimize the clustering objectives consists of two loops - an inner loop to minimize the distance measure objective [either (1) or (2)] and an outer loop to minimize the clustering objective (4). Using coordinate descent in the outer loop results in dynamics similar to the EM algorithm [Jordan and Jacobs, 1994] for clustering. All variables occurring in the distance measure objective are held fixed during this phase. The inner loop uses coordinate ascent/descent which results in repeated row and column projections for m. The minimization of m, and the distance measure variables [either t, A of (1) or 1', (f of (2)], occurs in an incremental fashion, that is their values are saved after each inner loop call from within the outer loop and are then used as initial values for the next call to the inner loop. This tracking of the values of the distance measure variables in the inner loop is essential to the efficiency of the algorithm since it greatly speeds up each inner loop optimization. Most coordinate ascent/descent phases are computed analytically, further speeding up the algorithm. Some local minima are avoided, by deterministic annealing in both the outer and inner loops. The mUlti-phase dynamics maybe described as a clocked objective. Let {D} be the set of distance measure variables excluding m. The algorithm is as follows: Initialize {D} to the equivalent of an identity transform, Y to random values Begin Outer Loop Begin Inner Loop Initialize {D} with previous values Find m, {D} for each ia pair: Find m by softmax, projecting across j, then k, iteratively Find {D} by coordinate descent End Inner Loop Find M ,Y using fixed values of m, {D}, determined in inner loop: Find M by softmax, across i Find Y by coordinate descent Increase f3M, f3m End Outer Loop When analytic solutions are computed for Y the outer loop takes a form similar to fuzzy ISODATA clustering, with annealing on the fuzziness parameter. 4 Methods and Experimental Results Four series of experiments were ran with randomly generated data to evaluate the learning algorithms. Point sets were clustered in the first three experiments and weighted graphs were clustered in the fourth. In each experiment a set of object 718 Steven Gold, Anand Rangarajan, Eric Mjolsness models were randomly generated. Then from each object model a set of object instances were created by transforming the object model according to the problem domain assumed for that experiment. For example, an object represented by points in two dimensional space was translated, rotated, scaled, sheared, and permuted to form a new point set. A object represented by a weighted graph was permuted. Noise was added to further distort the object. Parts of the object were deleted and spurious features (points) were added. In this manner, from a set of object models, a larger number of object instances were created. Then with no knowledge of the original objects models or cluster memberships, we clustered the object instances using the algorithms described above. The results were evaluated by comparing the object prototypes (cluster centers) formed by each experimental run to the object models used to generate the object instances for that experiment. The distance measures used in the clustering were used for this comparison, i.e. to calculate the distance between the learned prototype and the original object. Note that this distance measure also incorporates the transformations used to create the object instances. The mean and standard deviations of these distances were plotted (Figure 1) over hundreds of experiments, varying the object instance generation noise. The straight line appearing on each graph displays the effect of the noise only. It is the expected object model-object prototype distance if no transformations were applied, no features were deleted or added, and the cluster memberships of the object instances were known. It serves as an absolute lower bound on our learning algorithm. The noise was increased in each series of experiments until the curve flattened - that is the object instances became so distorted by noise that no information about the original objects could be recovered by the algorithm. In the first series of experiments (Figure 1a), point set objects were translated, rotated, scaled, and permuted. Initial object models were created by selecting points with a uniform distribution within a unit square. The transformations to create the object instance were selected with a uniform distribution within the following bounds; translation: ?.5, rotation: ?27?, log(scale): ? log(.5). 100 object instances were generated from 10 object models. All objects contained 20 points.The standard deviation of the Gaussian noise was varied by .02 from .02 to .16. 15 experiments were run at each noise level. The data point at each error bar represents 150 distances (15 experiments times 10 model-prototype distances for each experiment). In the second and third series of experiments (Figures 1b and 1c), point set objects were translated, rotated, scaled, sheared (obliquely and vertically), and permuted. Each object point had a 10% probability of being deleted and a 5% probability of generating a spurious point. The point sets and transformations were randomly generated as in the first experiment, except for these bounds; log(scale): ? log(.7), log(verticalshear): ?log(.7), and log(obliqueshear): ?log(.7). In experiment 2, 64 object instances and 4 object models of 15 points each were used. In experiment 3, 256 object instances and 8 object models of 20 points each were used. Noise levels like experiment 1 were used, with 20 experiments run at each noise level in experiment 2 and 10 experiments run at each noise level in experiment 3. In experiment 4 (Figure 1d), object models were represented by fully connected weighted graphs. The link weights in the initial object models were selected with a uniform distribution between 0 and 1. The objects were then randomly permuted Learning with Preknowledge 719 (a) (b) 0.05 0.1 0.15 standard deviation (c) 0.05 0.1 0.15 standard deviation (d) 3r---~----~----~, Ir-'~(" CD2 ~ .- "01 0.05 0.1 0.15 standard deviation or o V ~,/ ~ ("It-' V - 0.05 0.1 0.15 standard deviation Figure 1: (a): 10 clusters, 100 point sets, 20 points each , scale ,rotation, translation, 120 experiments (b): 4 clusters, 64 point sets, 15 points each, affine, 10 % deleted , 5 % spurious, 140 experiments (c): 8 clusters, 256 point sets , 20 points each, affine , 10 % deleted, 5 % spurious, 70 experiments (d): 4 clusters, 64 graphs, 10 nodes each , 360 experiments to form the object instance and uniform noise was added to the link weights. 64 object instances were generated from 4 object models consisting of 10 node graphs with 100 links. The standard deviation of the noise was varied by .01 from .01 to .12. 30 experiments where run at each noise level. In most experiments at low noise levels (~ .06 for point sets, ~ .03 for graphs), the object prototypes learned were very similar to the object models. Even at higher noise levels object prototypes similar to the object models are formed, though less consistently. Results from about 700 experiments are plotted. The objective for experiment 3 contained close to one million variables and converged in about 4 hours on an SGI Indigo workstation. The convergence times of the objectives of experiments 1, 2 and 4 were 120, 10 and 10 minutes respectively. 5 Conclusions It has long been argued by many, that learning in complex domains typically associated with human intelligence requires some type of prior structure or knowledge. We have begun to develop a set of tools that will allow the incorporation of prior 720 Steven Gold, Anand Rangarajan, Eric Mjolsness structure within learning. Our models incorporate many features needed in complex domains like vision - noise, missing and spurious features, non-rigid transformations. They can learn objects with inherent structure, like graphs. Many experiments have been run on experimentally generated data sets. Several directions for future research hold promise. One might be the learning of OCR data [Gold et al., 1995]. Second a supervised learning stage could be added to our algorithms. Finally the power of the distance measures can be enhanced to operate on attributed relational graphs with deleted nodes and links [Gold and Rangarajan, 1995]. Acknowledgements ONR/DARPA: NOOOI4-92-J-4048 , AFOSR: F49620-92-J-0465 and Yale CTAN. References S. Geman , E. Bienenstock, and R. Doursat. (1992) Neural networks and the bias/variance dilemma. Neural Computation 4:1-58. S. Gold, E . Mjolsness and A. Rangarajan. (1994) Clustering with a domain-specific distance measure. In J . Cowan et al., (eds.), NIPS 6. Morgan Kaufmann. S. Gold, C. P. Lu, A. Rangarajan, S. Pappu and E. Mjolsness. (1995) New algorithms for 2D and 3D point matching: pose estimation and correspondence. In G. Tesauro et al., (eds.) , NIPS 7. San Francisco, CA: Morgan Kaufmann. S. Gold and A. Rangarajan (1995) A graduated assignment algorithm for graph matching. YALEU/DCS/TR-I062, Yale Univ., CS Dept. M. I. Jordan and R. A. Jacobs. (1994) Hierarchical mixtures of experts and the EM algorithm . Neural Computation, 6:181-214. E. Mjolsness. Visual grammars and their neural networks. (1992) SPIE Conference on the Science of Artificial Neural Networks, 1110:63-85 . C. Peterson and B. Soderberg. A new method for mapping optimization problems onto neural networks. (1989) International Journal of Neural Systems,I(1):3-22. A. Rangarajan , S. Gold and E. Mjolsness. (1994) A novel optimizing network architecture with applications. YALEU /DCS/TR- I036, Yale Univ. , CS Dept . R. Shepard. (1989). Internal representation of universal regularities: A challenge for connectionism. In L. Nadel et al. , (eds.), Neural Connections, Mental Computation. Cambridge, MA , London, England: Bradford/MIT Press. P. Simard, Y. Le Cun, and J. Denker. (1993) Efficient pattern recognition using a transformation distance. In S. Hanson et ai., (eds.), NIPS 5. San Mateo, CA: Morgan Kaufmann. C. Williams , R. Zemel, and M. Mozer. (1993) Unsupervised learning of object models. AAAI Tech. Rep . FSS-93-04, Univ. of Toronto, CS Dept. A. L. Yuille and J .J. Kosowsky. (1994) . Statistical physics algorithms that converge. Neural Computation, 6:341-356 .
933 |@word decomposition:1 jacob:2 tr:2 yaleu:2 initial:3 series:4 selecting:1 recovered:1 comparing:1 analytic:1 designed:1 intelligence:1 selected:2 oblique:1 compo:1 mental:1 node:3 toronto:1 constructed:1 consists:1 manner:2 introduce:1 expected:1 multi:1 brain:1 ctan:1 decomposed:2 cpu:1 begin:2 bounded:1 null:1 fuzzy:2 finding:1 transformation:16 scaled:3 unit:1 f3m:2 organize:1 engineering:1 local:1 vertically:1 might:1 mateo:1 projective:2 yj:1 pappu:1 mjkllxj:2 logz:1 universal:1 significantly:2 matching:17 projection:1 pre:1 indigo:1 onto:3 unlabeled:3 close:1 equivalent:2 imposed:1 deterministic:3 missing:3 map:2 center:4 williams:2 coordinate:5 diego:1 today:1 enhanced:1 us:1 recognition:1 geman:2 steven:5 solved:1 calculate:1 connected:1 mjolsness:10 decrease:1 yk:4 ran:1 mozer:1 transforming:1 dynamic:2 yuille:2 upon:2 dilemma:1 eric:5 efficiency:1 translated:3 darpa:1 routinely:1 represented:4 univ:3 fast:1 london:1 artificial:1 zemel:1 outside:2 larger:1 otherwise:1 grammar:2 ability:1 transform:3 noisy:3 combining:1 loop:17 gold:12 amplification:1 description:1 obliquely:1 sgi:1 convergence:1 cluster:13 rangarajan:11 regularity:1 generating:1 incremental:1 gab:2 object:54 help:1 rotated:3 recurrent:1 develop:1 pose:2 c:5 direction:1 saved:1 human:1 adjacency:1 argued:2 clustered:3 connectionism:1 adjusted:1 hold:1 scope:1 cognition:1 bj:1 mapping:1 nents:1 estimation:1 individually:1 create:2 tool:1 weighted:8 minimization:2 mit:1 gaussian:1 varying:1 derived:1 consistently:1 greatly:1 tech:1 inference:1 rigid:1 membership:3 lj:1 typically:1 spurious:7 bienenstock:1 orientation:1 softmax:2 initialize:2 biologist:1 construct:2 represents:1 emj:1 unsupervised:2 future:1 minimized:4 others:1 haven:1 employ:1 few:1 inherent:2 randomly:5 replaced:2 phase:3 consisting:2 mixture:1 held:1 edge:1 respective:1 logmjk:1 plotted:2 instance:14 column:3 increased:1 assignment:1 cost:1 deviation:7 subset:1 hundred:1 uniform:4 international:1 subscribe:1 probabilistic:1 physic:1 aaai:1 expert:1 simard:2 doubt:1 li:2 vi:3 minimize:6 square:1 formed:2 ir:1 became:1 variance:2 kaufmann:3 efficiently:1 maximized:1 generalize:1 bayesian:1 lu:1 researcher:1 straight:1 mia:5 converged:1 ed:4 distort:1 associated:1 attributed:1 spie:1 workstation:2 treatment:1 begun:2 noooi4:1 knowledge:3 carefully:1 higher:1 supervised:1 evaluated:1 though:1 generality:1 stage:1 until:1 effect:1 hypothesized:1 multiplier:2 regularization:1 analytically:1 iteratively:1 during:1 self:1 lastname:1 clocked:3 image:8 ef:2 recently:1 novel:1 rotation:6 permuted:5 shear:6 dji:1 shepard:2 million:3 extend:1 functionally:1 cambridge:1 ai:4 approx:2 rasa:1 had:1 add:1 optimizing:1 jolla:1 tesauro:1 inequality:1 onr:1 rep:1 morgan:3 minimum:1 tabula:1 converge:1 lik:1 reduces:1 match:5 england:1 long:1 lai:1 va:2 nadel:1 enhancing:1 vision:2 achieved:1 separately:1 annealing:4 operate:1 doursat:1 ascent:2 cowan:1 anand:5 incorporates:1 jordan:2 call:2 iii:1 xj:3 graduated:1 architecture:2 inner:10 idea:1 prototype:6 f:1 algebraic:3 maybe:1 mai:11 generate:1 promise:1 four:1 deleted:6 graph:25 enforced:2 run:6 fourth:1 nascent:1 distorted:1 place:1 locke:1 scaling:2 bound:3 ct:1 display:1 yale:5 correspondence:4 incorporation:1 constraint:12 speed:1 min:1 department:2 according:1 combination:2 across:2 em:2 cun:1 ykll:2 projecting:1 invariant:7 remains:1 slack:1 needed:1 mind:2 end:2 serf:1 epm:3 denker:1 ocr:1 hierarchical:1 enforce:2 appropriate:2 appearing:1 original:3 clustering:12 objective:24 added:5 occurs:1 distance:41 link:4 outer:7 evaluate:1 mail:2 trivial:1 yjk:1 vai:3 vertical:1 descent:5 regularizes:1 relational:1 looking:2 excluding:1 dc:2 ucsd:2 varied:2 lb:1 community:1 introduced:1 pair:1 kosowsky:2 connection:1 hanson:1 california:1 learned:3 hour:1 nip:3 address:1 bar:1 below:1 perception:1 firstname:1 pattern:1 challenge:1 convincingly:1 ia:1 power:1 mjk:8 isodata:1 lk:4 irrespective:1 created:3 speeding:1 prior:3 acknowledgement:1 tangent:1 afosr:1 fully:1 permutation:7 prototypical:2 generation:1 soderberg:2 degree:2 affine:9 viewpoint:1 translation:6 row:4 preknowledge:5 bias:4 allow:1 peterson:2 barrier:1 absolute:1 f49620:1 curve:1 calculated:1 author:1 san:3 avoided:1 lz:2 assumed:1 francisco:1 xi:6 iterative:2 quantifies:2 nature:1 learn:3 ca:3 expanding:1 complex:2 domain:11 bounding:1 noise:16 repeated:1 fashion:3 position:1 third:1 minute:1 specific:2 cd2:1 essential:1 importance:1 flattened:1 dissimilarity:3 occurring:1 sheared:2 lt:1 explore:1 saddle:3 visual:3 lagrange:2 contained:2 tracking:1 miad:1 egm:3 ma:1 identity:1 formulated:4 fuzziness:1 towards:1 experimentally:1 determined:1 except:2 operates:3 total:1 gij:1 bradford:1 experimental:3 la:2 ya:6 internal:1 incorporate:2 dept:3
7,172
934
Reinforcement Learning Predicts the Site of Plasticity for Auditory Remapping in the Barn Owl Alexandre Pougett Cedric Deffayett Terrence J. Sejnowskit cedric@salk.edu terry@salk.edu alex@salk .edu tHoward Hughes Medical Institute The Salk Institute La Jolla, CA 92037 Department of Biology University of California, San Diego and tEcole Normale Superieure 45 rue d'Ulm 75005 Paris, France Abstract The auditory system of the barn owl contains several spatial maps . In young barn owls raised with optical prisms over their eyes, these auditory maps are shifted to stay in register with the visual map, suggesting that the visual input imposes a frame of reference on the auditory maps. However, the optic tectum, the first site of convergence of visual with auditory information, is not the site of plasticity for the shift of the auditory maps; the plasticity occurs instead in the inferior colliculus, which contains an auditory map and projects into the optic tectum . We explored a model of the owl remapping in which a global reinforcement signal whose delivery is controlled by visual foveation. A hebb learning rule gated by reinforcement learned to appropriately adjust auditory maps. In addition, reinforcement learning preferentially adjusted the weights in the inferior colliculus, as in the owl brain, even though the weights were allowed to change throughout the auditory system. This observation raises the possibility that the site of learning does not have to be genetically specified, but could be determined by how the learning procedure interacts with the network architecture . 126 Alexandre Pouget, Cedric Deffayet, Te"ence J. Sejnowski c,an c:::======:::::? ? ~_m Visual System Optic Tectum Inferior Colllc-ulus External nucleua .- t Forebrain Field L U~ a~ t Ovold.H. Nucleull ?"bala:m.ic Relay Inferior Colltculu. Cenlnll Nucleus C1ec) t Cochlea Figure 1: Schematic view of the auditory pathways in the barn owl. 1 Introduction The barn owl relies primarily on sounds to localize prey [6] with an accuracy vastly superior to that of humans. Figure 1A illustrates some of the nuclei involved in processing auditory signals. The barn owl determines the location of sound sources by comparing the time and amplitude differences of the sound wave between the two ears. These two cues are combined together for the first time in the shell and core of the inferior colliculus (ICc) which is shown at the bottom of the diagram . Cells in the ICc are frequency tuned and subject to spatial aliasing. This prevents them from unambiguously encoding the position of objects. The first unambiguous auditory map is found at the next stage, in the external capsule of the inferior colliculus (ICx) which itself projects to the optic tectum (OT). The OT is the first subforebrain structure which contains a multimodal spatial map in which cells typically have spatially congruent visual and auditory receptive fields. In addition, these subforebrain auditory pathways send one major collateral toward the forebrain via a thalamic relay. These collaterals originate in the ICc and are thought to convey the spatial location of objects to the forebrain [3]. Within the forebrain, two major structures have been involved in auditory processing: the archistriatum and field L. The archistriatum sends a projection to both the inferior colliculus and the optic tectum . Knudsen and Knudsen (1985) have shown that these auditory maps can adapt to systematic changes in the sensory input. Furthermore, the adaptation appears to be under the control of visual input, which imposes a frame of reference on the incoming auditory signals. In owls raised with optical prisms, which introduce a systematic shift in part of the visual field, the visual map in the optic tectum was identical to that found in control animals, but the auditory map in the ICx was shifted by the amount of visual shift introduced by the prisms. This plasticity ensures that the visual and auditory maps stay in spatial register during growth Reinforcement Learning Predicts the Site of Plasticity for Auditory Remapping 127 and other perturbations to sensory mismatch. Since vision instructs audition, one might expect the auditory map to shift in the optic tectum, the first site of visual-auditory convergence. Surprisingly, Brainard and Knudsen (1993b) observed that the synaptic changes took place between the ICc and the ICx, one synapse before the site of convergence. These observations raise two important questions: First, how does the animal knows how to adapt the weights in the ICx in the absence of a visual teaching signal? Second, why does the change take place at this particular location and not in the aT where a teaching signal would be readily available? In a previous model [7], this shift was simulated using backpropagation to broadcast the error back through the layers and by constraining the weights changes to the projection from the ICc to ICx. There is, however, no evidence for a feedback projection between from the aT to the ICx that could transmit the error signal; nor is there evidence to exclude plasticity at other synapses in these pathways. In this paper, we suggest an alternative approach in which vision guides the remapping of auditory maps by controlling the delivery of a scalar reinforcement signal. This learning proceeds by generating random actions and increasing the probability of actions that are consistently reinforced [1, 5] . In addition, we show that reinforcement learning correctly predicts the site of learning in the barn owl, namely at the ICx-ICc synapse, whereas backpropagation [8] does not favor this location when plasticity is allowed at every synapse. This raises a general issue: the site of synaptic adjustment might be imposed by the combination of the architecture and learning rule, without having to restrict plasticity to a particular synapse. 2 2.1 Methods Network Architecture The network architecture of the model based on the barn owl auditory system, shown in figure 2A, contains two parallel pathways. The input layer was an 8x21 map corresponding to the ICc in which units responded to frequency and interaural phase differences. These responses were pooled together to create auditory spatial maps at subsequent stages in both pathways. The rest of the network contained a series of similar auditory maps, which were connected topographically by receptive fields 13 units wide. We did not distinguish between field L and the archistriatum in the forebrain pathways and simply used two auditory maps, both called FBr. We used multiplicative (sigma-pi) units in the aT whose activities were determined according to: Yi = L,. w~Br yfBr WfkBr yfc:c (1) j The multiplicative interaction between ICx and FBr activities was an important assumption of our model. It forced the ICx and FBr to agree on a particular position before the aT was activated. As a result, if the ICx-aT synapses were modified during learning, the ICx-FBr synapses had to be changed accordingly. 128 Alexandre Pouget, Cedric Deffayet, Terrence J. Sejnowski Figure 2: Schematic diagram of weights (white blocks) in the barn owl auditory system. A) Diagram of the initial weights in the network. B) Pattern of weights after training with reinforcement learning on a prism-induced shift offour units. The remapping took place within the ICx and FBr. C) Pattern of weights after training with backpropagation. This time the ICx-OT and FBr-OT weights changed. Weights were clipped between 5.0 and 0.01, except for the FBr-ICx connections whose values were allowed to vary between 8.0 and 0.01. The minimum values were set to 0.01 instead of zero to prevent getting trapped in unstable local minima which are often associated with weights values of zero. The strong coupling between FBr and ICx was another important assumption of the model whose consequence will be discussed in the last section. Examples were generated by simply activating one unit in the ICc while keeping the others to zero, thereby simulating the pattern of activity that would be triggered by a single localized auditory stimulus. In all simulations, we modeled a prism-induced shift of four units. 2.2 Reinforcement learning We used stochastic units and trained the network using reinforcement learning [1]. The weighted sum of the inputs, neti, passed through a sigmoid, f(x) , is interpreted as the probability, Pi, that the unit will be active: Pi = f(neti) * 0.99 + 0.01 (2) were the output of the unit Yi was: ._{a y, - with probability 1 - Pi 1 with probability Pi (3) Reinforcement Learning Predicts the Site of Plasticity for Auditory Remapping 129 Because of the form of the equation for Pi, all units in the network had a small probability (0.01) of being spontaneously active in the absence of any inputs. This is what allowed the network to perform a stochastic search in action space to find which actions were consistently associated with positive reinforcement. We ensured that at most one unit was active per trial by using a winner-take-all competition in each layer. Adjustable weights in the network were updated after each training examples with hebb-like rule gated by reinforcement: (4) A trial consisted in choosing a random target location for auditory input (ICc) and the output of the OT was used to generate a head movement . The reinforcement , r , was then set to 1 for head movements resulting in the foveation of the stimulus and to -0.05 otherwise. 2.3 Backpropagation For the backpropagation network , we used deterministic units with sigmoid activation functions in which the output of a unit was given by: (5) where neti is the weighted sum of the inputs as before. The chain rule was used to compute the partial derivatives of the squared error, E , with respect to each weights and the weights were updated after each training example according to: (6) The target vectors were similar to the input vectors, namely only one OT units was required to be activated for a given pattern, but at a position displaced by 4 units compared to the input. 3 3.1 Results Learning site with reinforcement In a first set of simulation we kept the ICc-ICx and ICc-FBr weights fixed. Plasticity was allowed at these site in later simulations. Figure 2A shows the initial set of weights before learning starts. The central diagonal lines in the weight diagrams illustrate the fact that each unit receives only one non-zero weight from the unit in the layer below at the same location. 130 Alexandre Pouget, Cedric Deffayet, Terrence J. Sejnowski There are two solutions to the remapping: either the weights change within the ICx and FBr, or from the ICx and the FBr to the ~T. As shown in figure 2B , reinforcement learning converged to the first solution. In contrast, the weights between the other layers were unaltered, even though they were allowed to change. To prove that the network could have actually learned the second solution, we trained a network in which the ICc-ICx weights were kept fixed . As we expected, the network shifted its maps simultaneously in both sets of weights converging onto the OT, and the resulting weights were similar to the ones illustrated in figure 2C. However, to reach this solution , three times as many training examples were needed. The reason why learning in the ICx and FBr were favored can be attributed to probabilistic nature of reinforcement learning. If the probability of finding one solution is p, the probability of finding it twice independently is p2. Learning in the ICx and FBR is not independent because of the strong connection from the FBr to the ICx. When the remapping is learned in the FBR this connection automatically remapped the activities in the ICx which in turn allows the ICx-ICx weights to remap appropriately. In the OT on the other hand, the multiplicative connection between the ICx and FBr weights prevent a cooperation between this two sets of weights. Consequently, they have to change independently, a process which took much more training. 3.2 Learning at the ICc-ICx and ICc-FBr synapses The aliasing and sharp frequency tuning in the response of ICc neurons greatly slows down learning at the ICc-ICx and ICc-FBr synapses. We found that when these synapses were free to change, the remapping still took place within the ICx or FBr (figure 3). 3.3 Learning site with backpropagation In contrast to reinforcement learning, backpropagation adjusted the weights in two locations: between the ICx and the OT and between the Fbr and OT (figure 2C). This is the consequence of the tendency of the backpropagation algorithm to first change the weights closest to where the error is injected. 3.4 Temporal evolution of weights Whether we used reinforcement or supervised learning, the map shifted in a very similar way. There was a simultaneous decrease of the original set of weights with a simultaneous increase of the new weights, such that both sets of weights coexisted half way through learning. This indicates that the map shifted directly from the original setting to the new configuration without going through intermediate shifts. This temporal evolution of the weights is consistent the findings of Brainard and Knudsen (1993a) who found that during the intermediate phase of the remapping, cells in the inferior colli cuI us typically have two receptive fields. More recent work however indicates that for some cells the remapping is more continuous(Brainard and Knudsen , personal communication) , a behavior that was not reproduced by either of the learning rule. Reinforcement Learning Predicts the Site of Plasticity for Auditory Remapping 131 Figure 3: Even when the ICc-ICx weights are free to change, the network update the weights in the ICx first. A separate weight matrix is shown for each isofrequency map from the ICc to ICx. The final weight matrices were predominantly diagonal; in contrast, the weight matrix in ICx was shifted. 4 Discussion Our simulations suggest a biologically plausible mechanism by which vision can guide the remapping of auditory spatial maps in the owl's brain. Unlike previous approaches, which relied on visual signals as an explicit teacher in the optic tectum [7], our model uses a global reinforcement signal whose delivery is controlled by the foveal representation of the visual system. Other global reinforcement signals would work as well. For example, a part of the forebrain might compare auditory and visual patterns and report spatial mismatch between the two. This signal could be easily incorporated in our network and would also remap the auditory map in the inferior colli cuI us. Our model demonstrates that the site of synaptic plasticity can be constrained by the interaction between reinforcement learning and the network architecture. Reinforcement learning converged to the most probably solution through stochastic search. In the network, the strong lateral coupling between ICx and FBr and the multiplicative interaction in the OT favored a solution in which the remapping took place simultaneously in the ICx and FBr. A similar mechanism may be at work in the barn owl's brain. Colaterals from FBr to ICx are known to exist, but the multiplicative interaction has not been reported in the barn owl optic tectum. Learning mechanisms may also limit synaptic plasticity. NMDA receptors have been reported in the ICx, but they might not be expressed at other synapses. There may, however, be other mechanisms for plasticity. The site of remapping in our model was somewhat different from the existing observations. We found that the change took place within the ICx whereas Brainard and Knudsen [3] report that it is between the ICc and the ICx. A close examination of their data (figure 11 in [3]) reveals that cells at the bottom of ICx were not 132 Alexandre Pouget, Cedric Deffayet, Terrence J. Sejnowski remapped, as predicted by our model, but at the same time, there is little anatomical or physiological evidence for a functional and hierarchical organization within the ICx. Additional recordings are need to resolve this issue. We conclude that for the barn owl's brain, as well as for our model, synaptic plasticity within ICx was favored over changes between ICc and ICx. This supports the hypothesis that reinforcement learning is used for remapping in the barn owl auditory system. Acknowledgments We thank Eric Knudsen and Michael Brainard for helpful discussions on plasticity in the barn owl auditory system and the results of unpublished experiments. Peter Dayan and P. Read Montague helped with useful insights on the biological basis of reinforcement learning in the early stages of this project. References [1] A.G. Barto and M.1. Jordan. Gradient following without backpropagation in layered networks. Proc. IEEE Int. Conf. Neural Networks, 2:629-636, 1987. [2] M.S. Brainard and E.1. Knudsen. Dynamics of the visual calibration of the map of interaural time difference in the barn owl's optic tectum. In Society For Neuroscience Abstracts, volume 19, page 369.8, 1993. [3] M.S. Brainard and E.!. Knudsen. Experience-dependent plasticity in the inferior colliculus: a site for visual calibration of the neural representation of auditory space in the barn owl. The journal of Neuroscience, 13:4589-4608, 1993. [4] E. Knudsen and P. Knudsen. Vision guides the adjustment of auditory localization in the young barn owls. Science, 230:545-548, 1985. [5] P.R. Montague, P. Dayan, S.J. Nowlan, A. Pouget, and T.J. Sejnowski. Using aperiodic reinforcement for directed self-organization during development. In S.J. Hanson, J.D. Cowan, and C.L. Giles, editors, Advances in Neural Information Processing Systems, volume 5. Morgan-Kaufmann, San Mateo, CA, 1993. [6] R.S. Payne. Acoustic location of prey by barn owls (tyto alba). Journal of Experimental Biology, 54:535-573, 1970. [7] D.J. Rosen, D.E. Rumelhart, and E.I. Knudsen. A connectionist model of the owl's sound localization system. In Advances in Neural Information Processing Systems, volume 6. Morgan-Kaufmann, San Mateo, CA, 1994. [8] D.E. Rumelhart, G.E. Hinton, and R.J . Williams. Learning internal representations by error propagation. In D. E. Rumelhart, J. L. McClelland, and the PDP Research Group, editors, Parallel Distributed Processing, volume 1, chapter 8, pages 318-362. MIT Press, Cambridge, MA, 1986. Dynamic Modelling of Chaotic Time Series with Neural Networks Jose C. Principe, Jyh-Ming Kuo Computational NeuroEngineering Laboratory University of Florida, Gainesville, FL32611 principe@synapse.ee.ufi.edu Abstract This paper discusses the use of artificial neural networks for dynamic modelling of time series. We argue that multistep prediction is more appropriate to capture the dynamics of the underlying dynamical system, because it constrains the iterated model. We show how this method can be implemented by a recurrent ANN trained with trajectory learning. We also show how to select the trajectory length to train the iterated predictor for the case of chaotic time series. Experimental results corroborate the proposed method. 1.0 Introduction The search for a model of an experimental time series has been an important problem in science. For a long time the linear model was almost exclusively used to describe the system that produced the time series [1], but recently nonlinear models have also been proposed to replace the linear ones [2]. Lapedes and Farber [3] showed how artificial neural networks (ANNs) can be used to identify the dynamics of the unknown system that produced the time series. He simply used a multilayer perceptron to predict the next point in state space, and trained this topology with backpropagation. This paper explores more complex neural topologies and training methods with the goal of improving the quality of the identification of the dynamical system, and to understand better the issues of dynamic modelling with neural networks which are far from being totally understood. According to Takens' embedding theorem, a map F: Jilm + 1 ~ Jilm + 1 exists that transforms the current reconstructed state y(t) to the next state y(t + 1) , i.e. y(t+ 1) = F(y(t? (1) 312 Jose Principe, Jyh-Ming Kuo or where m is the estimated dimension of the unknown dynamical system cI>. Note that the map contains several trivial (nonlinear) filters and a predictor. The predictive mapping r: Jilm +1~ R can be expressed as x(t+ 1) = r(x(t? (2) where x (t) = [x (t - 2m) ... x (t - 1) x (t)] T. This is actually the estimated nonlinear autoregressive model of the input time series. The existence of this predictive model lays a theoretical basis for dynamic modelling in the sense that we can build from a vector time series a model to approximate the mapping r . If the conditions of Takens embedding theorem are met, this mapping captures some of the properties of the unknown dynamical system cI> that produced the time series [7]. Presently one still does not have a capable theory to guarantee if the predictor has successfully identified the original model cI>. The simple point by point comparison between the original and predicted time series used as goodness of fit for non-chaotic time series breaks down for chaotic ones [5]. Two chaotic time series can be very different pointwise but be produced by the same dynamical system (two trajectories around the same attractor). The dynamic invariants (correlation dimension, Lyapunov exponents) measure global properties of the attractor, so they should be used as the rule to decide about the success of dynamic modelling. Hence, a pragmatic approach in dynamic modelling is to seed the predictor with a point in state space, feed the output to its input as an autonomous system, and create a new time series. If the dynamic invariants computed from this time series match the ones from the original time series, then we say that dynamic modelling was successful [5]. The long term behavior of the autonomous predictive model seems to be the key factor to find out if the predictor identified the original model. This is the distinguishing factor between prediction of chaotic time series and dynamic modelling. The former only addresses the instantaneous prediction error, while the latter is interested in long term behavior. In order to use this theory, one needs to address the choices of predictor implementation. Due to the universal mapping characteristics of multilayer perceptrons (MLPs) and the existence of well established learning rules to adapt the MLP coefficients, this type of network appears as an appropriate choice [3]. However, one must realize that the MLP is a static mapper, and in dynamic modelling we are dealing with time varying signals, where the past of the signal contains vital information to describe the mapping. The design considerations to select the neural network topology are presented elsewhere [4]. Wejust would like to say that the MLP has to be enhanced with short term memory mechanisms, and that the estimation of the correlation dimension should be used to set the size of the memory layer. The main goal of the paper is to establish the methodology to efficiently train neural networks for dynamic modelling. Dynamic Modelling of Chaotic Time Series with Neural Networks 313 2. Iterated versus Single Step Prediction. From eqn. 2 it seems that the resulting dynamic model F can be obtained through single step prediction. This has been the conventional way to handle dynamic modelling [2],[3]. The predictor is adapted by minimizing the error E _.1 L L = I dist (x (i + 1) -F (1 (i? ) (3) =2m + 1 _.1 where L is the length of the time series, x(i) is the itb data sample, F is the map developed by the predictor and dist() is a distance measure (normally the L2 norm). Notice that the training to obtain the mapping is done independently from sample to sample, i.e. _.1 x(i+ 1) = F (xU? +5 1 _.1 x (i + j) =F (1 (i + j -1? + 5j where 5j are the instantaneous prediction errors, which are minimized during training. Notice that the predictor is being optimized under the assumption that the previous point in state space is known without error. The problem with this approach can be observed when we iterate the predictor as an autonomous system to generate the time series samples. If one wants to produce two samples in the future from sample i the predicted sample i+ 1 needs to be utilized to generate sample i+2. The predictor was not optimized to do this job, because during training the true i+ 1 sample was assumed known. As long as 51 is nonzero (as will be always the case for nontrivial problems), errors will accumulate rapidly. Single step prediction is more associated with extrapolation than with dynamic modelling, which requires the identification of the unique mapping that produces the time series. When the autonomous system generates samples, past values are used as inputs to generate the following samples, which means that the training should constrain also the iterates of the predictive mapping. Putting it in a simple way, we should train the predictor in the same way we are going to use it for testing (Le. as an autonomous system). We propose multistep prediction (or trajectory learning) as the way to constrain the iterates of the mapping developed by the predictor. Let us define L" E = I (4) dist(x(i + I)-xU + 1? =2m + 1 x where k is the number of prediction steps (length ofthe trajectory) and (i + 1) is an estimate of the predictive map _.1 x(i+l) = F (i(i-2m), ... ,i(i? (5) 314 Jose Principe, Jyh-Ming Kuo with j (i) =[ X 0 SiS 2m (i) _.1. F (x(i-2m-l), ... ,x(i-l? Equation (5) states that (for i>2m), i.e. i (i) _ .1. -.1. i>2m is the i-2m iterate of the predictive part of the map - .1. i(i+l) = (F (F ( ... F (x(2m??) t- _.1. = (F 2m (x(2m?) (6) Hence, minimizing the criterion expressed by equation (4) an optimal multistep predictor is obtained. The number of constraints that are imposed during learning is associated with k, the number of prediction steps, which corresponds to the number of iterations of the map. The more iterations, the less likely a sub-optimal solution is found, but note that the training time is being proportionally increased. In a chaotic time series there is a more important consideration that must be brought into the picture, the divergence of nearby trajectories, as we are going to see in a following section. 3. Multistep prediction with neural networks Figure 1 shows the topology proposed in [4] to identify the nonlinear mapping. Notice that the proposed topology is a recurrent neural network. with a global feedback loop. This topology was selected to allow the training of the predictor in the same way as it will be used in testing, i.e. using the previous network outputs to predict the next point. This recurrent architecture should be trained with a mechanism that will constrain the iterates of the map as was discussed above. Single step prediction does not fit this requirement. With multistep prediction, the model system can be trained in the same way as it is used in testing. We seed the dynamic net with a set of input samples, disconnect the input and feed back the predicted sample to the input for k steps. The mean square error between the predicted and true sample at each step is used as the cost function (equation (4?. If the network topology was feed forward , batch learning could be used to train the network, and static backpropagation applied to train the net. However, as a recurrent topology is utilized, a learning paradigm such as backpropagation through time (BPTT) or real time recurrent learning (RTRL) must be utilized [6]. The use of these training methods should not come as a surprise since we are in fact fitting a trajectory over time, so the gradients are time varying. This learning method is sometimes called "trajectory learning" in the recurrent learning literature [6]. A criterion to select the length of the trajectory k will be presented below. The procedure described above must be repeated for several different segments of the time series. For each new training segment, 2m+ 1 samples of the original time series are used to seed the predictor. To ease the training we suggest that successive training sequences of length k overlap by q samples (q<k). For chaotic time series we also suggest that the error be weighted according to the largest Lyapunov exponent. Hence Dynamic Modelling of Chaotic Time Series with Neural Networks 315 the cost function becomes E = Lr L" (7) h(i) ?dist(x(i+jq+l)-i(i+jq+l? J = 01 = 2m+ I where r is the number of training sequences, and h (i) = (e ~ (8) A I -(i-2m-I) III/IX ) In this equation A.max is the largest Lyapunov exponent and L\t the sampling interval. With this weighting the errors for later iteration are given less credit, as they should since due to the divergence of trajectories a small error is magnified proportionally to the largest Lyapunov exponent [7]. 4. Finding the length of the trajectory From the point of view of dynamic modelling, each training sequences should preferably contain enough information to model the attractor. This means that each sequence should be no shorter than the orbital length around the attractor. We proposed to estimate the orbital length as the reciprocal of the median frequency of the spectrum of the time series [8]. Basically this quantity is the average time required for a point to return to the same neighborhood in the attractor. The length of the trajectory is also equivalent to the number of constraints we impose on the iterative map describing the dynamical model. However, in a chaotic time series there is another fundamental limitation imposed on the trajectory length - the natural divergence of trajectories which is controlled by A.max ' the largest Lyapunov exponent. If the trajectory length is too long, then instabilities in the training can be expected. A full discussion of this topic is beyond the scope of this paper, and is presented elsewhere [8]. We just want to say that when A.max is positive there is an uncertainty region around each predicted point that is a function of the number of prediction steps (due to cummulative error). If the trajectory length is too long the uncertainty regions from two neighboring trajectories will overlap, creating conflicting requirements for training (the model is requried to develop a map to follow both segments A and B- Figure 2). It turns out that one can approximately find the number of iterations is that will guarantee no overlap of uncertainty regions [8]. The length of the principal axis of the uncertainty region around a signal trajectory at iteration i can be estimated as ?; = toe ~III/lXiAI (9) where ?0 is the initial separation. Now assuming that the two principal axis of nearby trajectories are colinear, we should choose the number of iterations is such that the distance d j between trajectories is larger than the uncertainty region, i.e. d; ~ 2?; . I I The estimate of is should be averaged over a number of neighboring training sequences (-50 depending on the signal dynamics). Hence, to apply this method three quantities must be estimated: the largest Lyapunov 316 Jose Principe, Jyh-Ming Kuo exponent, using one of the accepted algorithms. The initial separation can be estimated from the one-step predictor. And is by averaging local divergence. The computation time required to estimate these quantities is usually much less than setting by trial and error the length of the trajectory until a reasonable learning curve is achieved. We also developed a method to train predictors for chaotic signals with large A.max ' but it will not be covered in this paper [8]. 5. Results We used this methodology to model the Mackey-Glass system (d=30, sampled at 116 Hz). A signal of 500 samples was obtained by 4th order Runge-Kutta integration and normalized between -1,1. The largest Lyapunov exponent for this signal is 0.0071 nats/sec. We selected a time delay neural network (TDNN) with topology 8-14-1. The output unit is linear, and the hidden layer has sigmoid nonlinearities. The number of taps in the delay line is 8. We trained a one-step predictor and the multistep predictor with the methodology developed in this paper to compare results. The single step predictor was trained with static backpropagation with no momentum and step size of 0.001. Trained was stopped after 500 iterations. The final MSE was 0.000288. After training, the predictor was seeded with the first 8 points of the time series and iterated for 3,000 times. Figure 3a shows the corresponding output. Notice that the waveform produced by the model is much more regular that the Mackey-Glass signal, showing that some fine detail of the attractor has not been captured. Next we trained the same TDNN with a global feedback loop (TDNNGF). The estimate of the is over the neighboring orbits provided an estimate of 14, and it is taken as the length of the trajectory. We displaced each training sequence by 3 samples (q=3 in eqn 7). BPTT was used to train the TDNNGF for 500 iterations over the same signal. The final MSE was 0.000648, higher than for the TDNN case. We could think that the resulting predictor was worse. The TDNNGF predictor was initialized with the same 8 samples of the time series and iterated for 3,000 times. Figure 3b shows the resulting waveform. It "looks" much closer to the original Mackey-Glass time series. We computed the average prediction error as a function of iteration for both predictors and also the theoretical rate of divergence of trajectories assuming an initial error EO (Casdagli conjecture, which is the square of eqn 9) [7]. As can be seen in Figure 4 the TDNNGF is much closer to the theoretical limit, which means a much better model. We also computed the correlation dimension and the Lyapunov exponent estimated from the generated time series, and the figures obtained from TDNNGF are closer to the original time series. Figure 5 shows the instability present in the training when the trajectory length is above the estimated value of 14. For this case the trajectory length is 20. As can be seen the MSE decreases but then fluctuates showing instability in the training. 6. Conclusions This paper addresses dynamic modelling with artificial neural networks. We showed 317 Dynamic Modelling of Chaotic Time Series with Neural Networks that the network topology should be recurrent such that the iterative map is constrained during learning. This is a necessity since dynamic modelling seeks to capture the long term behavior of the dynamical system. These models can also be used as a sample by sample predictors. Since the network topology is recurrent, backpropagation through time or real time recurrent learning should be used in training. In this paper we showed how to select the length of the trajectory to avoid instability during training. A lot more work needs to be done to reliably capture dynamical properties of time series and encapsulate them in artificial models. But we believe that the careful analysis of the dynamic characteristics and the study of its impact on the predictive model performance is much more promising than guess work. According to this (and others) studies, modelling of chaotic time series of low Amax seems a reality. We have extended some of this work for time series with larger Amax , and successfully captured the dynamics of the Lorenz system [8]. But there, the parameters for learning have to be much more carefully selected, and some of the choices are still arbitrary. The main issue is that the trajectories diverge so rapidly that predictors have a hard time to capture information regarding the global system dynamics. It is interesting to study the limit of predictability of this type of approach for high dimensional and high Amax chaos. Predictor Corr. Dim. Lyapunov MG30 2.70+/-0.05 0.0073+/-0.000 1 TDNNGF 2.65+/-0.03 0.0074+/-0.0001 TDNN 1.60+1-0.10 0.0063+/-0.0001 segment B Figure 1. Prop_osed recurrent architecture Figure 2. State space representation in (IDNNGF) training a model 7. Acknowledgments This work was partially supported by NSF grant #ECS-9208789, and ONR #1494-941-0858. 318 Jose Principe, Jyh-Ming Kuo -OJ! -OJ! m ~ ~ _ ~ ~ w ~ _ ~ _ Figure 3a. Generated sequence with the TDNN ! m TDNNerror 002 015 TDNNGF j /~ .005 o o asdagli conjecture ? ?? 5 , 0.2 / e~ ~ /1,/ 0.01 C -I~~~~!:--=--=--:!~~~--::!=----= o ~ '110. 110 ICIO ~ W JIO ~ . . II1II Figure 3b. uenerated sequence with the TDNNGF ? ????? ?? ?? to .-' ~ / _._.:::~ IS ~ ~ ? Figure 4. Comparison of predictors Figure 5. Instability in training 8. References [1] Box, G. E., and G. M. Jenkins, Time Series Analysis, Forecasting and Control, Holden Day, San Francisco, 1970. [2] Weigend, A. S., B. A. Huberman, and D. E. Rumelhart, "Predicting the future: a connectionist approach," International Journal of Neural Systems, vol. 1, pp. 193209, 1990. [3] Lapedes, R., and R. Farber, "Nonlinear signal processing using neural networks: prediction and system modelling," Technical Report LA-UR87-2662, Los Alamos National Laboratory, Los Alamos, New Mexico, 1987. [4] Kuo J-M., Principe J.C., "A systematic approach to chaotic time series modeling with neural networks", in IEEE Workshop on Neural Nets for Signal Processing, Ermioni, Greece, 1994. [5] Principe, J. C., A. Rathie, and J. M. Kuo, "Prediction of chaotic time series with neural networks and the issue of dynamic modeling," International Journal of Biburcation and Chaos, vol. 2, no. 4, pp. 989-996, 1992. [6] Hertz, J, A. Krogh, and R. G. Palmer, Introduction to the Theory of Neural Computation, Addison-Wesley, Redwood City, CA, 1991. [7] Casdagli, M., "Nonlinear prediction of chaotic time series," Physica D 35, pp.335-356, 1989. [8] Kuo, J.M., "Nonlinear Dynamic Modelling with Artificial neural networks", Ph.D. dissertation, University of FLorida, 1993.
934 |@word trial:3 unaltered:1 seems:3 norm:1 casdagli:2 bptt:2 simulation:4 seek:1 gainesville:1 thereby:1 ulus:1 necessity:1 configuration:1 contains:6 series:41 foveal:1 exclusively:1 initial:5 tuned:1 lapedes:2 past:2 existing:1 current:1 comparing:1 nowlan:1 activation:1 si:1 must:5 readily:1 realize:1 subsequent:1 plasticity:17 update:1 mackey:3 cue:1 half:1 selected:3 guess:1 accordingly:1 reciprocal:1 core:1 short:1 dissertation:1 lr:1 iterates:3 location:8 successive:1 instructs:1 prove:1 pathway:6 interaural:2 fitting:1 introduce:1 orbital:2 expected:2 behavior:4 nor:1 dist:4 aliasing:2 brain:4 ming:5 automatically:1 resolve:1 little:1 increasing:1 totally:1 project:3 becomes:1 underlying:1 provided:1 remapping:16 what:1 tyto:1 interpreted:1 yfc:1 offour:1 developed:4 finding:4 magnified:1 guarantee:2 temporal:2 every:1 preferably:1 growth:1 ensured:1 demonstrates:1 control:3 unit:18 medical:1 normally:1 grant:1 encapsulate:1 before:4 positive:2 understood:1 local:2 limit:3 consequence:2 receptor:1 encoding:1 multistep:6 approximately:1 might:4 twice:1 mateo:2 ease:1 palmer:1 averaged:1 directed:1 acknowledgment:2 unique:1 spontaneously:1 testing:3 hughes:1 block:1 backpropagation:14 chaotic:17 procedure:2 coexisted:1 universal:1 thought:1 projection:3 regular:1 suggest:4 onto:1 close:1 layered:1 instability:5 conventional:1 map:36 imposed:3 deterministic:1 equivalent:1 send:1 williams:1 independently:3 pouget:5 rule:7 insight:1 amax:3 embedding:2 handle:1 autonomous:5 sejnowskit:1 transmit:1 updated:2 diego:1 tectum:10 controlling:1 ulm:1 target:2 enhanced:1 us:1 distinguishing:1 hypothesis:1 rumelhart:4 utilized:3 lay:1 predicts:5 bottom:2 observed:2 capture:5 region:5 ensures:1 connected:1 movement:2 decrease:2 constrains:1 nats:1 colli:2 dynamic:31 personal:1 trained:10 raise:3 colinear:1 segment:4 topographically:1 predictive:7 localization:2 eric:1 basis:2 multimodal:1 easily:1 montague:2 chapter:1 train:7 forced:1 describe:2 sejnowski:5 artificial:5 choosing:1 neighborhood:1 whose:5 fluctuates:1 larger:2 plausible:1 say:3 otherwise:1 favor:1 think:1 itself:1 final:3 reproduced:1 runge:1 triggered:1 sequence:8 net:3 took:6 propose:1 interaction:4 adaptation:1 neighboring:3 payne:1 loop:2 rapidly:2 competition:1 getting:1 los:2 convergence:3 requirement:2 congruent:1 generating:1 produce:2 object:2 brainard:7 coupling:2 illustrate:1 recurrent:10 develop:1 depending:1 job:1 p2:1 krogh:1 implemented:1 predicted:6 strong:3 come:1 met:1 lyapunov:9 waveform:2 aperiodic:1 farber:2 filter:1 stochastic:3 human:1 owl:23 activating:1 biological:1 adjusted:2 neuroengineering:1 physica:1 around:4 credit:1 barn:18 ic:1 seed:3 mapping:10 predict:2 scope:1 major:2 vary:1 early:1 relay:2 estimation:1 proc:1 largest:6 create:2 successfully:2 city:1 weighted:3 mit:1 brought:1 always:1 modified:1 normale:1 avoid:1 varying:2 barto:1 consistently:2 modelling:21 indicates:2 greatly:1 contrast:3 sense:1 glass:3 helpful:1 dim:1 dayan:2 dependent:1 typically:2 holden:1 hidden:1 jq:2 going:3 france:1 interested:1 issue:5 favored:3 exponent:8 takens:2 development:1 animal:2 spatial:8 raised:2 integration:1 constrained:2 field:7 having:1 sampling:1 biology:2 identical:1 look:1 future:2 minimized:1 others:2 stimulus:2 report:3 connectionist:2 primarily:1 rosen:1 simultaneously:2 divergence:5 national:1 phase:2 attractor:6 organization:2 mlp:3 possibility:1 adjust:1 activated:2 chain:1 capable:1 partial:1 closer:3 experience:1 collateral:2 shorter:1 initialized:1 orbit:1 theoretical:3 stopped:1 increased:1 modeling:2 giles:1 ence:1 corroborate:1 goodness:1 cost:2 predictor:29 alamo:2 delay:2 successful:1 too:2 reported:2 teacher:1 combined:1 explores:1 fundamental:1 international:2 stay:2 systematic:3 terrence:4 probabilistic:1 diverge:1 michael:1 together:2 vastly:1 squared:1 ear:1 central:1 broadcast:1 choose:1 worse:1 external:2 conf:1 audition:1 derivative:1 creating:1 return:1 suggesting:1 exclude:1 nonlinearities:1 pooled:1 sec:1 int:1 alba:1 coefficient:1 disconnect:1 register:2 multiplicative:5 view:2 later:2 helped:1 break:1 extrapolation:1 lot:1 wave:1 start:1 thalamic:1 parallel:2 relied:1 mlps:1 square:2 accuracy:1 responded:1 kaufmann:2 who:1 characteristic:2 reinforced:1 efficiently:1 identify:2 ofthe:1 identification:2 iterated:5 produced:5 basically:1 trajectory:27 archistriatum:3 converged:2 anns:1 simultaneous:2 synapsis:7 reach:1 synaptic:5 isofrequency:1 frequency:4 involved:2 pp:3 toe:1 associated:4 attributed:1 static:3 sampled:1 auditory:39 nmda:1 amplitude:1 greece:1 carefully:1 actually:2 back:2 appears:2 alexandre:5 feed:3 higher:1 wesley:1 supervised:1 follow:1 unambiguously:1 response:2 methodology:3 synapse:5 day:1 done:2 though:2 box:1 furthermore:1 just:1 stage:3 correlation:3 until:1 hand:1 receives:1 eqn:3 nonlinear:7 propagation:1 quality:1 believe:1 consisted:1 true:2 contain:1 normalized:1 evolution:2 hence:4 former:1 seeded:1 spatially:1 read:1 laboratory:2 nonzero:1 illustrated:1 white:1 during:9 self:1 inferior:10 unambiguous:1 criterion:2 instantaneous:2 consideration:2 recently:1 chaos:2 predominantly:1 superior:1 sigmoid:3 functional:1 winner:1 volume:4 discussed:2 he:1 accumulate:1 cambridge:1 tuning:1 teaching:2 had:2 mapper:1 calibration:2 closest:1 recent:1 ufi:1 showed:3 jolla:1 prism:5 success:1 onr:1 yi:2 morgan:2 minimum:2 additional:1 somewhat:1 impose:1 captured:2 eo:1 seen:2 paradigm:1 signal:22 full:1 sound:4 technical:1 match:1 adapt:3 long:7 controlled:3 schematic:2 converging:1 prediction:18 impact:1 multilayer:2 vision:4 cochlea:1 iteration:9 sometimes:1 achieved:1 cell:5 addition:3 whereas:2 want:2 fine:1 interval:1 diagram:4 median:1 source:1 sends:1 appropriately:2 ot:11 rest:1 unlike:1 probably:1 subject:1 induced:2 recording:1 cummulative:1 cowan:1 hz:1 jordan:1 ee:1 constraining:1 intermediate:2 vital:1 iii:2 enough:1 iterate:2 fit:2 architecture:7 restrict:1 topology:11 identified:2 regarding:1 br:1 shift:8 whether:1 passed:1 remap:2 forecasting:1 peter:1 action:4 useful:1 proportionally:2 covered:1 amount:1 transforms:1 ph:1 mcclelland:1 generate:4 exist:1 nsf:1 shifted:6 notice:4 trapped:1 neuroscience:2 correctly:1 per:1 estimated:7 anatomical:1 vol:2 group:1 key:1 four:1 putting:1 localize:1 prevent:2 prey:2 kept:2 sum:2 colliculus:6 weigend:1 jose:5 injected:1 uncertainty:5 place:6 throughout:1 clipped:1 almost:1 decide:1 reasonable:1 separation:2 delivery:3 layer:7 distinguish:1 bala:1 activity:4 nontrivial:1 adapted:1 optic:10 constraint:2 alex:1 constrain:3 nearby:2 generates:1 optical:2 conjecture:2 department:1 jio:1 according:5 combination:1 cui:2 hertz:1 rtrl:1 biologically:1 presently:1 invariant:2 taken:1 equation:5 agree:1 forebrain:6 turn:2 discus:1 mechanism:6 neti:3 needed:1 know:1 describing:1 addison:1 rathie:1 available:1 jenkins:1 apply:1 hierarchical:1 appropriate:2 simulating:1 alternative:1 batch:1 florida:2 existence:2 original:9 x21:1 build:1 establish:1 society:1 question:1 quantity:3 occurs:1 receptive:3 interacts:1 diagonal:2 subforebrain:2 gradient:2 kutta:1 distance:2 separate:1 thank:1 simulated:1 lateral:1 originate:1 topic:1 argue:1 unstable:1 trivial:1 toward:1 reason:1 assuming:2 length:18 modeled:1 pointwise:1 minimizing:2 preferentially:1 mexico:1 sigma:1 slows:1 implementation:1 design:1 reliably:1 adjustable:1 gated:2 perform:1 unknown:3 observation:3 displaced:2 neuron:1 knudsen:12 hinton:1 communication:1 head:2 incorporated:1 frame:2 pdp:1 perturbation:1 extended:1 redwood:1 sharp:1 arbitrary:1 introduced:1 namely:2 paris:1 specified:1 required:3 optimized:2 connection:4 unpublished:1 tap:1 hanson:1 california:1 acoustic:1 learned:3 icx:44 conflicting:1 established:1 address:3 beyond:1 proceeds:1 below:2 pattern:5 mismatch:2 dynamical:8 usually:1 remapped:2 genetically:1 deffayet:4 max:4 memory:2 oj:2 terry:1 overlap:3 natural:1 examination:1 predicting:1 eye:1 picture:1 axis:2 tdnn:5 ur87:1 literature:1 l2:1 icc:21 cedric:6 expect:1 interesting:1 limitation:1 itb:1 versus:1 localized:1 nucleus:2 consistent:1 imposes:2 editor:2 pi:6 elsewhere:2 changed:2 cooperation:1 surprisingly:1 last:1 keeping:1 free:2 supported:1 guide:3 allow:1 understand:1 perceptron:1 institute:2 wide:1 distributed:1 feedback:3 dimension:4 curve:1 autoregressive:1 sensory:2 forward:1 reinforcement:27 san:4 far:1 ec:1 reconstructed:1 approximate:1 dealing:1 global:7 active:3 incoming:1 reveals:1 conclude:1 assumed:1 francisco:1 jyh:5 spectrum:1 search:3 continuous:1 iterative:2 why:2 reality:1 promising:1 nature:1 capsule:1 ca:4 improving:1 mse:3 complex:1 rue:1 did:1 main:2 allowed:6 repeated:1 convey:1 xu:2 site:17 hebb:2 salk:4 predictability:1 icio:1 sub:1 position:3 momentum:1 explicit:1 weighting:1 ix:1 young:2 down:2 theorem:2 showing:2 explored:1 physiological:1 evidence:3 exists:1 workshop:1 fbr:23 lorenz:1 corr:1 ci:3 te:1 illustrates:1 surprise:1 simply:3 likely:1 visual:18 prevents:1 expressed:3 adjustment:2 contained:1 partially:1 scalar:1 corresponds:1 determines:1 relies:1 ma:1 shell:1 goal:2 ii1ii:1 consequently:1 ann:1 careful:1 replace:1 absence:2 change:13 hard:1 determined:2 foveation:2 except:1 huberman:1 averaging:1 principal:2 called:2 kuo:8 accepted:1 tendency:1 la:2 experimental:3 perceptrons:1 select:4 principe:8 pragmatic:1 internal:1 support:1 latter:1 superieure:1
7,173
935
A Model for Chemosensory Reception Rainer Malaka J Thomas Ragg Institut fUr Logik, Komplexitat und Oeduktionssysteme Universitat Karlsruhe, PO Box 0-76128 Karlsruhe, Germany e-mail: malaka@ira.uka.de.ragg@ira.uka.de Martin Hammer Institut fur Neurobiologie Freie Universitat Berlin 0-14195 Berlin, Germany e-mail: mhammer@castor.zedat.fu-berlin.de Abstract A new model for chemosensory reception is presented. It models reactions between odor molecules and receptor proteins and the activation of second messenger by receptor proteins. The mathematical formulation of the reaction kinetics is transformed into an artificial neural network (ANN). The resulting feed-forward network provides a powerful means for parameter fitting by applying learning algorithms. The weights of the network corresponding to chemical parameters can be trained by presenting experimental data. We demonstrate the simulation capabilities of the model with experimental data from honey bee chemosensory neurons. It can be shown that our model is sufficient to rebuild the observed data and that simpler models are not able to do this task. 1 INTRODUCTION Terrestrial animals, vertebrates and invertebrates, have developed very similar solutions for the problem of recognizing volatile substances [Vogt et ai., 1989]. Odor molecules bind to receptor proteins (receptor sites) at the cell membrane of the sensory cell. This interaction of odor molecules and receptor proteins activates a G-protein mediated second 62 Rainer Malaka, Thomas Ragg, Martin Hammer _____ odor molecules 'f! ~ /2 " 4! ___ Ins" "----- second messengers ----- . action potentials Ionic Influx Figure 1: Reaction cascade in chemosensory neurons. Volatile odor molecules reach receptor proteins at the surface of the chemosensory neuron. The odor bound binding proteins activate second messengers (e.g. G-proteins). The activated second messengers cause a change in the conductivity of ion channels. Through ionic influx a depolarization can build an action potential. messenger process. The concentrations of cAMP or IP3 rise rapidly and activate cyclicnucleotide-gated ion channels or IP3-gated ion channels [Breer et al., 1989, Shepherd, 1991]. As a result of this second messenger reaction cascade the conductivity of ion channels is changed and the cell can be hyperpolarized or depolarized, which can cause the generation of action potentials. It has been shown that one odor is able to activate different second messenger processes and that there is some interaction between the different second messenger processes [Breer & Boekhoff, 1992]. Figure 1 shows schematically the cascade of reactions from odor molecules over receptor proteins and second messengers up to the changing of ion channel conductance and the generation of action potentials. Responses of sensory neurons can be very complex. The response as a function of the odor concentration is highly non-linear. The response to mixtures can be synergistic or inhibitory relative to the response to the components of the compound. A synergistic effect occurs, if the response of one sensory cell to a binary mixture of two odors AI and A2 with concentrations [AI]' [A2l is higher than the sum of the responses to the odors AI, A2 at concentrations [A tl, [A2] alone. An inhibitory effect occurs, if the response to the mixture is smaller than either response to the single odors. In bee subplacode and placode recordings both effects can be observed [Akers & Getz, 1993]. Models of chemosensory reception should be complex enough to simulate the inhibitory and synergistic effects observed in sensory neurons, and they must provide a means for parameter fitting. We want to introduce a computational model which is constructed analogously to the chemical reaction cascade in the sensory neuron. The model can be expressed as an ANN and all unknown parameters can be trained with a learning algorithm. A Model for Chemosensory Reception 2 63 THE RECEPTOR TRANSDUCER MODEL The first step of odor reception is done by receptor proteins located at the cell membrane. There may be many receptor protein types in sensory cells at different concentrations and with different sensitivity to various odors. There is the possibility for different odors ligands Ai to react with a receptor protein Rj, but it is also possible for a single odor to react with different receptor proteins. The second step is the activation of second messengers. Ennis proposed a modelling of these complex reactions by a reaction step of activated odor-receptor complexes with transducer mechanisms [Ennis, 1991]. These transducers are a simplification of the second messengers processes. In Ennis' model transducers and receptor proteins are odor specific. We generalize Ennis' model by introducing transducer mechanisms T" that can be activated by odor-receptor complexes, and as with odors and receptor proteins we allow receptor proteins and transducers to react in any combination. Receptor proteins and transducer proteins are not required to be odor specific. The kinetics of the two reactions are given by + Rj AiRj + T" Ai ~ AiRj ~ (1) AiRjT". In a first reaction odor ligands Ai bind to receptor proteins Rj and build odor-receptor complexes AiRj, which can activate transducer mechanisms T" in a second reaction. Affinities lcij and Ij" describe the possibility of reactions between odor ligands Ai and receptor proteins Rj or between odor-receptor complexes Ai Rj and transducers T", respectively. The mass action equations are [AiRj] = lcij[Ad[Rj] (2) [Ai RjT,,] = Ij,,[AiRj][T,,]. The binding of odor-receptor complexes with transducer mechanisms is not dependent on the specific odor which is bound to the receptor protein, i.e. Ij" does not depend on i. It is only necessary that the receptor protein is bound. A sensory neuron can now be defined by the total concentration (or amount) of receptor proteins [H] and transducers [1']. The total concentration of either type corresponds to the sum of the free sites and the bound sites: [Hj] [Rj] + I)AiRj] (3) [T,,] + I)AiRjT,,] .1 (4) i,j Activated transducer mechanisms may elicit an excitatory or inhibitory effect depending on the kind of ion channel they open. Thus we divide the transducers T" into two types: inhibitory and excitatory transducers. With 8" = 1 {+ -1 , if transducer T" is excitatory , if transducer T" is inhibitory (5) IWe use the simplification [flj] = [Rj] + L:i[AiRJ' ] instead of [flj] = [Rj] + L:.[AiRj] L:i,k [AiRjTk], which is sufficient for [flj] > [tk], see also [Malaka & Ragg, 1993]. + 64 Rainer Malaka, Thomas Ragg, Martin Hammer Figure 2: ANN equivalent to the full receptor-transducer model. The input layer corresponds to the concentration of odor ligands [Ad, the first hidden layer to activated receptor protein types, the second to activated transducer mechanisms. The output neuron computes the effect E of the sensory cell. the effect can be set to the sum of all activated excitatory transducers minus the sum of all inhibitory transducers relative to the total amount of transducers. An additive constant (J is used to model spontaneous reactions. With this the effect of an odor can be set to (6) With Eqs.(2,4) and the hyperbolic function hyp(x) Eq.(6) can be reformulated to E= 1 A 2::L[Tk] '" = x/(l + x) the effect E defined in 2: hyp (2: Ijk[AiRj ]) 15 k ['h] L ?. '" t ,) + (J . (7) Analogously, we eliminate [AiRj] and [Rj): E= 1 A Lk[n] 2:hyp (2:Ijk[Rj]hYP(2:kij[Ad)) 8k[Tk]+(J k j . (8) i Equation (8) can now be regarded as an ANN with 4 feed-forward layers. The concentrations of the odor ligands [Ad represent the input layer, the two hidden layers correspond to activated receptor proteins and activated transducers, and one output element in layer 4 represents the effect caused by the input odor. The weight between the i-th element of the input layer to the j-th element of the first hidden layer is kij and from there to the k-th neuron of the second hidden layer Ij k [Hj]. The weight from element k of hidden layer 2 to the output element is 15k [1'k]/ 2::k[1'k]. The adaptive elements ofthe hidden layers have the hyperbolic activation functions hypo The network structure is shown in Figure 2. A Model for Chemosensory Reception 65 6 5 4 3 6 10 recept or protein types 20 50 mecanisms Figure 3: Mean error in spikes per output neuron for the model with different network sizes. Network sizes are varied in the number ofreceptor protein types and the number of transducing mechanisms. 3 SIMULATION RESULTS Applying learning algorithms like backpropagation or RProp to the model network, it is possible to find parameter settings for optimal (or local optimal) simulations of chemosensory cell responses with given response characteristics. In our simulations the best training results were achieved by using the fast learning algorithm RProp, which is an imprOVed version of backpropagation [Riedmiller & Braun, 1993]. For our simulations we used extracellular recordings made by Akers and Getz from single sensilla placodes of honey bee workers applying different stimuli and their binary mixtures to the antenna (see [Akers & Getz, 1992] for material and methods). The data set for training the ANNs consists of responses of 54 subplacodes to the four odors, geraniol, citral, limonene, linalool, their binary mixtures, and a mixture of all of four odors each at two concentration levels and to a blank stimulus, i.e. 23 responses to different odor stimulations for each subplacode. In a series of training runs with varying numbers of receptor protein types and transducer types the full model was trained to fit the data set. The networks were able to simulate the responses of the subplacodes, dependent on the network size. The size of the first hidden layer corresponds to the number of receptor protein types (R) in the model, the size of the second hidden layer corresponds to the number of transducing mechanisms (T). Figure 3 shows the mean error per output neuron in spikes for all combinations of one to six receptor types and one to six transducer mechanisms and for combinations with ten, twenty and fifty receptor protein types. The mean response over all subplacode responses is 18.15 spikes. The best results with errors less than two spikes per response were achieved with models with at least three receptor protein types and at least three transducer mechanisms. A model with only two transducer types is not sufficient to simulate the data. For generalization tests we generated a larger pattern set with our model. This training set 66 Rainer Malaka, Thomas Ragg, Martin Hammer spi kes spi kes Figure 4: Simulation results of our model (a.b) and the Ennis model (c.d). The responses of simulated sensory cells is given in spikes. The left column (a,c) represents receptor neuron responses to mixtures of geraniol and citral, the right column (b,d) represents sensory cell responses to mixtures of limonene and linalool. The concentrations of the odorants are depicted on a logarithmic scale from 2- 5 to 26 micrograms (0.03 to 64 micrograms). Measurement points and deviations from simulated data are given by crosses in the diagrams. was divided in a set of 23 training patterns and 88 test patterns. The training set had the same structure as the experimental data. Training of new randomly initialized networks provided a mean error on the test set that was approximately 1.6 times higher than on the training set. An overfitting effect was not observable during the training sequence of 10000 A Model for Chemosensory Reception 67 learning epochs. It is also possible to transform many other models for chemosensory perception into ANN s. We fitted the stimulus summation model and the stimulus substitution model [Carr & Derby, 1986] as well as the models proposed by Ennis [Ennis, 1991]. All of the other models were not able to reproduce the complex response functions observed in honey bee sensory neurons. Some of them are able to simulate synergistic responses to binary mixtures, but none were able to produce inhibitory effects. Figure 4 shows the simulation of a sensory neuron that shows very similar spike rates for the single odors geraniol and citral and to their binary mixture at the same concentration, while the mixture interaction of limonene and linalool shows a strong synergistic effect, i.e. the response to mixture of both odors is much higher than the responses to the single odors. As shown in Figure 4a) and b) our model is able to simulate this behavior, while the Ennis model is not sufficient to show the two different types of interaction for the binary mixtures geraniol-citral and limonene-linalool, as shown in Figure 4c) and d). The error for the Ennis model is greater than four spikes per output neuron and the error for our model with six receptor types and four transducer mechanisms is smaller than one spike per output neuron. The stimulus summation and stimulus substitution model have very similar results as the Ennis model, Figure 4 e) and t) show the simulation of the stimulus summation. 4 CONCLUSIONS Artificial neural networks are a powerful tool for the simulation of the responses of chemosensory cells. The use of ANNs is consistent with theoretical modelings. Many previously proposed models are expressible as ANNs. The new receptor transducer model described in this paper is also expressible as an ANN. The use of learning algorithms is a means to fit parameters for the simulation with given experimental response data. With this method it is possible to create simulation models of chemosensory cells, that can be used in further modelings of olfactory and chemosensory systems. Applying data from honey bee placode recordings we could also investigate the necessary complexity of chemosensory models. It could be shown that only the full receptor transducer model is able to simulate the complex response characteristics observed in honey bee chemosensory cells. Most other models can show only low synergistic mixture interactions and none of the other models is able to simulate inhibitory effects in odor perception. The found parameters of the ANN do not have to correspond to physiological entities, such as affinities between molecules. The learning or parameter fitting optimizes the parameters in a way that the difference between experimental data and simulation results is minimized. If there are several solutions to this task, one solution will be found, which might differ from the actual values. But it can be said, that a model is not sufficient if the learning algorithm is not able to fit the experimental data This implies that the smallest model, which is able to simulate the given data covers the minimum of complexity necessary. For honey bees this means that a competitive receptor transducer model is necessary with at least two transducer mechanisms and three receptor protein types. Any other model, such as the stimulus summation model, the stimulus substitution model and the Ennis model, is not sufficient. The model is not restricted to insect olfactory receptor neurons and can also be applied to many types of olfactory or gustatory receptor neurons in invertebrates and vertebrates. 68 Rainer Malaka, Thomas Ragg, Martin Hammer Acknowledgments We want to thank Pat Akers and Wayne Getz for giving us subplacode response data to train the ANNs used in our model, Heinrich Braun and Wayne Getz for fruitful discussions on our work. This work was supported by grants of the Deutsche Forschungsgemeinschaft (DFG), SPP Physiologie und Theorie neuronaler Netze, and the State of Baden-WUrttemberg. References [Akers & Getz, 1992] R.P. Akers & W.M. Getz. A test of identified response classes among olfactory receptor neurons in the honeybee worker. Chemical Senses, 17(2):191-209, 1992. [Akers & Getz, 1993] RP. Akers & W.M. Getz. Response of olfactory receptor neurons in honey bees to odorants and their binary mixtures. 1. Compo Physiol. A, 173:169-185, 1993. [Breer & Boekhoff, 1992] H. Breer & I. Boekhoff. Second messenger signalling in olfaction. Current Opinion in Neurobiology, 2:439-443,1992. [Breer et al., 1989] H. Breer, I. Boekhoff, J. Strotmann, K. Rarning, & E. Tareilus. Molecular elements of olfactory signal transduction in insect antennae. In D. Schild, editor, Chemosensory Information Processing, pages 75-86. Springer, 1989. [Carr & Derby, 1986] W.E.S . Carr & C.D. Derby. Chemically stimulated feeding behavior in marine animals: the importance of chemical mixtures and the involvement of mixture interactions. 1.Chem.Ecol., 12:987-1009,1986. [Ennis,1991] D.M. Ennis. Molecular mixture models based on competitive and noncompetitive agonism. Chemical Senses, 16(1):1-17,1991. [Malaka & Ragg, 1993] R. Malaka & T. Ragg. Models for chemosensory receptors: An approach using artificial neural networks. Interner Bericht 18/93, Institut fUr Logik, Komplexitlit und Deduktionssysteme, Universitlit Karlsruhe, 1993. [Riedmiller & Braun, 1993] M. Riedmiller & H. Braun. A direct adaptive method for faster backpropagation learning: The rprop algorithm. In Proceedings of the ICNN, 1993. [Shepherd, 1991] G.M. Shepherd. Computational structure of the olfactory system. In J.L. Davis & H. Eichenbaum, editors, Olfaction - A Model System for Computational Neuroscience, chapter 1, pages 3-41. MIT Press, 1991. [Yogt et al., 1989] RG. Yogt, R Rybczynski, & M.R. Lerner. The biochemistry of odorant reception and transduction. In D. Schild, editor, Chemosensory Information Processing, pages 33-76. Springer, 1989.
935 |@word version:1 hyperpolarized:1 vogt:1 open:1 simulation:12 minus:1 substitution:3 series:1 reaction:13 blank:1 current:1 activation:3 must:1 physiol:1 additive:1 recept:1 alone:1 signalling:1 marine:1 compo:1 provides:1 simpler:1 mathematical:1 constructed:1 direct:1 transducer:31 consists:1 fitting:3 olfactory:7 introduce:1 behavior:2 odorants:2 schild:2 actual:1 vertebrate:2 provided:1 deutsche:1 mass:1 kind:1 depolarization:1 developed:1 freie:1 braun:4 honey:7 wayne:2 conductivity:2 grant:1 bind:2 local:1 receptor:45 approximately:1 reception:8 might:1 physiologie:1 acknowledgment:1 backpropagation:3 riedmiller:3 elicit:1 cascade:4 hyperbolic:2 protein:32 synergistic:6 applying:4 equivalent:1 fruitful:1 react:3 regarded:1 spontaneous:1 flj:3 element:7 located:1 observed:5 getz:9 und:3 complexity:2 heinrich:1 trained:3 depend:1 po:1 various:1 chapter:1 train:1 fast:1 describe:1 activate:4 artificial:3 larger:1 transform:1 antenna:2 sequence:1 interaction:6 rapidly:1 produce:1 tk:3 depending:1 ij:4 eq:2 strong:1 implies:1 differ:1 hammer:5 opinion:1 material:1 feeding:1 generalization:1 icnn:1 summation:4 kinetics:2 ecol:1 biochemistry:1 a2:3 smallest:1 create:1 tool:1 mit:1 activates:1 hj:2 varying:1 rainer:5 ira:2 fur:3 modelling:1 camp:1 rebuild:1 dependent:2 eliminate:1 hidden:8 expressible:2 transformed:1 reproduce:1 germany:2 among:1 deduktionssysteme:1 insect:2 animal:2 represents:3 minimized:1 stimulus:9 randomly:1 lerner:1 dfg:1 olfaction:2 conductance:1 hyp:4 highly:1 possibility:2 investigate:1 mixture:18 activated:9 sens:2 fu:1 worker:2 necessary:4 institut:3 divide:1 initialized:1 theoretical:1 fitted:1 kij:2 column:2 modeling:2 cover:1 introducing:1 deviation:1 recognizing:1 universitat:2 neurobiologie:1 sensitivity:1 analogously:2 baden:1 potential:4 de:3 caused:1 ad:4 ennis:13 competitive:2 capability:1 characteristic:2 correspond:2 ofthe:1 generalize:1 rjt:1 ionic:2 none:2 anns:4 messenger:12 reach:1 akers:8 rprop:3 derby:3 feed:2 higher:3 response:29 improved:1 formulation:1 done:1 box:1 karlsruhe:3 effect:14 chemical:5 odorant:1 during:1 davis:1 presenting:1 demonstrate:1 carr:3 volatile:2 stimulation:1 measurement:1 ai:10 had:1 surface:1 involvement:1 optimizes:1 compound:1 binary:7 minimum:1 greater:1 signal:1 full:3 rj:11 ip3:2 faster:1 cross:1 divided:1 molecular:2 represent:1 achieved:2 cell:13 ion:6 schematically:1 want:2 diagram:1 uka:2 fifty:1 depolarized:1 shepherd:3 recording:3 forschungsgemeinschaft:1 enough:1 fit:3 spi:2 identified:1 six:3 reformulated:1 cause:2 action:5 amount:2 ten:1 inhibitory:9 neuroscience:1 per:5 logik:2 four:4 changing:1 kes:2 sum:4 run:1 powerful:2 chemically:1 layer:13 bound:4 simplification:2 influx:2 invertebrate:2 simulate:8 martin:5 extracellular:1 eichenbaum:1 combination:3 chemosensory:19 airj:10 membrane:2 smaller:2 restricted:1 honeybee:1 equation:2 previously:1 mechanism:12 odor:38 rp:1 thomas:5 a2l:1 giving:1 build:2 occurs:2 spike:8 concentration:12 said:1 affinity:2 thank:1 berlin:3 simulated:2 entity:1 mail:2 terrestrial:1 theorie:1 rise:1 unknown:1 gated:2 twenty:1 neuron:20 pat:1 neurobiology:1 varied:1 iwe:1 required:1 able:11 pattern:3 perception:2 spp:1 komplexitat:1 transducing:2 lk:1 mediated:1 epoch:1 bee:8 relative:2 generation:2 sufficient:6 consistent:1 editor:3 excitatory:4 changed:1 supported:1 free:1 allow:1 computes:1 sensory:12 forward:2 made:1 adaptive:2 observable:1 overfitting:1 gustatory:1 stimulated:1 channel:6 molecule:7 complex:10 ragg:9 site:3 tl:1 transduction:2 lcij:2 specific:3 substance:1 physiological:1 hypo:1 importance:1 rg:1 depicted:1 logarithmic:1 expressed:1 binding:2 ligand:5 springer:2 corresponds:4 ann:7 change:1 total:3 experimental:6 ijk:2 chem:1
7,174
936
SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu Abstract A self-organizing neural network for sequence classification called SARDNET is described and analyzed experimentally. SARDNET extends the Kohonen Feature Map architecture with activation retention and decay in order to create unique distributed response patterns for different sequences. SARDNET yields extremely dense yet descriptive representations of sequential input in very few training iterations. The network has proven successful on mapping arbitrary sequences of binary and real numbers, as well as phonemic representations of English words. Potential applications include isolated spoken word recognition and cognitive science models of sequence processing. 1 INTRODUCTION While neural networks have proved a good tool for processing static patterns, classifying sequential information has remained a challenging task. The problem involves recognizing patterns in a time series of vectors, which requires forming a good internal representation for the sequences. Several researchers have proposed extending the self-organizing feature map (Kohonen 1989, 1990), a highly successful static pattern classification method, to sequential information (Kangas 1991; Samarabandu and Jakubowicz 1990; Scholtes 1991). Below, three of the most recent of these networks are briefly described. The remainder of the paper focuses on a new architecture designed to overcome the shortcomings of these approaches. 578 Daniel L. James, Risto Miikkulainen Recently, Chappel and Taylor (1993) proposed the Temporal Kohonen Map (TKM) architecture for classifying sequences. The TKM keeps track of the activation history of each node by updating a value called leaky integrator potential, inspired by the membrane potential in biological neural systems. The activity of a node depends both on the current input vector and the previous input vectors, represented by the node's potential. A given sequence is processed by mapping one vector at a time, and the last winning node serves to represent the entire sequence. This way, there needs to be a separate node for every possible sequence, which is a disadvantage when the number of sequences to be classified is large. The TKM also suffers from loss of context. Which node wins depends almost entirely upon the most recent input vectors. For example, the string baaaa would most likely map to the same node as aaaaa, making the approach applicable only to short sequences. The SOFM-S network proposed by van Harmelen (1993) extends TKM such that the activity of each map node depends on the current input vector and the past activation of all map nodes. The SOFM-S is an improvement of TKM in that contextual information is not lost as quickly, but it still uses a single node to represent a sequence. The TRACE feature map (Zandhuis 1992) has two feature map layers. The first layer is a topological map of the individual input vectors , and is used to generate a trace (i.e. path) of the input sequence on the map. The second layer then maps the trace pattern to a single node . In TRACE, the sequences are represented by distributed patterns on the first layer, potentially allowing for larger capacity, but it is difficult to encode sequences where the same vectors repeat, such as baaaa. All a-vectors would be mapped on the same unit in the first layer, and any number of a-vectors would be indistinguishable. The architecture described in this paper, SARDNET (Sequential Activation Retention and Decay NETwork), also uses a subset of map nodes to represent the sequence of vectors. Such a distributed approach allows a large number of representations be "packed" into a small map-like sardines. In the following sections, we will examine how SARDNET differs from conventional self-organizing maps and how it can be used to represent and classify a large number of complex sequences. 2 THE SARDNET ARCHITECTURE Input to SARDNET consists of a sequence of n-dimensional vectors S = V I, V 2 , V 3 , ... , VI (figure 1) . The components of each vector are real values in the interval [0,1]. For example, each vector might represent a sample of a speech signal in n different frequencies, and the entire sequence might constitute a spoken word. The SARDNET input layer consists of n nodes, one for each component in the input vector, and their values are denoted as A = (aI, a2, a3, ... , an). The map consists of m x m nodes with activation Ojk , 1 ~ j, k ~ m. Each node has an n-dimensional input weight vector Wjk, which determines the node's response to the input activation. In a conventional feature map network as well as in SARDNET, each input vector is mapped on a particular unit on the map, called the winner or the maximally responding unit . In SARDNET, however, once a node wins an input, it is made SARDNET: A Self-Organizing Feature Map for Sequences 579 Sequence of Input vectors S V1 V2 V3 V4 V, ---II Previous winners Input weight vector wJk.l Winning unit jlc Figure 1: The SARDNET architecture. A sequence of input vectors activates units on the map one at a time. The past winners are excluded from further competition, and their activation is decayed gradually to indicate position in the sequence. INITIALIZATION: Clear all map nodes to zero. MAIN LOOP: While not end of seihence 1. Find unactivated weight vector t at best matches the input. 2. Assign 1.0 activation to that unit. 3. Adjust weight vectors of the nodes in the neighborhood. 4. Exclude the winning unit from subse~ent competition. S. Decrement activation values for all ot er active nodes. RESULT: Sequence representation = activated nodes ordered by activation values Table 1: The SARDNET training algorithm. uneligible to respond to the subsequent inputs in the sequence. This way a different map node is allocated for every vector in the sequence. As more vectors come in, the activation of the previous winners decays. In other words, each sequence of length 1 is represented by 1 active nodes on the map, with their activity indicating the order in which they were activated. The algorithm is summarized in table 1. Assume the maximum length ofthe sequences we wish to classify is I, and each input vector component can take on p possible values. Since there are pn possible input vectors, Ipn map nodes are needed to represent all possible vectors in all possible positions in the sequence, and a distributed pattern over the Ipn nodes can be used to represent all pnl different sequences. This approach offers a significant advantage over methods in which pnl nodes would be required for pnl sequences. The specific computations of the SARDNET algorithm are as follows: The winning node (j, k) in each iteration is determined by the Euclidean distance Djk of the 580 Daniel L. James, Risto Miikkulainen input vector A and the node 's weight vector W j k: n Djk = 1)Wjk ,i - a;)2. (1) i=O The unit with the smallest distance is selected as the winner and activated with 1.0. The weights of this node and all nodes in its neighborhood are changed according to the standard feature map adaptation rule: (2) where a denotes the learning rate. As usual, the neighborhood starts out large and is gradually decreased as the map becomes more ordered. As the last step in processing an input vector , the activation 7]jk of all active units in the map are decayed proportional to the decay parameter d: O<d<1. (3) As in the standard feature map, as the weight vectors adapt, input vectors gradually become encoded in the weight vectors of the winning units. Because weights are changed in local neighborhoods , neighboring weight vectors are forced to become as similar as possible , and eventually the network forms a topological layout of the input vector space. In SARDNET, however , if an input vector occurs multiple times in the same input sequence, it will be represented multiple times on the map as well. In other words, the map representation expands those areas of the input space that are visited most often during an input sequence. 3 EXPERIMENTS SARDNET has proven successful in learning and recognizing arbitrary sequences of binary and real numbers , as well as sequences of phonemic representations for English words. This section presents experiments on mapping three-syllable words. This dat a was selected because it shows how SARDNET can be applied to complex input derived from a real-world task . 3.1 INPUT DATA The phonemic word representations were obtained from the CELEX database of the Max Planck Institute for Psycholinguistics and converted into International Phonetic Alphabet (IPA)-compliant representation, which better describes similarities among the phonemes. The words vary from five to twelve phonemes in length. Each phoneme is represented by five values: place, manner, sound, chromacity and sonority. For example, the consonant p is represented by a single vector (bilabial, stop, unvoiced, nil , nil) , or in terms of real numbers, (.125, .167, .750,0 , 0). The diphthong sound ai as in "buy" , is represented by the two vectors (nil , vowel, voiced, front, low) and (nil, vowel , voiced , front-center, hi-mid), or in real numbers , (0, 1, .25, .2, 1) and (0, 1, .25, .4, .286). There are a total of 43 phonemes in this data set, including 23 consonants and 20 vowels. To represent all phonemic sequences of length 12, TKM and SOFM-S would SARDNET: A Self-Organizing Feature Map for Sequences 581 o.ee 0.118 0.87 0.118 0.85 .......- - Figure 2: Accuracy of SARDNET for different map and data set sizes. The accuracy is measured as a percentage of unique representations out of all word sequences. need to have 45 12 ~ 6.919 map nodes, whereas SARDNET would need only 45 x 12 = 540 nodes . Of course, only a very small subset of the possible sequences actually occur in the data. Three data sets consisting of 713 , 988, and 1628 words were used in the experiments. If the maximum number of occurrences of phoneme i in any single sequence is Cj I then the number of nodes SARDNET needs is C = L:~o Cj I where N is the number of phonemes. This number of nodes will allow SARDNET to map each phoneme in each sequence to a unit with an exact representation of that phoneme in its weights . Calculated this way, SARDNET should scale up very well with the number of words: it would need 81 nodes for representing the 713 word set , 84 for the 988 set and 88 for the 1628 set. 3.2 DENSENESS AND ACCURACY A series of experiments with the above three data sets and maps of 16 to 81 nodes were run to see how accurately SARDNET can represent the sequences. Self-organization was quite fast: each simulation took only about 10 epochs, with a = 0.45 and the neighborhood radius decreasing gradually from 5-1 to zero. Figure 2 shows the percentage of unique representations for each data set and map SIze. SARDNET shows remarkable representational power: accuracy for all sets is better than 97.7%, and SARDNET manages to pack 1592 unique representations even on the smallest 16-node map. Even when there are not enough units to represent each phoneme in each sequence exactly, the map is sometimes able to "reuse" units to represent multiple similar phonemes. For example, assume units with exact representations for the phonemes a and b exist somewhere on the map, and the input data does not contain pairs of sequences such as aba-abb, in which it is crucial to distinguished the second a from the second b. In this case, the second occurrence of both phonemes could be represented by the same unit with a weight vector that is the average of a and b. This is exactly what the map is doing: it is finding the most descriptive representation of the data, given the available resources. 582 Daniel L. James, Risto Miikkulainen Note that it would be possible to determine the needed C = L:f:o Cj phoneme representation vectors directly from the input data set, and without any learning or a map structure at all, establish distributed representations on these vectors with the SARDNET algorithm. However, feature map learning is necessary ifthe number of available representation vectors is less than C. The topological organization of the map allows finding a good set of reusable vectors that can stand for different phonemes in different sequences, making the representation more efficient. 3.3 REPRESENTING SIMILARITY Not only are the representations densely packed on the map, they are also descriptive in the sense that similar sequences have similar representations. Figure 3 shows the final activation patterns on the 36-unit, 713-word map for six example words. The first two words, "misplacement" and "displacement," sound very similar, and are represented by very similar patterns on the map. Because there is only one m in "displacement" , it is mapped on the same unit as the initial m of "misplacement." Note that the two IDS are mapped next to each other, indicating that the map is indeed topological, and small changes in the input cause only small changes in the map representation. Note also how the units in this small map are reused to represent several different phonemes in different contexts. The other examples in figure 3 display different types of similarities with "misplacement". The third word, "miscarried", also begins with "mis", and shares that subpart of the representation exactly. Similarly, "repayment" shares a similar tail and "pessimist" the subsequence "mis" in a different part or the word. Because they appear in a different context, these subsequences are mapped on slightly different units, but still very close to their positions with "misplacement." The last word, "burundi" sounds very different, as its representation on the map indicates. Such descriptive representations are important when the map has to represent information that is incomplete or corrupted with noise. Small changes in the input sequence cause small changes in the pattern, and the sequence can still be recognized. This property should turn out extremely important in real-world applications of SARDNET, as well as in cognitive science models where confusing similar patterns with each other is often plausible behavior. 4 DISCUSSION AND FUTURE RESEARCH Because the sequence representations on the map are distributed , the number of possible sequences that can be represented in m units is exponential in m, instead of linear as in most previous sequential feature map architectures. This denseness together with the tendency to map similar sequences to similar representations should turn out useful in real-world applications, which often require scale-up to large and noisy data sets. For example, SARDNET could form the core of an isolated word recognition system. The word input would be encoded in durationnormalized sequences of sound samples such as a string of phonemes, or perhaps representations of salient transitions in the speech signal. It might also be possible to modify SARDNET to form a more continuous trajectory on the map so that SARDNET itself would take care of variability in word duration. For example, a SARDNEf: A Self-Organizing Feature Map for Sequences (a) ~~t 583 (b) (c) (e) (f) ? ript.ItDt (d) Figure 3: Example map representations. sequence of redundant inputs could be reduced to a single node if all these inputs fall within the same neighborhood. Even though the sequence representations are dense, they are also descriptive. Category memberships are measured not by labels of the maximally responding units, but by the differences in the response patterns themselves. This sort of distributed representation should be useful in cognitive systems where sequential input must be mapped to an internal static representation for later retrieval and manipulation. Similarity-based reasoning on sequences should be easy to implement, and the sequence can be easily recreated from the activity pattern on the map. Given part of a sequence, SARDNET may also be modified to predict the rest of the sequence. This can be done by adding lateral connections between the nodes in the map layer. The lateral connections between successive winners would be strengthened during training. Thus, given part of a sequence, one could follow the strongest lateral connections to complete the sequence. 584 5 Daniel L. James, Risto Miikkulainen CONCLUSION SARDNET is a novel feature map architecture for classifying sequences of input vectors. Each sequence is mapped on a distributed representation on the map, making it possible to pack a remarkable large number of category representations on a small feature map . The representations are not only dense, they also represent the similarities of the sequences, which should turn out useful in cognitive science as well as real-world applications of the architecture. Acknowledgments Thanks to Jon Hilbert for converting CELEX data into the International Phonetic Alphabet format used in the experiments. This research was supported in part by the National Science Foundation under grant #IRI-9309273. References Chappel, G. J., and Taylor, J. G. (1993). The temporal Kohonen map. Neural Networks, 6:441-445. Kangas, J. (1991). Time-dependent self-organizing maps for speech recognition. In Proceedings of the International Conference on Artificial Neural Networks (Espoo, Finland), 1591-1594. Amsterdam; New York: North-Holland. Kohonen, T. (1989). Self-Organization and Associative Memory. Berlin; Heidelberg; New York: Springer. Third edition. Kohonen, T . (1990) . The self-organizing map. Proceedings of the IEEE, 78:14641480. Samarabandu, J. K., and Jakubowicz, O. G. (1990). Principles of sequential feature maps in multi-level problems. In Proceedings of the International Joint Conference on Neural Networks (Washington, DC), vol. II, 683-686. Hillsdale, NJ: Erlbaum. Scholtes, J. C. (1991). Recurrent Kohonen self-organization in natural language processing. In Proceedings of the International Conference on Artificial Neural Networks (Espoo, Finland), 1751-1754. Amsterdam; New York: NorthHolland. van Harmelen, H. (1993). Time dependent self-organizing feature map for speech recognition. Master's thesis, University of Twente, Enschede, the Netherlands. Zandhuis, J. A. (1992). Storing sequential data in self-organizing feature maps. Internal Report MPI-NL-TG-4/92, Max-Planck-Institute fur Psycholinguistik, Nijmegen, the Netherlands.
936 |@word briefly:1 reused:1 risto:6 simulation:1 initial:1 series:2 daniel:5 past:2 current:2 contextual:1 activation:13 yet:1 must:1 subsequent:1 designed:1 selected:2 short:1 core:1 node:39 successive:1 five:2 scholtes:2 become:2 consists:3 manner:1 indeed:1 behavior:1 themselves:1 examine:1 multi:1 integrator:1 inspired:1 decreasing:1 becomes:1 begin:1 what:1 string:2 spoken:2 finding:2 nj:1 temporal:2 every:2 expands:1 exactly:3 unit:21 grant:1 appear:1 planck:2 retention:2 local:1 modify:1 id:1 path:1 might:3 initialization:1 challenging:1 unique:4 acknowledgment:1 lost:1 implement:1 differs:1 displacement:2 area:1 word:22 close:1 context:3 conventional:2 map:68 center:1 layout:1 iri:1 duration:1 rule:1 exact:2 us:2 recognition:4 jk:1 updating:1 database:1 upon:1 easily:1 joint:1 psycholinguistics:1 represented:10 tx:1 alphabet:2 forced:1 fast:1 shortcoming:1 artificial:2 neighborhood:6 quite:1 encoded:2 larger:1 plausible:1 noisy:1 itself:1 final:1 associative:1 sequence:65 descriptive:5 advantage:1 ifthe:1 took:1 remainder:1 adaptation:1 kohonen:7 neighboring:1 loop:1 organizing:11 representational:1 competition:2 wjk:3 ent:1 extending:1 recurrent:1 measured:2 phonemic:4 c:1 involves:1 indicate:1 come:1 radius:1 repayment:1 hillsdale:1 require:1 assign:1 biological:1 mapping:3 predict:1 vary:1 finland:2 a2:1 smallest:2 applicable:1 label:1 visited:1 utexas:1 create:1 tool:1 activates:1 modified:1 pn:1 encode:1 derived:1 focus:1 improvement:1 fur:1 indicates:1 sense:1 dependent:2 membership:1 entire:2 djk:2 classification:2 among:1 espoo:2 denoted:1 once:1 washington:1 aba:1 jon:1 future:1 report:1 few:1 densely:1 national:1 individual:1 consisting:1 vowel:3 organization:4 highly:1 adjust:1 analyzed:1 nl:1 activated:3 necessary:1 incomplete:1 taylor:2 euclidean:1 ipn:2 isolated:2 classify:2 disadvantage:1 tg:1 subset:2 recognizing:2 successful:3 erlbaum:1 front:2 corrupted:1 thanks:1 decayed:2 international:5 twelve:1 v4:1 compliant:1 recreated:1 together:1 quickly:1 thesis:1 cognitive:4 potential:4 exclude:1 converted:1 summarized:1 north:1 depends:3 vi:1 later:1 doing:1 start:1 sort:1 voiced:2 accuracy:4 phoneme:16 yield:1 ofthe:1 accurately:1 manages:1 trajectory:1 researcher:1 history:1 classified:1 strongest:1 suffers:1 frequency:1 james:5 mi:2 static:3 stop:1 proved:1 cj:3 hilbert:1 actually:1 follow:1 response:3 maximally:2 done:1 though:1 perhaps:1 contain:1 excluded:1 indistinguishable:1 during:2 self:14 abb:1 mpi:1 sofm:3 complete:1 reasoning:1 novel:1 recently:1 winner:6 tail:1 jlc:1 significant:1 ai:2 similarly:1 language:1 similarity:5 recent:2 manipulation:1 phonetic:2 binary:2 care:1 converting:1 recognized:1 determine:1 v3:1 redundant:1 signal:2 ii:2 multiple:3 sound:5 match:1 adapt:1 offer:1 retrieval:1 iteration:2 represent:14 sometimes:1 whereas:1 interval:1 decreased:1 allocated:1 crucial:1 ot:1 rest:1 ee:1 enough:1 easy:1 architecture:9 texas:1 six:1 reuse:1 speech:4 york:3 cause:2 constitute:1 useful:3 clear:1 netherlands:2 mid:1 processed:1 category:2 reduced:1 generate:1 percentage:2 exist:1 ipa:1 track:1 subpart:1 vol:1 reusable:1 salient:1 v1:1 run:1 master:1 respond:1 sonority:1 extends:2 almost:1 place:1 confusing:1 entirely:1 layer:7 hi:1 syllable:1 display:1 topological:4 activity:4 occur:1 extremely:2 format:1 department:1 according:1 membrane:1 describes:1 slightly:1 making:3 gradually:4 resource:1 turn:3 eventually:1 needed:2 serf:1 end:1 available:2 v2:1 occurrence:2 distinguished:1 responding:2 denotes:1 include:1 somewhere:1 establish:1 dat:1 harmelen:2 occurs:1 usual:1 win:2 distance:2 jakubowicz:2 separate:1 mapped:7 capacity:1 lateral:3 berlin:1 length:4 difficult:1 potentially:1 trace:4 nijmegen:1 packed:2 allowing:1 unvoiced:1 variability:1 dc:1 kangas:2 arbitrary:2 pair:1 required:1 misplacement:4 connection:3 able:1 below:1 pattern:13 max:2 including:1 memory:1 power:1 natural:1 representing:2 epoch:1 loss:1 proportional:1 proven:2 remarkable:2 foundation:1 principle:1 classifying:3 share:2 storing:1 austin:2 course:1 changed:2 repeat:1 last:3 supported:1 english:2 denseness:2 allow:1 institute:2 fall:1 leaky:1 distributed:8 van:2 overcome:1 calculated:1 world:4 stand:1 transition:1 made:1 miikkulainen:5 keep:1 active:3 buy:1 consonant:2 subsequence:2 continuous:1 table:2 pack:2 northholland:1 heidelberg:1 complex:2 dense:3 main:1 decrement:1 noise:1 tkm:6 edition:1 celex:2 strengthened:1 position:3 wish:1 winning:5 exponential:1 third:2 remained:1 specific:1 er:1 decay:4 a3:1 sequential:8 adding:1 pnl:3 likely:1 forming:1 amsterdam:2 ordered:2 holland:1 springer:1 determines:1 experimentally:1 change:4 determined:1 called:3 nil:4 total:1 tendency:1 indicating:2 ojk:1 internal:3
7,175
938
Estimating Conditional Probability Densities for Periodic Variables Chris M Bishop and Claire Legleye Neural Computing Research Group Department of Computer Science and Applied Mathematics Aston University Birmingham, B4 7ET, U.K. c.m.bishop@aston.ac.uk Abstract Most of the common techniques for estimating conditional probability densities are inappropriate for applications involving periodic variables. In this paper we introduce three novel techniques for tackling such problems, and investigate their performance using synthetic data. We then apply these techniques to the problem of extracting the distribution of wind vector directions from radar scatterometer data gathered by a remote-sensing satellite. 1 INTRODUCTION Many applications of neural networks can be formulated in terms of a multi-variate non-linear mapping from an input vector x to a target vector t. A conventional neural network approach , based on least squares for example, leads to a network mapping which approximates the regression of t on x. A more complete description of the data can be obtained by estimating the conditional probability density of t, conditioned on x , which we write as p(tlx). Various techniques exist for modelling such densities when the target variables live in a Euclidean space. However, a number of potential applications involve angle-like output variables which are periodic on some finite interval (usually chosen to be (0,271")). For example , in Section 3 642 Chris M. Bishop, Claire Legleye we consider the problem of determining the wind direction (a periodic quantity) from radar scatterometer data obtained from remote sensing measurements. Most of the existing techniques for conditional density estimation cannot be applied in such cases. A common technique for unconditional density estimation is based on mixture models of the form m pet) = L Cki?i(t) (1) i=l where Cki are called mixing coefficients, and the kernel functions ?i(t) are frequently chosen to be Gaussians. Such models can be used as the basis of techniques for conditional density estimation by allowing the mixing coefficients, and any parameters governing the kernel functions, to be general functions of the input vector x. This can be achieved by relating these quantities to the outputs of a neural network which takes x as input, as shown in Figure 1. Such an approach forms the basis of conditional probability density n p(tlx) 1\ 1\ 1\ parameter (> vector U mixture model z neural network Figure 1: A general framework for conditional density estimation is obtained by using a feed-forward neural network whose outputs determine the parameters in a mixture density model. The mixture model then represents the conditional probability density of the target variables, conditioned on the input vector to the network. the 'mixture of experts' model (Jacobs et al., 1991) and has also been considered by a number of other authors (White, 1992; Bishop, 1994; Lui, 1994). In this paper we introduce three techniques for estimating conditional densities of periodic variables, based on extensions of the above formalism for Euclidean variables. Estimating Conditional Probability Densities for Periodic Variables 2 643 DENSITY ESTIMATION FOR PERIODIC VARIABLES In this section we consider three alternative approaches to estimating the conditional density p(Blx) of a periodic variable B, conditioned on an input vector x. They are based respectively on a transformation to an extended domain representation, the use of adaptive circular normal kernel functions, and the use of fixed circular normal kernels. 2.1 TRANSFORMATION TO AN EXTENDED VARIABLE DOMAIN The first technique which we consider involves finding a transformation from the periodic variable B E (0,27r) to a Euclidean variable X E (-00,00), such that standard techniques for conditional density estimation can be applied in X-space. In particular, we seek a conditional density function p(xlx) which is to be modelled using a conventional Gaussian mixture approach as described in Section 1. Consider the transformation 00 L p(Blx) = p(B + L27rlx) (2) L=-oo Then it is clear by construction that the density model on the left hand side satisfies the periodicity requirement p(B + 27rlx) = p(Blx). Furthermore, if the density function p(xlx) is normalized, then we have {21f Jo p(Blx) dB = (3) and so the corresponding periodic density p(Olx) will also be normalized. We now model the density function p(xlx) using a mixture of Gaussians of the form m p(xlx) = L fri(X)4>i(xlx) (4) i=l where the kernel functions are given by 1 ( {X - Xi(X)P) 4>i(xlx) = (27r)1/2Uj(x) exp 2u;(x) (5) and the parameters fri(X), Uj(x) and Xi(X) are determined by the outputs of a feed-forward network. In particular, the mixing coefficients frj(x) are governed by a 'softmax' activation function to ensure that they lie in the range (0,1) and sum to Chris M. Bishop, Claire Legleye 644 unity; the width parameters O'i(X) are given by the exponentials ofthe corresponding network outputs to ensure their positivity; and the basis function centres Xi(X) are given directly by network output variables. The network is trained by maximizing the likelihood function, evaluated for set of training data, with respect to the weights and biases in the network. For a training set consisting of N input vectors xn and corresponding targets (r, the likelihood is given by N .c = II peon Ixn)p(xn) (6) n=l where p(x) is the unconditional density of the input data. Rather than work with .c directly, it is convenient instead to minimize an error function given by the negative log of the likelihood. Making use of (2) we can write this in the form E = -In.c ~ - 2: In 2:P'(on + L271'1xn) n (7) L where we have dropped the term arising from p(x) since it is independent of the network weights. This expression is very similar to the one which arises if we perform density estimation on the real axis, except for the extra summation over L, which means that the data point on recurs at intervals of 271' along the x-axis. This is not equivalent simply to replicating the data, however, since the summation over L occurs inside the logarithm, rather than outside as with the summation over data points n. In a practical implementation, it is necessary to restrict the summation over L. For the results presented in the next section, this summation was taken over 7 complete periods of 271' spanning the range (-771', 771'). Since the Gaussians have exponentially decaying tails, this represents an extremely good approximation in almost all cases, provided we take care in initializing the network weights so that the Gaussian kernels lie in the central few periods. Derivatives of E with respect to the network weights can be computed using the rules of calculus, to give a modified form of back-propagation. These derivatives can then be used with standard optimization techniques to find a minimum of the error function. (The results presented in the next section were obtained using the BFGS quasi-Newton algorithm). 2.2 MIXTURES OF CIRCULAR NORMAL DENSITIES The second approach which we introduce is also based on a mixture of kernel functions of the form (1), but in this case the kernel functions themselves are periodic, thereby ensuring that the overall density function will be periodic. To motivate this approach, consider the problem of modelling the distribution of a velocity vector v in two dimensions (this arises, for example, in the application considered in Section 3). Since v lives in a Euclidean plane, we can model the density function p(v) using a mixture of conventional spherical Gaussian kernels, where each kernel has 645 Estimating Conditional Probability Densities for Periodic Variables the form ? (V x , Vy ) - flx F = 211"cr1 2 exp ( { Vx 2cr 2 {V y - fly 2cr 2 F) (8) where (v x , v y ) are the Cartesian components of v, and (flx, fly) are the components of the center I-' of the kernel. From this we can extract the conditional distribution of the polar angle () of the vector v, given a value for v = I/vll. This is easily done with the transformation Vx = v cos (), Vy = v sin (), and defining ()o to be the polar angle of 1-', so that flx = fl cos ()o and fly = fl sin ()o, where fl = 111-'11? This leads to a distribution which can be written in the form ?(()) = 1 (A) exp {A cos(() - ()on 211"10 (9) where the normalization coefficient has been expressed in terms of the zeroth order modified Bessel function of the first kind, Io(A) . The distribution (9) is known as a circular normal or von Mises distribution (Mardia, 1972) . The parameter A (which depends on v in our derivation) is analogous to the (inverse) variance parameter in a conventional normal distribution. Since (9) is periodic, we can construct a general representation for the conditional density of a periodic variable by considering a mixture of circular normal kernels, with parameters given by the outputs of a neural network. The weights of the network can again be determined by maximizing the likelihood function defined over a set of training data. 2.3 FIXED KERNELS The third approach introduced here is again based on a mixture model in which the kernel functions are periodic, but where the kernel parameters (specifying their width and location) are fixed. The only adaptive parameters are the mixing coefficients, which are again determined by the outputs of a feed-forward network having a softmax final-layer activation function. Here we consider a set of equally-spaced circular normal kernels in which the width parameters are chosen to give a moderate degree of overlap between the kernels so that the resulting representation for the density function will be reasonably smooth. Again, a maximum likelihood formalism is employed to train the network. Clearly a major drawback of fixed-kernel methods is that the number of kernels must grow exponentially with the dimensionality of the output space. For a single output variable, however , they can be regarded as practical techniques. 3 RESULTS In order to test and compare the methods introduced above, we first consider a simple problem involving synthetic data, for which the true underlying distribution function is known . This data set is intended to mimic the central properties of the real data to be discussed in the next section. It has a single input variable x and an output variable () which lies in the range (0,211"). The distribution of () is governed 646 Chris M. Bishop, Claire Leg/eye by a mixture of two triangular functions whose parameters (locations and widths) are functions of x. Here we present preliminary results from the application of the method introduced in section 2.1 (involving the transformation to Euclidean space) to this data. Figure 2 shows a plot of the reconstructed conditional density in both the extended X variable, and in the reconstructed polar variable (), for a particular value of the input variable x . 0.6 r-----r-------,~-_, x=0.5 0.5,..-----"'T""-----., -network -- - - true p(X) sinO 0.0 0.3 0.0 L..-_---L_"----'-----';.......oooooL..-.........- - I -21? 21? o x ~-~---+-----~ , -0.5 '--_ _ _ _...L..-_ _ _ _.... -0.5 0.0 cosO 0.5 Figure 2: The left hand plot shows the predicted density (solid curve) together with the true density (dashed curve) in the extended X space. The right hand plot shows the corresponding densities in the periodic () space. In both cases the input variable is fixed at x = 0.5. One of the original motivations for developing the techniques described in this paper was to provide an effective, principled approach to the analysis of radar scatterometer data from satellites such as the European Remote Sensing Satellite ERS-I. This satellite is equipped with three C-band radar antennae which measure the total backscattered power (called 0"0) along three directions relative to the satellite track, as shown in Figure 3. When the satellite passes over the ocean, the strengths of the backscattered signals are related to the surface ripples of the water (on length-scales of a few cm.) which in turn are determined by the low level winds. Extraction of the wind speed and direction from the radar signals represents an inverse problem which is typically multi-valued. For example, a wind direction of (}1 will give rise to similar radar signals to a wind direction of (}1 + 1r. Often, there are additional such 'aliases' at other angles . A conventional neural network approach to this problem , based on least-squares, would predict wind directions which were given by conditional averages of the target data. Since the average of several valid wind directions is typically not itself a valid direction, such an approach would clearly fail. Here we aim to extract the complete distribution of wind directions (as a function of the three 0"0 values and on the angle of incidence of the radar beam) and hence avoid Estimating Conditional Probability Densities for Periodic Variables 647 satellite fore beam 785km mid beam aft beam 5COkm Figure 3: Schematic illustration of the ERS-l satellite showing the footprints of the three radar scatterometers. such difficulties. This approach also provides the most complete information for the next stage of processing (not considered here) which is to 'de-alias' the wind directions to extract the most probable overall wind field. A large data set of ERS-l measurements, spanning a wide range of meteorological conditions, has been assembled by the European Space Agency in collaboration with the UK Meteorological Office. Labelling of the data set was performed using wind vectors from the Meteorological Office Numerical Weather Prediction code. An example of the results from the fixed-kernel method of Section 2.3 are presented in Figure 4. This clearly shows the existence of a primary alias at an angle of 1r relative to the principal direction, as well as secondary aliases at ?1r/2. Acknowledgements We are grateful to the European Space Agency and the UK Meteorological Office for making available the ERS-l data. We would also like to thank lain Strachan and Ian Kirk of AEA Technology for a number of useful discussions relating to the interpretation of this data. References Bishop C M (1994). Mixture density networks. Neural Computing Research Group Report, NCRG/4288, Department of Computer Science, Aston University, Birmingham, U.K. Jacobs R A, Jordan M I, Nowlan S J and Hinton G E (1991). Adaptive mixtures 648 Chris M. Bishop, Claire Leg/eye 0.5 ~--~---r---~--""'" 0.0 1--------.....;;::~~-__1 -0.5 -1.0 -1.5 ~ -1.5 _ _~_----II.....-_---I_ _- - J -1.0 -0.5 0.0 0.5 Figure 4: An example of the results obtained with the fixed-kernel method applied to data from the ERS-1 satellite. As well as the primary wind direction, there are aliases at 1r and ?1r/2. of local experts. Neural Computation, 3 79-87. Lui Y (1994) Robust parameter estimation and model selection for neural network regression. Advances in Neural Information Processing Systems 6 Morgan Kaufmann, 192-199 .. Mardia K V (1972) . Statistics of Directional Data. Academic Press, London. White H (1992). Parametric statistical estimation with artificial neural networks. University of California, San Diego, Technical Report.
938 |@word km:1 calculus:1 seek:1 jacob:2 thereby:1 solid:1 existing:1 legleye:3 incidence:1 nowlan:1 activation:2 tackling:1 written:1 must:1 aft:1 numerical:1 plot:3 plane:1 provides:1 location:2 along:2 inside:1 introduce:3 themselves:1 scatterometer:3 frequently:1 multi:2 spherical:1 inappropriate:1 considering:1 equipped:1 provided:1 estimating:8 underlying:1 kind:1 cm:1 finding:1 transformation:6 uk:3 dropped:1 local:1 io:1 zeroth:1 specifying:1 co:3 range:4 practical:2 footprint:1 weather:1 convenient:1 cannot:1 selection:1 live:1 conventional:5 equivalent:1 center:1 maximizing:2 rule:1 regarded:1 vll:1 analogous:1 target:5 construction:1 diego:1 velocity:1 coso:1 fly:3 initializing:1 remote:3 principled:1 agency:2 radar:8 trained:1 motivate:1 grateful:1 basis:3 easily:1 various:1 derivation:1 train:1 effective:1 london:1 artificial:1 outside:1 whose:2 valued:1 triangular:1 statistic:1 antenna:1 itself:1 final:1 mixing:4 description:1 requirement:1 satellite:9 ripple:1 oo:1 ac:1 predicted:1 involves:1 direction:13 drawback:1 vx:2 preliminary:1 probable:1 summation:5 extension:1 considered:3 normal:7 exp:3 mapping:2 predict:1 major:1 estimation:9 polar:3 birmingham:2 clearly:3 gaussian:3 aim:1 modified:2 rather:2 avoid:1 cr:2 office:3 modelling:2 likelihood:5 recurs:1 typically:2 quasi:1 overall:2 softmax:2 field:1 construct:1 having:1 extraction:1 represents:3 blx:4 mimic:1 report:2 few:2 intended:1 consisting:1 tlx:2 investigate:1 circular:6 mixture:15 unconditional:2 necessary:1 euclidean:5 logarithm:1 formalism:2 cr1:1 periodic:18 synthetic:2 density:35 cki:2 rlx:1 together:1 jo:1 von:1 central:2 again:4 positivity:1 expert:2 derivative:2 potential:1 bfgs:1 de:1 coefficient:5 depends:1 performed:1 wind:13 decaying:1 xlx:6 minimize:1 square:2 variance:1 kaufmann:1 gathered:1 ofthe:1 spaced:1 directional:1 modelled:1 fore:1 mi:1 dimensionality:1 back:1 feed:3 evaluated:1 done:1 furthermore:1 governing:1 stage:1 hand:3 propagation:1 meteorological:4 normalized:2 true:3 hence:1 white:2 sin:2 width:4 complete:4 novel:1 common:2 b4:1 exponentially:2 ncrg:1 tail:1 discussed:1 approximates:1 relating:2 interpretation:1 measurement:2 mathematics:1 centre:1 replicating:1 surface:1 moderate:1 life:1 morgan:1 minimum:1 additional:1 care:1 employed:1 determine:1 fri:2 period:2 bessel:1 dashed:1 ii:2 signal:3 smooth:1 technical:1 academic:1 equally:1 ensuring:1 schematic:1 involving:3 regression:2 prediction:1 kernel:21 normalization:1 achieved:1 beam:4 interval:2 grow:1 extra:1 pass:1 db:1 olx:1 jordan:1 extracting:1 variate:1 restrict:1 expression:1 aea:1 useful:1 clear:1 involve:1 mid:1 band:1 exist:1 vy:2 arising:1 track:1 write:2 group:2 lain:1 sum:1 angle:6 inverse:2 flx:3 almost:1 fl:3 layer:1 frj:1 strength:1 speed:1 extremely:1 department:2 developing:1 unity:1 making:2 leg:2 taken:1 turn:1 fail:1 available:1 gaussians:3 apply:1 ocean:1 alternative:1 existence:1 original:1 ensure:2 newton:1 uj:2 quantity:2 occurs:1 parametric:1 primary:2 thank:1 chris:5 spanning:2 pet:1 water:1 length:1 code:1 illustration:1 negative:1 rise:1 implementation:1 perform:1 allowing:1 i_:1 finite:1 defining:1 extended:4 hinton:1 introduced:3 california:1 assembled:1 usually:1 power:1 overlap:1 difficulty:1 ixn:1 aston:3 technology:1 eye:2 alias:5 axis:2 extract:3 acknowledgement:1 determining:1 relative:2 degree:1 scatterometers:1 collaboration:1 claire:5 periodicity:1 l_:1 side:1 bias:1 wide:1 curve:2 dimension:1 xn:3 valid:2 forward:3 author:1 adaptive:3 san:1 reconstructed:2 xi:3 reasonably:1 robust:1 european:3 domain:2 motivation:1 exponential:1 lie:3 governed:2 mardia:2 third:1 kirk:1 ian:1 bishop:8 showing:1 er:5 sensing:3 labelling:1 conditioned:3 cartesian:1 simply:1 expressed:1 satisfies:1 conditional:19 formulated:1 determined:4 except:1 lui:2 principal:1 called:2 total:1 secondary:1 arises:2
7,176
939
Learning Prototype Models for Tangent Distance Trevor Hastie? Statistics Department Sequoia Hall Stanford University Stanford, CA 94305 email: trevor@playfair .stanford .edu Patrice Simard AT&T Bell Laboratories Crawfords Corner Road Holmdel, NJ 07733 email: patrice@neural.att.com Eduard Siickinger AT &T Bell Laboratories Crawfords Corner Road Holmdel, NJ 07733 email: edi@neural.att.com Abstract Simard, LeCun & Denker (1993) showed that the performance of nearest-neighbor classification schemes for handwritten character recognition can be improved by incorporating invariance to specific transformations in the underlying distance metric - the so called tangent distance. The resulting classifier, however, can be prohibitively slow and memory intensive due to the large amount of prototypes that need to be stored and used in the distance comparisons. In this paper we develop rich models for representing large subsets of the prototypes. These models are either used singly per class, or as basic building blocks in conjunction with the K-means clustering algorithm. *This work was performed while Trevor Hastie was a member of the Statistics and Data Analysis Research Group, AT&T Bell Laboratories, Murray Hill, NJ 07974 . J 000 1 Trevor Hastie, Patrice Simard, Eduard Siickinger INTRODUCTION Local algorithms such as K-nearest neighbor (NN) perform well in pattern recognition, even though they often assume the simplest distance on the pattern space. It has recently been shown (Simard et al. 1993) that the performance can be further improved by incorporating invariance to specific transformations in the underlying distance metric - the so called tangent distance. The resulting classifier, however, can be prohibitively slow and memory intensive due to the large amount of prototypes that need to be stored and used in the distance comparisons. In this paper we address this problem for the tangent distance algorithm, by developing rich models for representing large subsets of the prototypes. Our leading example of prototype model is a low-dimensional (12) hyperplane defined by a point and a set of basis or tangent vectors. The components of these models are learned from the training set, chosen to minimize the average tangent distance from a subset of the training images - as such they are similar in flavor to the Singular Value Decomposition (SVD), which finds closest hyperplanes in Euclidean distance. These models are either used singly per class, or as basic building blocks in conjunction with K-means and LVQ. Our results show that not only are the models effective, but they also have meaningful interpretations. In handwritten character recognition, for instance, the main tangent vector learned for the the digit "2" corresponds to addition/removal of the loop at the bottom left corner of the digit; for the 9 the fatness of the circle. We can therefore think of some of these learned tangent vectors as representing additional invariances derived from the training digits themselves. Each learned prototype model therefore represents very compactly a large number of prototypes of the training set. 2 OVERVIEW OF TANGENT DISTANCE When we look at handwritten characters, we are easily able to allow for simple transformations such as rotations, small scalings, location shifts, and character thickness when identifying the character. Any reasonable automatic scheme should similarly be insensitive to such changes. Simard et al. (1993) finessed this problem by generating a parametrized 7dimensional manifold for each image, where each parameter accounts for one such invariance. Consider a single invariance dimension: rotation. If we were to rotate the image by an angle B prior to digitization, we would see roughly the same picture, just slightly rotated. Our images are 16 x 16 grey-scale pixelmaps, which can be thought of as points in a 256-dimensional Euclidean space. The rotation operation traces out a smooth one-dimensional curve Xi(B) with Xi(O) = Xi, the image itself. Instead of measuring the distance between two images as D(Xi,Xj) = IIX i - Xjll (for any norm 11?11), the idea is to use instead the rotation-invariant DI (Xi, Xj) = minoi,oj IIX i(B;) - Xj(Bj )11. Simard et al. (1993) used 7 dimensions of invariance, accounting fo:: horizontal and vertical location and scale, rotation and shear and character thickness. Computing the manifold exactly is impossible, given a digitized image, and would be impractical anyway. They approximated the manifold instead by its tangent Learning Prototype Models for Tangent Distance 1001 plane at the image itself, leading to the tangent model Xi(B) = Xi + TiB, and the tangent distance DT(Xi,Xj) = minoi,oj IIXi(Bd -Xj(Bj)ll. Here we use B for the 7-dimensional parameter, and for convenience drop the tilde. The approximation is valid locally, and thus permits local transformations. Non-local transformations are not interesting anyway (we don't want to flip 6s into 9s; shrink all digits down to nothing.) See Sackinger (1992) for further details. If 11?11 is the Euclidean norm, computing the tangent distance is a simple least-squares problem, with solution the square-root of the residual sum-of-squares of the residuals in the regression with response Xi - Xj and predictors (-Ti : Tj ). Simard et al. (1993) used DT to drive a 1-NN classification rule, and achieved the best rates so far-2.6%-on the official test set (2007 examples) of the USPS data base. Unfortunately, 1-NN is expensive, especially when the distance function is non-trivial to compute; for each new image classified, one has to compute the tangent distance to each of the training images, and then classify as the class of the closest. Our goal in this paper is to reduce the training set dramatically to a small set of prototype models; classification is then performed by finding the closest prototype. 3 PROTOTYPE MODELS In this section we explore some ideas for generalizing the concept of a mean or centroid for a set of images, taking into account the tangent families. Such a centroid model can be used on its own, or else as a building block in a K-means or LVQ algorithm at a higher level. We will interchangeably refer to the images as points (in 256 space). The centroid of a set of N points in d dimensions minimizes the average squared norm from the points: (1) 3.1 TANGENT CENTROID One could generalize this definition and ask for the point M that minimizes the average squared tangent distance: N MT = argm,Jn LDT(Xi,M)2 (2) i=l This appears to be a difficult optimization problem, since computation of tangent distance requires not only the image M but also its tangent basis TM. Thus the criterion to be minimized is 1002 Trevor Hastie, Patrice Simard, Eduard Sackinger where T(M) produces the tangent basis from M. All but the location tangent vectors are nonlinear functionals of M, and even without this nonlinearity, the problem to be solved is a difficult inverse functional. Fortunately a simple iterative procedure is available where we iteratively average the closest points (in tangent distance) to the current guess. Tangent Centroid Algorithm Initialize: Set M = ~ 2:~1 Xi, let TM = T(M) be the derived set of tangent vectors, and D = 2:i DT(Xi' M). Denote the current tangent centroid (tangent family) by M(-y) = M +TM"I. Iterate: each i find 11M + TM"I - Xi(8)11 1. For N 1 a 1'. and = min'Y.9 8i that solves 2. Set M +- N 2:'=1 (Xi(8i ) - TMi'i) and compute the new tangent subspace TM = T(M). 3. Compute D = 2:iDT(Xi,M) Until: D converges. A Note that the first step in Iterate is available from the computations in the third step. The algorithm divides the parameters into two sets: M in the one, and then TM, "Ii and 8, for each i in the other. It alternates between the two sets, although the computation of TM given M is not the solution of an optimization problem. It seems very hard to say anything precise about the convergence or behavior of this algorithm, since the tangent vectors depend on each iterate in a nonlinear way. Our experience has always been that it converges fairly rapidly ? 6 iterations). A potential drawback of this algorithm is that the TM are not learned, but are implicit in M. 3.2 TANGENT SUBSPACE Rather than define the model as a point and have it generate its own tangent subspace, we can include the subspace as part of the parametrization: M(-y) = M + V"I. Then we define this tangent subspace model as the minimizer of N MS(M, V) = L min 11M + V"Ii - Xi(8d1l 2 (3) . 1 'Yi. 9 i t= over M and V. Note that V can have an arbitrary number 0 ::; r ::; 256 of columns, although it does not make sense for r to be too large. An iterative algorithm similar to the tangent centroid algorithm is available, which hinges on the SVD decomposition for fitting affine subspaces to a set of points. We briefly review the SVD in this context. Let X be the N x 256 matrix with rows the vectors Xi - X where X = ~ 2:~1 Xi. Then SVD(X) = UDV T is a unique decomposition with UNxR and V25 6xR the 1003 Learning Prototype Models for Tangent Distance orthonormal left and right matrices of singular vectors, and R = rank( X). D Rx R is a diagonal matrix of decreasing positive singular values. A pertinent property of the SVD is: Consider finding the closest affine, rank-r subspace to a set of points, or 2 N min M,v(r),{9i} 2: i=1 IIXi - M - v(r)'hll where v(r) is 256 x r orthonormal. The solution is given by the SVD above, with M = X and v(r) the first r columns of V, and the total squared distance E;=1 D;j. The V( r) are also the largest r principal components or eigenvectors of the covariance matrix of the Xi. They give in sequence directions of maximum spread, and for a given digit class can be thought of as class specific invariances. We now present our Tangent subspace algorithm for solving (3); for convenience we assume V is rank r for some chosen r, and drop the superscript. Tangent subspace algorithm Initialize: Set M = ~ Ef:l Xi and let V correspond to the first r right singular vectors of X. Set D = E;=1 D;j, and let the current tangent subspace model be M(-y) = M + V-y. Iterate: 1. For each i find that (ji which solves IIM(-y) - Xi (8)11 = min N 2. Set M +- ~ Ei=1 (Xi (8 i )) and replace the rows of X by Xi({jd - M. Compute the SVD of X, and replace V by the first r right singular vectors. 3. Compute D = E;=l D;j A Until: D converges. The algorithm alternates between i) finding the closest point in the tangent subspace for each image to the current tangent subspace model, and ii) computing the SVD for these closest points. Each step of the alternation decreases the criterion, which is positive and hence converges to a stationary point of the criterion. In all our examples we found that 12 complete iterations were sufficient to achieve a relative convergence ratio of 0.001. One advantage of this approach is that we need not restrict ourselves to a sevendimensional V - indeed, we have found 12 dimensions has produced the best results . The basis vectors found for each class are interesting to view as images. Figure 1 shows some examples of the basis vectors found, and what kinds of invariances in the images they account for. These are digit specific features; for example, a prominent basis vector for the family of 2s accounts for big versus small loops. Trevor Hastie, Patrice Simard, Eduard Siickinger 1004 Each of the examples shown accounts for a similar digit specific invariance. None of these changes are accounted for by the 7-dimensional tangent models, which were chosen to be digit nonspecific. Figure 1: Each column corresponds to a particular tangent subspace basis vector for the given digit . The top image is the basis vector itself, and the remaining 3 images correspond to the 0.1 , 0.5 and 0.9 quantiles for the projection indices for the training data for that basis vector, showing a range of image models for that basis, keeping all the others at o. 4 SUBSPACE MODELS AND K-MEANS CLUSTERING A natural and obvious extension of these single prototype-per-class models, is to use them as centroid modules in a K-means algorithm. The extension is obvious, and space permits only a rough description. Given an initial partition of the images in a class into K sets: 1. Fit a separate prototype model to each of the subsets; 2. Redefine the partition based on closest tangent distance to the prototypes found in step 1. In a similar way the tangent centroid or subspace models can be used to seed LVQ algorithms (Kohonen 1989), but so far we have not much experience with them. 5 RESULTS Table 1 summarizes the results for some of these models. The first two lines correspond to a SVD model for the images fit by ordinary least squares rather than least tangent squares. The first line classifies using Euclidean distance to this model, the second using tangent distance. Line 3 fits a single 12-dimensional tangent subspace model per class, while lines 4 and 5 use 12-dimensional tangent subspaces as cluster Learning Prototype Models for Tangent Distance 1005 Table 1: Test errors for a variety of situations. In all cases the training data were 7291 USPS handwritten digits, and the test data the "official" 2007 USPS test digits. Each entry describes the model used in each class, so for example in row 5 there are 5 models per class, hence 50 in all. 0 1 2 3 4 5 6 7 8 Prototype 1-NN 12 dim SVD subspace 12 dim SVD subspace 12 dim Tangent subspace 12 dim Tangent subspace 12 dim Tangent subspace Tangent centroid (4) U (6) 1-NN Metric Euclidean Euclidean Tangent Tangent Tangent Tangent Tangent Tangent Tangent # Prototypes7Class Error Rate ~ ~ 700 1 1 1 3 5 20 23 700 0.053 0.055 0.045 0.041 0.038 0.038 0.038 0.034 0.026 centers within each class. We tried other dimensions in a variety of settings, but 12 seemed to be generally the best. Line 6 corresponds to the tangent centroid model used as the centroid in a 20-means cluster model per class; the performance compares with with K=3 for the subspace model. Line 7 combines 4 and 6, and reduces the error even further. These limited experiments suggest that the tangent subspace model is preferable, since it is more compact and the algorithm for fitting it is on firmer theoretical grounds. Figure 4 shows some of the misclassified examples in the test set. Despite all the matching, it seems that Euclidean distance still fails us in the end in some of these cases. 6 DISCUSSION Gold, Mjolsness & Rangarajan (1994) independently had the idea of using "domain specific" distance measures to seed K-means clustering algorithms. Their setting was slightly different from ours, and they did not use subspace models. The idea of classifying points to the closest subspace is found in the work of Oja (1989), but of course not in the context of tangent distance. We are using Euclidean distance in conjunction with tangent distance. Since neighboring pixels are correlated, one might expect that a metric that accounted for the correlation might do better. We tried several variants using Mahalanobis metrics in different ways, but with no success. We also tried to incorporate information about where the images project in the tangent subspace models into the classification rule. We thus computed two distances: 1) tangent distance to the subspace, and 2) Mahalanobis distance within the subspace to the centroid for the subspace. Again the best performance was attained by ignoring the latter distance. In conclusion, learning tangent centroid and subspace models is an effective way 1006 Trevor Hastie, Patrice Simard, Eduard Siickinger true: 6 true: 2 true: 5 true: 2 true: 9 true: 4 pred. pro). ( 0 ) pradoproj. ( 0 ) pred. pro). ( 8 ) pred. proj. ( 0 ) prado proj. ( 4 ) prado pro). ( 7 ) Figure 2: Some of the errorS for the test set corresponding to line (3) of table 4. Each case is displayed as a column of three images. The top is the true image, the middle the tangent projection of the true image onto the subspace model of its class, the bottom image the tangent projection of the image onto the winning class . The models are sufficiently rich to allow distortions that can fool Euclidean distance. to reduce the number of prototypes (and thus the cost in speed and memory) at a slight expense in the performance. In the extreme case, as little as one 12 dimensional tangent subspace per class and the tangent distance is enough to outperform classification using ~ 700 prototypes per class and the Euclidean distance (4.1 % versus 5.3% on the test data). References Gold, S., Mjolsness, E. & Rangarajan, A. (1994), Clustering with a domain specific distance measure, in 'Advances in Neural Information Processing Systems', Morgan Kaufman, San Mateo, CA. Kohonen, T. (1989), Self-Organization and Associative Memory (3rd edition), Springer-Verlag, Berlin. Oja, E. (1989), 'Neural networks, principal components, and subspaces', International Journal Of Neural Systems 1(1), 61-68. Sackinger, E. (1992), Recurrent networks for elastic matching in pattern recognition, Technical report, AT&T Bell Laboratories. Simard, P. Y, LeCun, Y. & Denker, J. (1993), Efficient pattern recognition using a new transformation distance, in 'Advances in Neural Information Processing Systems', Morgan Kaufman, San Mateo, CA, pp. 50-58.
939 |@word middle:1 briefly:1 norm:3 seems:2 grey:1 tried:3 decomposition:3 accounting:1 covariance:1 initial:1 att:2 ours:1 current:4 com:2 bd:1 partition:2 pertinent:1 drop:2 stationary:1 guess:1 plane:1 parametrization:1 argm:1 location:3 hyperplanes:1 fitting:2 combine:1 redefine:1 indeed:1 behavior:1 themselves:1 udv:1 roughly:1 decreasing:1 little:1 project:1 classifies:1 underlying:2 what:1 kind:1 kaufman:2 minimizes:2 finding:3 transformation:6 nj:3 impractical:1 ti:1 exactly:1 prohibitively:2 classifier:2 preferable:1 positive:2 local:3 despite:1 might:2 mateo:2 limited:1 range:1 finessed:1 unique:1 lecun:2 block:3 xr:1 digit:11 procedure:1 bell:4 thought:2 projection:3 matching:2 road:2 suggest:1 convenience:2 onto:2 context:2 impossible:1 center:1 nonspecific:1 independently:1 identifying:1 rule:2 orthonormal:2 anyway:2 recognition:5 approximated:1 expensive:1 bottom:2 tib:1 module:1 solved:1 mjolsness:2 decrease:1 depend:1 solving:1 basis:10 usps:3 compactly:1 easily:1 effective:2 stanford:3 say:1 distortion:1 statistic:2 think:1 itself:3 patrice:6 superscript:1 ldt:1 associative:1 sequence:1 advantage:1 kohonen:2 neighboring:1 loop:2 rapidly:1 achieve:1 gold:2 description:1 convergence:2 cluster:2 rangarajan:2 produce:1 generating:1 converges:4 rotated:1 develop:1 recurrent:1 nearest:2 solves:2 direction:1 drawback:1 extension:2 sufficiently:1 hall:1 eduard:5 ground:1 seed:2 bj:2 largest:1 rough:1 always:1 rather:2 conjunction:3 derived:2 rank:3 centroid:14 sense:1 dim:5 nn:5 proj:2 misclassified:1 pixel:1 classification:5 initialize:2 fairly:1 represents:1 look:1 minimized:1 others:1 idt:1 report:1 oja:2 ourselves:1 organization:1 extreme:1 tj:1 experience:2 euclidean:10 divide:1 circle:1 theoretical:1 instance:1 classify:1 column:4 measuring:1 ordinary:1 cost:1 subset:4 entry:1 predictor:1 too:1 stored:2 thickness:2 international:1 squared:3 again:1 corner:3 simard:11 leading:2 account:5 potential:1 performed:2 root:1 view:1 tmi:1 minimize:1 square:5 correspond:3 generalize:1 handwritten:4 produced:1 none:1 rx:1 drive:1 classified:1 fo:1 trevor:7 email:3 definition:1 pp:1 obvious:2 di:1 ask:1 appears:1 higher:1 dt:3 attained:1 response:1 improved:2 shrink:1 though:1 just:1 implicit:1 until:2 correlation:1 horizontal:1 sackinger:3 ei:1 nonlinear:2 building:3 concept:1 true:8 hence:2 laboratory:4 iteratively:1 mahalanobis:2 ll:1 interchangeably:1 self:1 anything:1 criterion:3 m:1 prominent:1 hill:1 complete:1 pro:3 image:27 ef:1 recently:1 rotation:5 shear:1 mt:1 functional:1 overview:1 ji:1 insensitive:1 interpretation:1 slight:1 refer:1 automatic:1 rd:1 similarly:1 nonlinearity:1 had:1 base:1 closest:9 own:2 showed:1 firmer:1 verlag:1 success:1 alternation:1 yi:1 morgan:2 additional:1 fortunately:1 ii:3 reduces:1 smooth:1 technical:1 variant:1 basic:2 regression:1 metric:5 iteration:2 achieved:1 addition:1 want:1 else:1 singular:5 member:1 enough:1 iterate:4 xj:6 fit:3 variety:2 hastie:6 restrict:1 reduce:2 idea:4 prototype:20 tm:8 intensive:2 shift:1 dramatically:1 generally:1 fool:1 eigenvectors:1 amount:2 singly:2 iixi:2 locally:1 simplest:1 generate:1 outperform:1 xjll:1 per:8 group:1 sum:1 angle:1 inverse:1 family:3 reasonable:1 summarizes:1 holmdel:2 scaling:1 speed:1 min:4 department:1 developing:1 alternate:2 describes:1 slightly:2 character:6 invariant:1 flip:1 end:1 available:3 operation:1 permit:2 denker:2 jn:1 jd:1 top:2 clustering:4 include:1 remaining:1 iix:2 hinge:1 murray:1 especially:1 diagonal:1 subspace:34 distance:41 separate:1 berlin:1 parametrized:1 digitization:1 manifold:3 trivial:1 index:1 ratio:1 difficult:2 unfortunately:1 expense:1 trace:1 perform:1 vertical:1 displayed:1 tilde:1 situation:1 precise:1 digitized:1 arbitrary:1 edi:1 pred:3 learned:5 address:1 able:1 pattern:4 oj:2 memory:4 natural:1 residual:2 representing:3 scheme:2 hll:1 picture:1 crawford:2 prior:1 review:1 tangent:70 removal:1 relative:1 expect:1 interesting:2 versus:2 affine:2 sufficient:1 classifying:1 row:3 course:1 accounted:2 keeping:1 iim:1 allow:2 neighbor:2 taking:1 curve:1 dimension:5 valid:1 rich:3 seemed:1 san:2 far:2 functionals:1 compact:1 xi:23 don:1 iterative:2 table:3 ca:3 elastic:1 ignoring:1 domain:2 official:2 did:1 main:1 spread:1 big:1 edition:1 nothing:1 quantiles:1 slow:2 fails:1 winning:1 third:1 down:1 specific:7 showing:1 incorporating:2 flavor:1 generalizing:1 explore:1 springer:1 corresponds:3 minimizer:1 goal:1 lvq:3 replace:2 change:2 hard:1 hyperplane:1 principal:2 called:2 total:1 invariance:9 svd:11 meaningful:1 playfair:1 latter:1 rotate:1 incorporate:1 correlated:1
7,177
94
195 LEARNING WITH TEMPORAL DERIVATIVES IN PULSE-CODED NEURONAL SYSTEMS Mark Gluck David B. Parker Eric S. Reifsnider Department of Psychology Stanford University Stanford, CA 94305 Abstract A number of learning models have recently been proposed which involve calculations of temporal differences (or derivatives in continuous-time models). These models. like most adaptive network models. are formulated in tenns of frequency (or activation), a useful abstraction of neuronal firing rates. To more precisely evaluate the implications of a neuronal model. it may be preferable to develop a model which transmits discrete pulse-coded information. We point out that many functions and properties of neuronal processing and learning may depend. in subtle ways. on the pulse-coded nature of the information coding and transmission properties of neuron systems. When compared to formulations in terms of activation. computing with temporal derivatives (or differences) as proposed by Kosko (1986). Klopf (1988). and Sutton (1988). is both more stable and easier when reformulated for a more neuronally realistic pulse-coded system. In reformulating these models in terms of pulse-coding. our motivation has been to enable us to draw further parallels and connections between real-time behavioral models of learning and biological circuit models of the substrates underlying learning and memory. INTRODUCTION Learning algorithms are generally defined in terms of continuously-valued levels of input and output activity. This is true of most training methods for adaptive networks. (e.g .? Parker. 1987; Rumelhart. Hinton. & Williams, 1986; Werbos. 1974; Widrow & Hoff, 1960). and also for behavioral models of animal and hwnan learning. (e.g. Gluck & Bower. 1988a, 1988b; Rescorla & Wagner. 1972). as well as more biologically oriented models of neuronal function (e.g .? Bear & Cooper, in press; Hebb, 1949; Granger. Abros-Ingerson, Staubli, & Lynch, in press; Gluck & Thompson, 1987; Gluck. Reifsnider. & Thompson. in press; McNaughton & Nadel. in press; Gluck & Rumelhart. in press). In spite of the attractive simplicity and utility of the "activation" construct 196 Parker, Gluck and Reifsnider neurons use discrete trains of pulses for the transmission of information from cell to cell. Frequency (or activation) is a useful abstraction of pulse trains. especially for bridging the gap between whole-animal and single neuron behavior. To more precisely evaluate the implications of a neuronal model. it may be preferable to develop a model which transmits discrete pulse-coded information; it is possible that many functions and properties of neuronal processing and learning may depend. in subtle ways. on the pulse-coded nature of the information coding and transmission properties of neuron systems. In the last few years, a number of learning models have been proposed which involve computations of temporal differences (or derivatives in continuous-time models). Klopf (1988) presented a formal real-time model of classical conditioning that predicts the magnitude of conditioned responses (CRs). given the temporal relationships between conditioned stimuli (eSs) and an unconditional stimulus (US). Klopf's model incorporates a "differential-Hebbian" learning algorithm in which changes in presynaptic levels of activity are correlated with changes in postsynaptic levels of activity. Motivated by the constraints and motives of engineering. rather than animal learning. Kosko (1986) proposed the same basic rule and provided extensive analytic insights into its properties. Sutton (1988) introduced a class of incremental learning procedures. called "temporal difference" methods. which update associative (predictive) weights according to the difference between temporally successive predictions. In addition to the applied potential of this class of algorithms. Sutton & Barto (1987) show how their model. like Klopf's (1988) model. provides a good fit to a wide range of behavioral data on classical conditioning. These models. all of which depend on computations involving changes over time in activation levels. have been successful both for predicting a wide range of behavioral animal learning data (Klopf. 1988; Sutton & Barto. 1987) and for solving useful engineering problems in adaptive prediction (Kosko. 1986; Sutton. 1988). The possibility that these models might represent the computational properties of individual neurons. seems, at first glance. highly unlikely. However. we show by reformulating these models for pulse-coded communication (as in neuronal systems) rather than in terms of abstract activation levels. the computational soundness as well as the biological relevance of the models is improved. By avoiding the use of unstable differencing methods in computing the time-derivative of activation levels. and by increasing the error-tolerance of the computations, pulse coding will be shown to improve the accuracy and reliability of these models. The pulse coded models will also be shown to lend themselves to a closer comparison to the function of real neurons than do models that operate with activation levels. As the ability of researchers to directly measure neuronal behavior grows. the value of such close comparisons will increase. As an example. we describe here a pulse-coded version of Klopf's differential-Hebbian model of classification learning. Further details are contained in Gluck. Parker. & Reifsnider. 1988. Learning with Temporal Derivatives Pulse-Coding in Neuronal Systems We begin by outlining the general theory and engineering advantages of pulse-coding and then describe a pulse-coded refonnulation of differential-Hebbian learning. The key idea is quite simple and can be summarized as follows: Frequency can be seen, loosely speaking, as an integral of pulses; conversely, therefore, pulses can be thought of as carrying infonnation about the derivatives of frequency. Thus, computing with the "derivatives of frequency" is analogous to computing with pulses. As described below, our basic conclusion is that differential-Hebbian learning (Klopf, 1988; Kosko, 1986) when refonnulated for a pulse-coded system is both more stable and easier to compute than is apparent when the rule is fonnulated in tenns of frequencies. These results have important implications for any learning model which is based on computing with timederivatives, such as Sutton's Temporal Difference model (Sutton, 1988; Sutton & Barto, 1987) There are many ways to electrically transmit analog information from point to point. Perhaps the most obvious way is to transmit the infonnation as a signal level. In electronic systems, for example, data that varies between 0 and 1 can be transmitted as a voltage level that varies between 0 volts and 1 volt This method can be unreliable, however, because the receiver of the information can't tell if a constant DC voltage offset has been added to the information, or if crosstalk has occurred with a nearby signal path. To the exact degree that the signal is interfered with, the data as read by the receiver will be erroneously altered. The consequences of faults appearing in the signal are particularly serious for systems that are based on derivatives of the signal. In such systems, even a small, but sudden, unintended change in signal level can drastically alter its derivative, creating large errors. A more reliable way to transmit analog information is to encode it as the frequency of a series of pulses. A receiver can reliably detennine if it has received a pulse, even in the face of DC voltage offsets or moderate crosstalk. Most errors will not be large enough to constitute a pulse, and thus will have no effect on the transmitted infonnation. The receiver can count the number of pulses received in a given time window to detennine the frequency of the pulses. Further infonnation on encoding analog infonnation as the frequency of a series of pulses can be found in many electrical engineeri.ng textbooks (e.g., Horowitz & Hill, 1980). As noted by Parker (1987), another advantage of coding an analog signal as the frequency of a series of pulses is that the time derivative of the signal can be easily and stably calculated: If x (t) represents a series of pulses (x equals 1 if a pulse is occuring at time t; otherwise it equals 0) then we can estimate the frequency, f (t), of the series of pulses using an exponentially weighted time average: f (t) = Jllx ('t)e-Jl{t-'t) d't 197 198 Parker, Gluck and Reifsnider where Jl is the decay constant. The well known formula for the derivative of 1 (t) is AtJP- = Jl~(t)-/(t)) Thus. the time derivative of pulse-coded information can be calculated without using any unstable differencing methods. it is simply a function of presence or absence of a pulse relative to the current expectation (frequency) of pulses. As described earlier. calculation of time derivatives is a critical component of the learning algorithms proposed by Klopf (1988). Kosko (1986) and Sutton (Sutton. 1988; Sutton & Barto 1987). They are also an important aspect of 2nd order (pseudo-newtonian) extensions of the backpropogation learning rule for multi-layer adaptive "connectionist" networks (parker. 1987). Summary 01 Klopf s Model Klopf (1988) proposed a model of classical conditioning which incorporates the same learning rule proposed by Kosko (1986) and which extends some of the ideas presented in Sutton and Barto's (1981) real-time generalization of Rescorla and Wagner's (1972) model of classical conditioning. The mathematical specification of Klopf s model consists of two equations: one which calculates output signals based on a weighted sum of input signals (drives) and one which determines changes in synapse efficacy due to changes in signal levels. The specification of signal output level is defined as where: y (t ) is the measure of postsynaptic frequency of firing at time t; Wi (t) is the efficacy (positive or negative) of the i th synapse; Xi (t) is the frequency of action potentials at the i th synapse; 9 is the threshold of firing; and n is the number of synapses on the "neuron". This equation expresses the idea that the postsynaptic firing frequency depends on the summation of the weighted presynaptic firing frequencies. Wi (t )Xi (t ). relative to some threshold. 9. The learning mechanism is defined as where: ~Wi (t) is the change in efficacy of the i th synapse at time t; ~y (t) is the change in postsynaptic firing at time t; 't' is the longest interstimulus interval over which delayed conditioning is effective. The Cj are empirically established learning rate constants -each corresponding to a different inter-stimulus interval. In order to accurately simulate various behavioral phenomena observed in classical conditioning. Klopf adds three ancillary assumptions to his model. First. he places a lower bound of 0 on the activation of the node. Second. he proposes that changes in synaptic 199 Learning with Temporal Derivatives weight, ~w; (t), be calculated only when the change in presynaptic signal level is positive -- that is, when Ax; (t-j) > O. Third, he proposes separate excitatory and inhibitory weights in contrast to the single real-valued associative weights in other conditioning models (e.g., Rescorla & Wagner, 1972; Sutton & Barto, 1981). It is intriguing to note that all of these assumptions are not only sufficiently justified by constraints from behavioral data but are also motivated by neuronal constraints. For a further examination of the biological and behavioral factors supporting these assumptions see Gluck, Parker, and Reifsnider (1988). The strength of Klopf's model as a simple formal behavioral model of classical conditioning is evident. Although the model has not yielded any new behavioral predictions, it has demonstrated an impressive ability to reproduce a wide, though not necessarily complete, range of Pavlovian behavioral phenomena with a minimum of assumptions. Klopf (1988) specifies his learning algorithm in terms of activation or frequency levels. Because neuronal systems communicate through the transmission of discrete pulses, it is difficult to evaluate the biological plausibility of an algorithm when so formulated. For this reason, we present and evaluate a pulse-coded reformulation of Klopf's model. A Pulse-Coded Reformulation of Klopf s Model We illustrate here a pulse-coded reformulation of Klopf's (1988) model of classical conditioning. The equations that make up the model are fairly simple. A neuron is said to have fired an output pulse at time t if vet) > e, where e is a threshold value and vet) is defined as follows: vet) = (l-d)v(t-l) + !:Wi(t)Xi(t) (1) where v (t) an auxiliary variable, d is a small positive constant representing the leakage or decay rate, Wi (t) is the efficacy of synapse i at time t, and Xi (t) is the frequency of presynaptic pulses at time t at synapse i. The input to the decision of whether the neuron will fire consists of the weights and efficacies of the synapses as well as information about previous activation levels at the neuronal output Note that the leakage rate, d, causes older information about activation levels to have less impact on current values of v (t) than does recent information of the same type. The output of the neuron, p (t), is: v (t) > e then p (t) = 1 (pulse generated) v (t ) ~ e then p (t) = 0 (no pulse generated) It is important that once p (t) has been determined, v (t) will need to be adjusted if 200 Parker, Gluck and Reifsnider To reflect the fact that the neuron has fired, (i.e., p (t) = 1) then v (t) = v (t) - 1. This decrement occurs after p (t) has been determined for the current t. Frequencies of pulses at the output node and at the synapses are calculated using the following equations: p (t) = 1. / (t) =/ (t-l) + 11/(t) where 11/ (t) = m(p (t) - / (t-l)) where / (t) is the frequency of outgoing pulses at time t; p (t) is the ouput (1 or 0) of the neuron at time t ; and m is a small positive constant representing a leakage rate for the frequency calculation. Following Klopf (1988), changes in synapse efficacy occur according to (2) where I1Wi(t) = Wi (t+l) - Wi(t) and l1y (t) and ru:i (t) are calculated analogously to 11/ (t); 't is the longest interstimulus interval (lSI) over which delay conditioning is effective; and Cj is an empirically established set of learning rates which govern the efficacy of conditioning at an lSI of j . Changes in Wi (t) are governed by the learning rule in Equation 2 which alters v (t) via Equation 1. Figure 1 shows the results of a computer simulation of a pulse-coded version of Klopf's conditioning model. The first graph shows the excitatory weight (dotted line) and inhibitory weight (dashed line) of the CS "synapse". Also on the same graph is the net synaptic weight (solid line), the sum of the excitatory and inhibitory weights. The subsequent graphs show CS input pulses, US input pulses, and the output (CR) pulses. The simulation consists of three acquisition trials followed by three extinction trials. 201 Learning with Temporal Derivatives en l! 0.4 .~ 02 . ; ~ (...............................J 0.6 0.0 ?0.2 .................................. (................................., ......... ~ ?V1???????????U .............................. ------------------------------------------~-----~----------- o 50 100 150 200 250 cycle Figure 1. Simulation of pulse.coded version of Klopf's conditioning model. Top panel shows excitatory and inhibitory weights as dashed lines and the net synaptic weight of the CS as a solid line. Lower panels show the CS and US inputs and the CR output. As expected, excitatory weight increases in magnitude over the three ~quisition trials, while inhibitory weight is stable. During the first two extinction trials, the excitatory and the net synaptic weights decrease in magnitude, while the inhibitory weight increases. Thus, the CS produces a decreasing amount of output pulses (the CR). During the third extinction trial the net synaptic weight is so low that the CS cannot produce output pulses, and so the CR is extinct. However, as net weight and excitatory weight remain positive, there are residual effects of the acquisition which will accelerate reacquisition. Because a threshold must be reached before a neuronal output pulse can be emitted, and because output must occur for weight changes to occur, pulse coding adds to the accelerated reacquisition effect that is evident in the original Klopf model; extinction is halted before net weight is zero, when pulses can no longer be produced. 300 202 Parker, Gluck and Reifsnider Discussion To facilitate comparison between learning algorithms involving temporal derivative computations and actual neuronal capabilities. we formulated a pulse-coded variation of Klopfs classical conditioning model. Our basic conclusion is that computing with temporal derivatives (or differences) as proposed by Kosko (1986). Klopf (1988). and Sutton (1988). is more stable and easier when reformulated for a more neuronally realistic. pulse-coded system. than when the rules are fonnulated in terms of frequencies or activation. It is our hope that further examination of the characteristics of pulse-coded systems may reveal facts that bear on the characteristics of neuronal function. In refonnulating these algorithms in terms of pulse-coding. our motivation has been to enable us to draw further parallels and connections between real-time behavioral models of learning and biological circuit models of the substrates underlying classical conditioning. (e.g .? Thompson. 1986; Gluck & Thompson. 1987; Donegan. Gluck. & Thompson. in press). More generally. noting the similarities and differences between algorithmic/behavioral theories and biological capabilities is one way of laying the groundwork for developing more complete integrated theories of the biological bases of associative learning (Donegan. Gluck. & Thompson. in press). Acknowledgments Correspondence should be addressed to: Mark A. Gluck. Dept of Psychology. Jordan Hall; Bldg. 420. Stanford. CA 94305. For their commentary and critique on earlier drafts of this and related papers. we are indebted to Harry Klopf. Bart Kosko. Richard Sutton. and Richard Thompson. This research was supported by an Office of Naval Research Grant to R. F. Thompson and M. A. Gluck. References Bear. M. F., & Cooper, L. N. (in press). Molecular mechanisms for synaptic modification in the visual cortex: Interaction between theory and experiment. In M. A. Gluck, & D. E. Rumelhart (Eds.), Neuroscience and Connectionist Theory. Hillsdale, N.J.: Lawrence Erlbaum Associates .. Donegan, N. H., Gluck, M. A., & Thompson, R. F. (1989). Integrating behavioral and biological models of classical conditioning. In R. D. Hawkins, & G. H. Bower (Eds.), Computational models of learning in simple neural systems (Volume 22 of the Psychology of Learning and Motivation). New York: Academic Press. Gluck, M. A., & Bower. G. H. (1988a). Evaluating an adaptive network model of human learning. Journal of Memory and Language, 27, 166-195. Gluck, M. A., & Bower, G. H. (1988b). From conditioning to category learning: An adaptive network model. Journal of Experimental Psychology: General, 117(3), 225-244. Learning with Temporal Derivatives Gluck, M. A., Parker, D. B., & Reifsnider, E. (1988). Some biological implications of a differential-Hebbian learning rule. Psychobiology, 16(3), 298-302. Gluck, M. A, Reifsnider, E. S., & Thompson, R. F. (in press). Adaptive signal processing and temporal coarse coding: Cerebellar models of classical conditioning and VOR Adaptation. In M. A. Gluck, & D. E. Rumelhart (Eds.), Neuroscience and Connectionist Theory. Hillsdale, N.1.: Lawrence Erlbaum Associates .. Gluck, M. A, & Rumelhart, D. E. (in press). Neuroscience and Connectionist Theory. Hillsdale, N.J.: Lawrence Erlbaum Associates .. Gluck, M. A., & Thompson, R. F. (1987). Modeling the neural substrates of associative learning and memory: A computational approach. Psychological Review, 94, 176-191. Granger, R., Ambros-Ingerson, 1., Staubli, U., & Lynch, G. (in press). Memorial operation of multipIe, interacting simulated brain structures. In M. A. Gluck, & D. E. Rumelhart (Eds.), Neuroscience and Connectionist Theory. Hillsdale, N.J.: Lawrence Erlbaum Associates .. Hebb, D. (1949). Organization of Behavior. New York: Wiley & Sons. Horowitz, P., & Hill, W. (1980). The Art of Electronics. Cambridge, England: Cambridge University Press. Klopf, A. H. (1988). A neuronal model of classical conditioning. Psychobiology, 16(2), 85-125. Kosko, B. (1986). Differential hebbian learning. In 1. S. Denker (Ed.), Neural Networksfor Computing, AlP Conference Proceedings 151 (pp. 265-270). New York: American Institute of Physics. McNaughton, B. L., & Nadel, L. (in press). Hebb-Marr networks and the neurobiological representation of action in space. In M. A. Gluck, & D. E. Rumelhart (Eds.), Neuroscience and Connectionist Theory. Hillsdale, N.J.: Lawrence Erlbaum Associates .. Parker, D. B. (1987). Optimal Algorithms for Adaptive Networks: Second Order Back Propagation, Second Order Direct Propagation, and Second Order Hebbian Learning. Proceedings of the IEEE First Annual Conference on Neural Networks. San Diego, California:, . Rescorla. R. A, & Wagner, A. R. (1972). A theory of Pavlovian conditioning: Variations in the effectiveness of reinforcement and non-reinforcement. In A. H. Black, & W. F. Prokasy (Eds.), Classical conditioning II: Current research and theory. New York: AppletonCentury-Crofts. RumeIhart, D. E., Hinton, G. E., & Williams, R. J. (1986). Learning internal representations by error propogation. In D. Rumelhart, & 1. McClelland (Eds.), Parallel distributed processing: Explorations in the microstructure of cognition (Vol. 1: Foundations). Cambridge, M.A.: MIT Press. Sutton, R. S. (1988). Learning to predict by the methods of temporal differences. Machine Learning, 3, 9-44. Sutton, R. S., & Barto, A. G. (1981). Toward a modem theory of adaptive networks: Expectation and prediction. Psychological Review, 88, 135-170. Sutton, R. S., & Barto, A. G. (1987). A temporal-difference model of classical conditioning. In Proceedings of the 9th Annual Conference of the Cognitive Science Society. Seattle, WA. Thompson, R. F. (1986). The neurobiology ofleaming and memory. Science, 233, 941-947. Werbos, P. (1974). Beyond regression: New tools for prediction and analysis in the behavioral sciences. Doctoral dissertation (Economics), Harvard University, Cambridge, Mass .. Widrow, B., & Hoff, M. E. (1960). Adaptive switching circuits. Institute of Radio Engineers, Western Electronic Show and Convention, Convention Record, 4,96-194. 203 Part II Application
94 |@word trial:5 version:3 seems:1 nd:1 extinction:4 pulse:57 simulation:3 solid:2 electronics:1 series:5 efficacy:7 unintended:1 groundwork:1 current:4 activation:13 intriguing:1 must:2 fonnulated:2 vor:1 subsequent:1 realistic:2 analytic:1 update:1 bart:1 es:1 dissertation:1 record:1 sudden:1 provides:1 draft:1 node:2 coarse:1 successive:1 mathematical:1 direct:1 differential:6 ouput:1 consists:3 behavioral:14 inter:1 expected:1 behavior:3 themselves:1 multi:1 brain:1 decreasing:1 actual:1 window:1 increasing:1 provided:1 begin:1 underlying:2 circuit:3 panel:2 mass:1 textbook:1 temporal:16 pseudo:1 preferable:2 grant:1 positive:5 before:2 engineering:3 consequence:1 switching:1 sutton:18 encoding:1 critique:1 firing:6 path:1 might:1 black:1 doctoral:1 conversely:1 range:3 acknowledgment:1 crosstalk:2 procedure:1 thought:1 integrating:1 spite:1 cannot:1 close:1 demonstrated:1 williams:2 economics:1 thompson:12 simplicity:1 rule:7 insight:1 marr:1 his:2 variation:2 mcnaughton:2 analogous:1 transmit:3 diego:1 exact:1 substrate:3 associate:5 harvard:1 rumelhart:8 particularly:1 werbos:2 predicts:1 observed:1 electrical:1 cycle:1 decrease:1 govern:1 depend:3 solving:1 carrying:1 predictive:1 eric:1 easily:1 accelerate:1 various:1 train:2 describe:2 effective:2 tell:1 quite:1 apparent:1 stanford:3 valued:2 otherwise:1 ability:2 soundness:1 associative:4 advantage:2 net:6 rescorla:4 interaction:1 adaptation:1 fired:2 interstimulus:2 staubli:2 seattle:1 transmission:4 produce:2 incremental:1 newtonian:1 illustrate:1 develop:2 widrow:2 received:2 auxiliary:1 c:6 convention:2 ancillary:1 human:1 alp:1 enable:2 exploration:1 hillsdale:5 microstructure:1 generalization:1 biological:9 summation:1 adjusted:1 extension:1 sufficiently:1 hall:1 hawkins:1 lawrence:5 algorithmic:1 cognition:1 predict:1 radio:1 infonnation:5 tool:1 weighted:3 hope:1 mit:1 lynch:2 rather:2 cr:5 barto:8 voltage:3 office:1 encode:1 ax:1 naval:1 longest:2 contrast:1 abstraction:2 motif:1 unlikely:1 integrated:1 reproduce:1 classification:1 proposes:2 animal:4 art:1 fairly:1 hoff:2 equal:2 construct:1 once:1 ng:1 represents:1 alter:1 connectionist:6 stimulus:3 serious:1 few:1 richard:2 oriented:1 individual:1 delayed:1 fire:1 organization:1 possibility:1 highly:1 unconditional:1 implication:4 integral:1 closer:1 loosely:1 psychological:2 earlier:2 modeling:1 halted:1 delay:1 successful:1 erlbaum:5 varies:2 propogation:1 physic:1 analogously:1 continuously:1 reflect:1 ambros:1 cognitive:1 creating:1 horowitz:2 derivative:19 american:1 potential:2 harry:1 coding:10 summarized:1 depends:1 reached:1 parallel:3 capability:2 accuracy:1 characteristic:2 neuronally:2 accurately:1 produced:1 multipie:1 psychobiology:2 researcher:1 drive:1 indebted:1 synapsis:3 synaptic:6 ed:8 acquisition:2 frequency:22 pp:1 obvious:1 transmits:2 cj:2 subtle:2 back:1 response:1 improved:1 synapse:8 formulation:1 though:1 propagation:2 glance:1 western:1 stably:1 reveal:1 perhaps:1 grows:1 facilitate:1 effect:3 true:1 reformulating:2 read:1 volt:2 attractive:1 during:2 noted:1 hill:2 evident:2 occuring:1 complete:2 recently:1 empirically:2 conditioning:22 exponentially:1 volume:1 jl:3 analog:4 occurred:1 he:3 cambridge:4 backpropogation:1 language:1 reliability:1 stable:4 specification:2 impressive:1 similarity:1 cortex:1 longer:1 add:2 base:1 recent:1 moderate:1 tenns:2 fault:1 seen:1 transmitted:2 minimum:1 commentary:1 signal:14 dashed:2 ii:2 hebbian:7 memorial:1 academic:1 england:1 plausibility:1 calculation:3 molecular:1 coded:20 calculates:1 prediction:5 nadel:2 basic:3 involving:2 impact:1 regression:1 expectation:2 represent:1 cerebellar:1 cell:2 justified:1 addition:1 interval:3 addressed:1 operate:1 incorporates:2 effectiveness:1 jordan:1 emitted:1 presence:1 noting:1 enough:1 fit:1 psychology:4 idea:3 whether:1 motivated:2 utility:1 bridging:1 reformulated:2 speaking:1 cause:1 constitute:1 action:2 york:4 useful:3 generally:2 involve:2 amount:1 category:1 mcclelland:1 specifies:1 lsi:2 inhibitory:6 alters:1 dotted:1 neuroscience:5 discrete:4 vol:1 express:1 key:1 reformulation:3 threshold:4 v1:1 graph:3 year:1 sum:2 bldg:1 communicate:1 extends:1 place:1 electronic:2 draw:2 decision:1 networksfor:1 layer:1 bound:1 followed:1 correspondence:1 yielded:1 activity:3 annual:2 strength:1 occur:3 precisely:2 constraint:3 nearby:1 erroneously:1 aspect:1 simulate:1 extinct:1 pavlovian:2 department:1 developing:1 according:2 electrically:1 ingerson:2 remain:1 son:1 postsynaptic:4 wi:8 biologically:1 modification:1 equation:6 granger:2 count:1 mechanism:2 detennine:2 operation:1 denker:1 appearing:1 original:1 top:1 especially:1 classical:14 society:1 leakage:3 added:1 occurs:1 said:1 separate:1 simulated:1 evaluate:4 presynaptic:4 unstable:2 reason:1 toward:1 laying:1 ru:1 relationship:1 differencing:2 difficult:1 negative:1 reliably:1 jllx:1 neuron:12 kosko:9 modem:1 supporting:1 hinton:2 communication:1 neurobiology:1 dc:2 interacting:1 david:1 introduced:1 extensive:1 connection:2 california:1 established:2 beyond:1 below:1 reliable:1 memory:4 lend:1 critical:1 examination:2 predicting:1 residual:1 representing:2 older:1 improve:1 altered:1 ofleaming:1 temporally:1 review:2 relative:2 bear:3 outlining:1 foundation:1 degree:1 excitatory:7 summary:1 supported:1 last:1 drastically:1 formal:2 institute:2 wide:3 face:1 wagner:4 tolerance:1 distributed:1 calculated:5 l1y:1 evaluating:1 adaptive:10 san:1 reinforcement:2 prokasy:1 neurobiological:1 unreliable:1 receiver:4 xi:4 continuous:2 vet:3 nature:2 ca:2 necessarily:1 decrement:1 motivation:3 whole:1 hwnan:1 neuronal:17 en:1 parker:12 hebb:3 cooper:2 wiley:1 governed:1 bower:4 third:2 croft:1 formula:1 offset:2 decay:2 magnitude:3 conditioned:2 interfered:1 gap:1 easier:3 gluck:27 simply:1 visual:1 contained:1 determines:1 formulated:3 absence:1 change:13 determined:2 engineer:1 called:1 experimental:1 klopf:24 internal:1 mark:2 relevance:1 accelerated:1 avoiding:1 dept:1 outgoing:1 phenomenon:2 correlated:1
7,178
940
? ?            !#"%$   "'&)(*,+.-0/ ? ? H IKJ` 13LNR 25^ 46MO25JI57cdPi83QIE9!e R!: T3S.TVPdjUWI5Z ;PX5Y[I IEZWklX]\ Pj gvh P_wxt IEeyo Z iWzR S ` R{ Rey|~} R ? R c?{ h I ep? `????d? ??? ???? ?V?[?V??? ?V??[?3?[?a?3? ?a? ??]? ??? ? ? ? ? Hm IEJ LNP<op^ Q =Z Q I5>@c_P_i? QI[R!R!Saqa`b`bA.RB!R ^C j!P J2EZWcdQGDGrFsuQ tIE25e3I = f f> X]X]Z Z I I5P P X X I I hn vP Z t IKeyopwxi z R!Sa`R {xR ey| h R ? R cd{ h I e6? ` ? ??d??? ? ?!?? ? ? ? ?3? ? ?d???[?? ? ?[?.?V?3? ? ?? ? ? ??? ?? ?b?]?E??V?d? ???6fcdZ ?xo@i JdR Pd?~?'?!? e.?I5? ? ? J ?]{xi RRle6SpI]R]o e iWI ? X]IV? |~i JZxPJdj{?ZWXix?|~Ia? Zx?R IP ? R| St?ZxRI ^edR!J S R e z \ Pd?H Z?? XIE\?epI ?P z X oyIi I ?G?? ?d?{ ?I]i ??[e P I Z \Pdj ? a ? ^ tdR!ql\ ?c? {xR I5c? ixJP cZ Pdi oj ? ZW??P X R R J^ ? JR ?o6?,w o R i PR ?j!R ??opi I5m ???xwx? ? IOo Z o6h ixc? Jd|EixI ZGe R tdP?Z oop???y? ?{xI5? IOP?ZxPI ej S R e IE^OJ ImP X]Io R PI5QxIEQW?? I ? ??P ZG? iRZW?X PdX5i ?d|R RS Qx?WP X RiZxZ? ^OI I J epo I?I eoprI5? I]P!oi ?bZ P R ^ ? ?dh |I { h Rh |E?i |]???P R Zn opPI ?W{x? ? Z o I epJd?[| ? JR I~ee6??W? ?I h I ? ? c ?? I#hR| S hh!?? ep? i??tI ? opi IEJ?? I | h J!e6? ZWX i wGRP o ? R c!e I ?!? I k ^I5P i oo ? R? i ? |Eiopi g h? } td?o ? zZWI ? ?opI~?d? ? ? r? opZ ??Pd? j U {xQZ ?dIP!??I PZ P!P?x? Pi I j| ep}EP w ?~oa\ | ?X]eI5IEI PixJiWiI e e?J?I?e6iop??I5mP!ZWiGX| ?ii wZxRRPP {Wo~IZ ?.P ???i P ? Zo IP ZW? jd?!?IEUG?G? ???R I?I e??l???? ?Ii \ X]j ? R? ?e ?dZWc? Io ^ RX]R? X]^Qx?m J P? ?dI hII ??^ ?Z }?R hUxi IZ? ? P o X i c ?I5PJw ? Z ? iGJ o e6 ?? ??? epX]I ZWR Xo Pdi Z S o R R xP?o ei o ?i ? R!R SViW? ?#rso P ??J?h ?OQWw ?td| IZ hh Zco j ? \ ^ R P oP I } ? ?iW? ? R I e o ?? o[?eyoS6?iR e ? I ? X | ?Z X Z? }Qx??I5R!P?S i c!iWc?epIPIw i?? ot  o ? I?I Jh?d?O ? e |EI ixX I| ? oJ!Z e6P? j {xZ IX i  ZxUGR ?P j oaI q epR? e[? iWZ ? h I } I5PPI]?!{G|~i z ? I t eaIQW??o |? I  J? o P!?}do3I|Vj epI IE ? IEe eyI]|~op?xIEI P } i ?|EQ ? ZxR R P eVh ?? ? RwG^o3}d? |Eix? ? Z X o ?I ? \? {I?}t RI P R ? ? P!} m z o Z j  ? ?X]m ? h w  I~epI P X5I ? I i ?  POiW? Ibi ? R? X ? ZxQWI X iWc e I o  ?   ? !"$#&%!%' (&)+*,# ?)- . - /!0+1 -324!15 %76$0!8 1 / 4!0 * %79 -!:<; =?>A@ BDCFEHGJIKGLMNC E I OQPSR I T R LVUSCXW Y[Z\&]M L&E C ^`_Kab a McEdI Efe  "$gh % % e  "ji )+*,k ) - . - /!0!l - g 4 15 % "+m % n op 4 q7r 4 /+s 4 * % tvu 0 " 4 /=*=%7wyx . 0 * gz 4+{|+4 8 1 }~ o " * i " *??&???  %         ! " #$ &(' )*+ ,.- /)102+ ' 354 68789:;7=<>@?A B C > 6>(D8E >FHGHI >JA K L :NM O P 9 789 J QH9 7R S FA B F.T P CUV>(: A K : A W P9YX[Z 4(\[]^4(_8`@acbd4egfih 4(_j24kmln op pqr fgs(tuseiv(w[syx \Nz84 k;{ |N}Y]4 _~8Y?4 e? j `8?;? 4ei4(_x _ ? _8k \ _84(}??[??h _8?;? ~ 4 f s1? ?k \???x[?.x;h 4 ?e? ?m?? xN\[?8?H?js j 4 t xN\N??4Hfi4?4 f?? _84 ??8x;h 4 ei4f ~wN? ftd?u??sxf ??@4 _ 4????_ 4 x t ? e??f` ? ` z 4?ct 4 kNk X ??~ x \ _ ?? `84(_8?[? k;{?fg? ??4 ?x;h 4 e _84~ e ??_4$?N?d?ei?fH?d?g?@`8? l@z 4y???8? ?eik;{?g?H?=? w _?4~ e? _84?[t ?e ?f ? 4 eg4 ??? _ ??td?Nx;?df ?[? `8?Y` ?i~8?4e z \[f??? 4 } _8?N_8? ? h 4 e 4 ? 4 _84 x[t ?e ? ? f ?? _ 4(` ? ? 4(` ? _ `8?N? 4? 4 _ ? 4 f ? 4(x[td4(4 _?? h 4 ? e 4$` ?;? x;??? `?? ? f 4 e z 4 l z ?.~84f.? ? x h 4d? 4eg?;4 f? ? 4 `? ? ? 4ei4 _8? 4f.? _? {??? _ ? 4??? ?;?4} ? h ? xN4 ? xN??4^?nsH]8?? l 4(?8???k;4?t?? ? h 4 xN\ ?Y4 ` 4(? s{ _4~ e ??_84(? t ?e ? } ? h\N?N??x;? 4? ?d? ?.? ??k 4 l ? _? ? x 4 ?;? ? ~w ? 4 e 4 f?? _ ? 4 _ 4 x t ?e?? ? x ? `8}Y? ? ~ ? 4??? f??? 4 ? _ ?N_ ]?? ?? ? n ?_.x;h 4 ? _84 h ? `? z? 4 t f? ?N?Y4 f 4 e ? 4(f ?^? ? ? ? x ? ? _.4 fgf 4_ xN? ? ? ? { ??_?_ k \N_ 4}?eg4(]?e 4 ?fi\N? _ ?? ? 4 js ? ? ??hs? t 4 } 4 l 4? ?N? ]??d?Nx;h=s ? ? ? 4 f 4 ?H4f \ ? ???\ ? ? k { \[]_?e 4(l? ? 4(?Y? ?eg? ` ? ? 4e?4 _ ?(4.X[????N4(} _ ?N_?]8? ? _ ? h84H?8x;h84e ? ??`? x[? 4 f s ? ? ? 4 ei4 _ xs?8?8ei?8s ? h?? ? x s `? ~ f ?2f ? h 4H? }@??4 ? 4egf ? ?@??? ? ? ? ? 4e4(_ ? 4(f ?4?[t 4(4 _ fi? ?? 4 f1f ?;z 4 ? ei? ? ?(? ?N? _ f? _ xN\;? 4 ? ?8~xNx ? _?8o????? ? ? ?N4 } _ ?N_]????? 4 4(_=f h?8t _ x ? ? 4 z84 e{ f ~8?(? 4f fg??? \N_?x;h4 ? ?_ x 4(?8? ? ] 4 ? f ~ ? h ? ? s ? ] ? } s`8? ] ?u?? ?dfg? ?84 s ??w[? ? ? ? _ ? 4fgs ~ eg?8? o ?8?  35? f?s ?84 e ? _ z 4 f x ? ]8s ? 4 f t hvx h 4e x;h 4 x[?=?;h 4 f ? ? 4 t h x` ? ? 4 e 4(_ x x ??? ?j x ?;?@4 f 4 ? 4f ?e ? ?  ? xN? ?_ ? 35\[?V? s? 4 e ? f??e ?? ? ` ? ? ?8? ? ? t fYx 4 e ? e ?N4de 4 z?;4 t ?;_ ? k 4 } _8\N_ ] x ? ? ~ kNx ?!if"N4 ? ? e 4 `8? K ? x ?N?8_ ?_?? h4  ?8? k ? ?  x \[? _ ? j  x ??f ~ ?4e z ? f ? ?N4(}Y_?;_8] ? f%$'& \[]~ e 4?o ?%( ? 4 ? x \N?8_*) ?mh 4_ ` 4 f ? j2~_ ?(? ? ? _ ??t 4 ?N?? ? 4 ? sx[s f 4 x ~ f+ ?N_ x;h4.4? ? 4?H? 4(_8x ?, ? ? 132 4(f54?4H? s? 4e ? ? 4 ? x ? ? _ / f ~0 ?N4(} _??N_ ]8? ? 4 ? x ?N? _ Yj ? ?~ f ? ?Hx ?;?@4 f 4e ? 4 f# ? ` ? ? _ ? ? ? ?  ei\ ?84 f.xNh84?} ? ?? x ? x ~ e 4 f ? ` ? ? ? \[? ? q.- e 4fi4_x[fx h 4Hei4 f ~? x f ? ? ? 798;:=< > * '.02'@?BA+ * '?+ 'C:[0 ' < > *D ,@E * </ )CF -G:=H 0IKJ H < LML *ON 6 4@?4 { ?;l ? ie ? ? ? ? ? ? _ f t h 4 ei4??Nh 4 4 ?54?]84_84e ? ? ?~ xNx[?_n? ? 3 ? 4 s ?4 x;h x R ? eg? 4ZY8? h ?P _ l ? ? vOQ ? _] ? }Y4 s` ? s ? 4_ x ?;_ x ?N? } 4 ? ? 4(` ?_ x;h?4H` ? ? e 4 fgf ? ? _ j ?e x h84?? f ? hs x 4R ?egfCS ~ f 4 ? j ?e@T?s l ? 4 4VU 3 ?[fd\[f ? W _ ? ? _ ? ??xH?N? x ?84 4e 4_8?(4H? 4(xNtCX?_ ? 4?e? ? ? ? ?N? t 4 ?N? h ? ~ ? ` s1? 4.[?k 4 ?=\ ?;_84(} ? 4 f  ?x f? ` 0/ )+ * J _ x ` 4 f? 4 _ x? ? s` ?NxN? ? _ ? _ ? `dx h 4?? ? ? 4 r ?V? ?] ? } 4@? ??? ?_ ? ? s?8?e??8s ? h f4 ez ? z ? ~ 4 ( ? ? ? ?8? z 4$_ ? { ^ ] ^prqs*t  _a`Cbdce3f9ghVb=iajklhOb%mon h u g o m j qwvp t h 4e 4xf?? ?=x;h 4 k;v} _ \ _y egsx;4 Pz5{ ^}| j ? l { b }Y4?x[t ??s `? s ?(4 _ x ?e ? ? ? x ?N? _ ? ? j x ?84 >~o A K ? 7 R > 9 : s t { ? f ??@ 4 T s` ?;4 _?8??x;?v.?e? ?;? ? ? ? _ x[}Y]84(x ?? ] ? W ?Vx;h 4 e 4 ? 4 _ ? { t 4 ? ? hx td? ? h????k ] ?? q ? ?O?? ? l s xx ?N? 4C? t ? x h e 4 f ? ??x ? ?=x;h 4 t 4 ? ]h ????x;h4 _4 x[t ? e ? ? ?2_*?V? ?s xN? ? _ ? o ?%? t 4 ~ fg4Yx;h 4Y? e 4 f 4 _ x t 4(?N? h8x[f.xN? ? ? ?(~? sx 4 ? ?4d?e 4 ` ?m? ? ? ? _ f hVb ? l?? ^ | j ? ` t 4 ~ fi4?x;h4 ? ? x?t 4(?[]h?2f xN? ????~? s x 4?x;h 4Y? ??x T s` ?N4 ? x[f ?2_??~ e 4 ?? 4 e ?N?Y4(_ x[fg?f ?;_ ? 4?h b \ f B h b.? { ? e ??s(]8s ?[? _ ] ? x ? h eg??? ] h s ? ? ~ x;? ~? ? j s _ ? ? k ? _84 ? ? ?8_8_^? x ?N? _ ? f x _ 4 xNtd?e? ? t 4 j ?V? ? ~?Nx ? k s { 4 e _?4 x t ?e? t \Nx h?h ? `8` 4$_@~_\N? f ? `?t 4=? s ? ??8e ? ? s(] (?N4 t.4 ? ]?h ? ? h ? ] 4 ? ? {?s?8? ? { ?N_ ] ? h4 ? h ? _ [ k 4 x ??x;h4=]e%? ?N4 ?8x s t hu t ?;x hYe 4(f ? ??? x[? x h4Yh?;`8` 4 _ ? {4e s ? ? ? z  x ? ? _ ? ~8_ ? ? ?[?_o? ? 4 z 4e ? z }Y\ ? ? f ? j ? ?]? 4?8? f ? ?U?? ? ? ? ? _ k;{ ? ?V? f?]egs` ?N4(_ x[f ? ??? ? _ ? 4 ?e 4 f 4(_ x? ? e ? ? ? ei4(` ? ? ?[?;? _ fr? h ^? { b | jm?? ? ? o ? ? ?8?xN\ _~ ? ?;{ s` l f]e?s ` ?;4 _ ? f t \ ?Nh _ ? t 4 ? ? ]hx ? _].? ? e 4 ? 4_ ? { ?? ? l u ?N_ x;h 4 ]4(_84 e?? ? ? 4?8? ?}]? t 4(\ ]h x f x?h45??;hd???x] e  ` ?N4? x ? {dse 4 ? v _ ? {?t 4 ? ? ]h x ? j ] U ? ?Vt ? ? ? ? 4 fih ? ? _ ?N_ ?;h4 f ~ ? f 4 ? ~ 4 _8x f ? x ?[? _??N_ ? ~ e 4 ? ? kN4?? ?}]? xN4 _ `8f x[?dwN4s`=x[? ? 4 ? 4 f x^e 4 f ~ k ? f j? ea]*?@? ~ _ ` ???) ? x;h4?? ? x ? ? ?  k z ? ~4.?8j ] ? ? _ ?? ? 4 ` 4(xN4 ? ? ?? ?? ?e f ? ?? _ ?\ ? ? 4f ?%?                "!$# &%   ' ( )*,+-!" ( . /10 324 5 6 798 :;<>=9?A@ABDCEGFAHI I JKMLAN @ LAO ;QPSRHUT9V W XYZ [D\1]_^`JGH abLc Fedf9H>g9h3=9H O i3j F L <UHUk Hl H kUmnd L < H_o F HI k L ;bd O F f H_i =9F= JGH,pqrsut a HvbO FAH k>@w H O`x k HyH a g hz H j F${}| mS~ P? L gbH ; j k H F?O ??? I9?,F?g h =9H k j F F L < H { a H ; O ? H a xb?? f H O x k H J g j FDLAO ;,gb? @ OJ? r????"qr ?&? r?$?&?? ?`? qr ?S?}???"?&?? O_c ?F F f H <>= ? d LA? k F H IUI JGK?Lc F LAO ;?^ J"O x9? H < L ; F O F f H3? ?DJ"??H R O`JG? P R H?9J k F iO`? ? O gbH&J"? j IbIbBA;b? k H ? = H; c H Obi I JGHabLc d LDO ; k ??a H k cJ Lx Ha? ???9= F F O ; p ?&?Q? ? ? ?? ? ? O J j ; m ? k F H ^?nH? I JGH abL c F LAO ; I J"O x ? H <,?eRH iO ? m kG? c c H k&k LAgbH,^ JG? Lc FA? O ; k Obi Ff H k ? HUFD?? H???`?? tu??U? ?et ?-? rt ??? ? ?"??? ? ? ? r ?? t?? p? ?? L k??Aw H I JKMLc F LAO ; j F-d L? H {?? O`J F f ?9H3? ? F L < H k Hl H k3?k FAHI k ? H j&?9?"?u? @ H j c f d L < H ? kG?AHI-? RH3?G? O J <?F R O k H F k Obi I J"H a9?? c FD? O ; k?? r ?>? a ? r ? ? x ?H a O ; Ff H O x k H J g j @ LAO ; I9? J p?u? P-?urs ~"?U? H c O`? H k I O ;b?9L ;Q R H BD f F`?bI ?9j F H j F`F L < He{ L ; gbOb? g H k-F f HeFAH < I O J"h?ab? ?1??H&J"H;bc H O ?QF f H?dfbH k HeH ? = Lg hH; ? I J H a Lc F L Ob; k ?? ? ? $? ? r?}????? ? r? ? ? ? ? `? r?`?Q? ? ? r ?e?`?,? ? ?? ?? ?G?9? ? = j dAB O ;?p ? ~ k f ObR kUFf j F @w H_? h  O J"L F fb< JK = c H kF O F f H ?9??h  O JGL? < iGO`J k L ;b ?AH ? k FAH I I J"H ? LAc FAB O ; k? k L ;bc H dAwbH JGH L k ; O? H < I O? h?k F? c d = J H ?JGHgbH h ??L ;_F L <>H>? L? ?`? F f H j c F= h g9h3=9H>? k j g9?}? j&x ?AH R L Ff,F f H ? J k F O x k H&JGg j F LAO ; ^ J p ?e? ? ? r ? ~ j F L < H { j ? F f H"J H iGO`JGH q  ? ? ?  q r? ? ? O RH g9H J P i`J j <>? F L k F H I ^ JGH ? L c FAB O ; ^ JGO x ? H < P ? < I9OJ  k = c d = J H?HbBDk k L ;}d f H JGH gbH ? j F L Ob; O ? @Af H3O x k  J g j F LAO ;}gbH c F O  k  O ; ? ? B ;bH PFf H ? O,?9L ? HJ H; F L j F H @ f H FDR O h  O`JGL F f9<Uk P L `= J H ? ? HIbB c @Dk?? f H j c ? I J"O I j ? j @ LAO ;UO ??H ? O`J k ?Qk L ;9 ?9? ? a ? ? H ?BA;b??h ?9OJ ?AFf9< k : ; ?"! PuH# OJ k ?>H? H ; H&J j F%$ x ? F f H k& =b?UH a?a L%'?H J H ; c H x H FDR}H H ;_I JK L c FAH a ? a,F(  HFug h ? = H k)* ; H FAR O`J ?,F J ;b?A;b k L%+ , ?? ? ? J B H k FAO < L ; BA< L- H?F f H H. O J i ? ; c F B O ; ? / ; 0&[ X 2 Z 3[ 2bX 4Q\ a L ??H JGH ;bc H x H FDR H H; ? w H , JGHa Lc d H ? ? ? r P O J k L + ^b? ? ? ? ~ ? a?F ? ?bH ? ? $ O 1 jg ? =9H k %? 5 q ? ? PSk H H p 7? 6  = JGH?&p j ~~ ? :; ?A? L  = J 9 H 8 p x ~ ~ ? ? ? H ? ;bL ;9 < L ; L < %L : H k j a L ;? &J"H;bFuH# O`J i =9; c @ LAO ;=< Ff H ? B '?H J H ; c H x H ?DR  H ; k =bc c H k k L g9H?I J H ? L c F LAO ;9k ? ?> { ? ? a ? ? ? ?&?{ | ? ? @? ? O FAHP L ; F w H>c ? H_O ? j ; ObL k H AH k"k FAB < H k ? l H Bk AeR H ? O_; O F$H C IbH c F$d O k H H j a L ??H J H ; c HL ; I9H DO J + ? c H x H FDR}H H ; ?9? ? a?  H? ;bL ;9bP E?Ff H j c F? h g j  =bH k O i F f H$F L < H k H J LH k ?>H j c c = J j @ H a H k cJ"L I d O J k O`??F f H k ? E"F H G < Fk O =9F I =9F  : ; @f H cj k H>O i j ;bO L k ?UFDL < H k HJ L H k P-R}H c O ; H"H c F= J H}F f j Jd I  H?U; L ;9 I J"O g LAabH k ;j  H FAFAH J FAH K c f L ;  k L  ; h Ff j ; k BA<?I ? ??= k LA; ? F f H ; O L E ? O x k H J g j x  =H MON ;>Ff PDkuI j ^bHJ P R H x HbL ; x ? c O < I ? L ;b>F f 1 H Q H JGi O ? ? c H O i ? ? ? aU? ?H ?>; P ;b? O ; j ? R R?; "O S k Ha H F H TVU; L k FD? c FD?W < H k H Y X H k  [\ [Bb c d egf [ Z b c d e Zp ? o ??"?B? ? h [B] q] i [^ Z^ i [B_ Zr ? ? ? ? ?y? [? ? [B? [B? \ ???%?B???7??? ????=?;? ? ?? i [B` Zs [a [Z a [Z ] [Z ^ [Z _ B[ Z b j \ c d k a e h l m n m a ? uv3w Z ]y? xz{ Z ^G? |}3~ Z _/?  ?? q s [Z b c t e Z / ? o ? ?"? ? ?/?y?y? ? ?? ? ? [ ?? [? ? [? ? ???B?3?=???????? ??? ?%? ??? ? ?@?G?? ? ? ? L `= JGHU???J? j c b? I J"O=? j  j FALDO`; iO J ? =bIbHJg L k"H ? ? a>? H <?I OJ j J? L ??H&JGH ; c H?AH??LA;b      !#"$ % & '    )+*,.-0/123,.45 *7698:<;>=?8 4 8 ( @BADCDEGFIHIJIKMLMNIAPORQSMTVUXWYHIZI[\ OS [^]G_IA `bacAdWYeG[fLMAgZhLjih[Dk.l m npoYqYrhst O S u T ZvW!wxiIyfA \ Az i#{}| A |G~YXE A? [f? ~ ? ??[^S? wXe C L ?^? A AYf HhZXC?A LASZIA?$??ZX|.A \ H y A$O.? H? | A |.?$E A!L ?$? \ WY[fC O}?D? \.??w efA w a [ d z ??? \ A?? ag[ ???Y? Z iI??iIES AL????RW![ CRO ? AdHXe A J S O.AS ? [? ??W Ag[ \ ~ SMA??K? [ S F ? ?f?? ih? AYw?? a [??AYi a [ SM? ? ? ?.Z w A W HI? S A!?da AS A W ex[?LMA \ ih[ W [ C?? ? A ? e.A??f?V? ??? ?????x?h?????Y??????? ? S ?Z HIyf` LZ \ ? e A C E _ ? Z K L iIA O O S u T Z W wIZI[\ i??9?#? ?V? Y? ?3?f?Y? ???????x?3?Y?9?#???g? @BA { S LM? ? W \ AwIa?[ S?f??? A ~ ? e??[ S C Z \ ? ~ L Z \ ?  A ? SMAY? i e A LMA S? ? A LP? @ A {x| A?[^E.? Of??i L W [7?$A L ? \ AYHha [ SM? e ??{ | A?? \ O?E H ? \ ZIw Ld? W![ ? AYL LMA SMZ?A LM??D??? e.ZI? ? A \ E ?fZ i Lg? ? ].Zv?? X? ? ??  ? ? [fK eGZ w AY? H ES A$W[??fLZ L HhLj[ ?#? |?A LMA ? ??~ HIA O SMAY? ZXW i Zh[\ Z W ? J [ \??? ? ??? S wXe Ad? ? ? e L ? A O?KM??A?P? Q SMA ?J W HhZI[ \ [ ? [ \ ?9? [ w y A ? SMAY?fZ W!wX? [ \ L??.?$iI? [ E ? e ?^???? ~ ?Ye ? [ \ ?fZ ? ? Hh[ i y A OG~ L i{x| A | ??E.ALR[?$? A?HhZX` A W?y [fLMAY??? ~ \ ? ~9L ? \G?? A ? ? \ A ? [ ? N ? ? i ? [S?A~ ? e ?  @ A L A W [ \ ? W e ZXH A WYH ? ? ? ? ??? ?g??? ? ??   | A ? S A ? W i??  [ ? L ? A ? f W[ ^L O [\ ?f?I???Hh[ i ef? Z \ Q E HhL"] E He ?$# ? i ? e e ZI??fA \&%?\ ?IHhL(' S A ? L ~ L XZ \ ?G \ A H a?[^SM? a??Iw y | A[ E H OfE whL  ?  @ A A ihac[^S ? e ? H y A?L! A { |fA ? )*+ ?-,  ? ?/. *01&2 A43 S65 \?[f\?L ?fC 8L 7 E ? ?A ? (A 9 [ S " d a A?Z ? e.wIZ ??? ?&: ? | A ? S AY?.U W H??  [ \.L <A ; ? ?g>= ?A@ \ [ C OCB AY? wI[dwXeGA ~ WYw ? ? | ??E A L ? ace ? ?IA ? \dHXe A HXe A L E O A S | J L A ?g_IA ? ? ? \.? W ??AY? HXe AOS A ? ?IW i ? [ \ L ? A W  D W ~ L A E A(9 [ S L ? A W ~ ?XW E _ ~ iIAY??].?9A???[ \ L? W ? AL L ?GF A OS AY?fZ W!w ZI[ \ L(HI ? J&KMLONQP(RSU S T V cdj? N  V W ?6X PZY\[]R_^  V NQP(` `8a N6b ` egf ? ?gh @ \?\ A ? ? [ SM??iXS 5 \ Z \.?ji ?dA ? L A ] ~ i W e?? O ? ~ i A Lk Z?? A ?ml ?on ?G~ i A ??AYZ ? ?fH ?pRiIAS A ~ W y O ? L i?? [ E ?7e H e.A w S6q \fZ \f? ? ~ i ~ 'srgAP?hag[ SM? H S6t \u?\ ? W [ \ i ? \ ?fA!L ? \ Hwv x ? H y AA 9 [ S [ \ i efA W S[fL L K | ? U ? ~ i ZI[f\  A ? \ Z \ ?j??? ? y AgSMAY??A \ W ? LA H L 3 ~ ?\ ~ N  L [ S ]y ? JI\ L w [9ZI? W S  ??A{zO| [ S \ A i ac[ SM? LH S q ? A ??] ?~} acA Z ? eGw ? S ? ??A L_?$[ C??~?\? ? ? ? ? ??? ? ?? ??? A E LMA?HXe A ? ? A(? ? ~ w ~?I? S [ C i y A???? r ???_???-?o?-? o(?Q? ? rQ? ? ??? '?? @BA ? ~ w ~?? A v \ iIA \ L ? H ? C A ?9E S A C?A \ ? L?[ ? ~?? ?  ??ASR? \ ~ W e.~?[i J ? L i ~ i A ? A/? yfZ ]GZ HQ? ???? [ S A ?y?]??_ Z ? A?? ? \C? Z W L ??? A E LMA i e A?[ i?f ? ? ? W [fC O A w ?Xi ZI[\O? ~ i ~$O [ u \fH L?? S i Sq \ JQ? ?s? ] E i??L[??(?? ? ?R? ? S i e A SjO [?I\ i L ? [ S W S[ LL K | ?gZX?x~PHIJh[ \ [ ? [ ? S C?[?? ?? ?G # ? ?f?f? ?pE? y APS ?.[ ? \ i?Ls? [ S ? A LMi ? \y? ? A ?fA ?C? ??S[C H e A W?? C??.APi ? ? ? [.? ? ?  A L Z?\R[ S ??A S H [ ? AYH.e ?h? e A S L w ~ 3 ZIL HhZ W ? LZ ?\ ? ? W ? W A? ??2s?Y5&? 4 ? ? 0 ? ? ????? ? c ????    ] A ? ? \?[ E ? ? ~  ? L ZIL ] ? C? ? [ ? ? Z \ ? ? y A L ; E ?  ? A ? o[ ? G? [ S C?? ZQ? ? ] ? 3 y ? A ? [ S A ~ ? y??? iXG e A { F A [ EfIi O ? H E \ Z H L-?? F ? Z? W D ?? ? ? ??? ? ~ ? E ?^W ?? ? [ ? [ ? H S65 \GXZ \.? ? C A ? [^S ].[fw ? ???????/??????? ?(???t ? ? ???]? ?(?>???? ? ?<?/???? ???{? ???6? ? ?6? ??? ? ?(? ???(???? ? ??? ?6? ???/? ??(?y? ? ?\? ? ? ? ?\?/?/? ? ? ?(??? ? ? ? ?????? ? ??? ?  ?6?]?   ?  ? ? ?]?  ? ?(? ?  ? ? /? ? ??? ? ?(? ? ? ? ?????   ?(! ? "#%$'&)(+* &-,/.,103254 ,7698 4;: ? ?=<#5>?<@BADC/8BE;? F38)0) A 8G( <;F5HI" 25J18 ?]?%KML 4)2)" ION ??? ?? ? ?  ? QS (UT&;, . ,10 2;VG,7WX85VZY <[/\ C]8/EF)830/^8 ( < @B_ $Q <5<R   ? f+gihGj ? ? ? ? ??k ??  8 . &5`a8cb Ped . P "5"# ?  ? ? ? ? ? ? ? ? l ?(?Cn? m  ?A? ? ? ? ? ? 9? o?? ? ? q? p s r           !" $# % &  ' ()+* !, - ./ 10243 54687 9:;=<8><8?@BA CDE>FG8HJILK <M >A4N4D!: OQP RTS<VUG8:XW Y[Z]\ :^ @ UG C N A G F G :XI N8_ C4`Ba G :XbQc : Z]N D c8dI HJC4< C FeC < >Z]fBghXi : Z D!: \ :I ^jk :h >< i >c lm N8nogh Zp ` n o9 n < i DE> I!d nA q <Qrst>Iu v C4F I!G C wBIU G : ax ni :y C4z I >< H={8g F>| U \+} g:~ < >< A????C D :]? ???8o:4???? ? I g: ;=<4??: ; o??p m F G :]<?o: c :]oeI C4?? ?8? n$N I!:X??J>I ?8??? <8?Xg: ; < : m??+? < F k8:=? gd ; <8???< ? Zn I!:4??? G d=??c : ?N D c4: I; :+? x m :$D : mV? <+Uk8: C x i : D U G8n a w :?? C N8o i :]? u : ?? w >U G Uk :1:$? CD \ IEC ?]?en Ue: m wB? U G=? ? @wB??QI I HJ@Bg : D?UG8??? R n < m IEC C4<?????> I ? I : ? u : Z a : m?? IE? ' < Z :1F G8:L? ? U k u4D!: m ??U ?' C < ?? ? >eI m DE> c :]< ? h F G8: ? D!: m ?? Z U?eC4< ? ? ??? ? ?4D!C? :Z U: i C4< :?!U :]u > <8FCF G :?? N F N D :4 ??: @ I C=< C U: U k ? U8?k : ? ? ? N x c : ? I?I K H=> g ;??C4? ?8C U G u8? nm ?e? H I ? I K < ? : F G8:?:? CD?>I??4D!> c :< ? h=Uk :I!? :1? ? I!:$D c : m c @ N :1? ? ? ? ?]??? ? ? ? ? ??? ?4?$???!?E????!? ? ? ? ? ? ?? ? ? ? ?E????E?!?? ?E? ? ?]?+?? ? ? ? ? ? ? ? ? ? ? ? ? ?? ???? ? ? ? ? ?? ? ?? ? ??? ?? ? ? ? ? ? ? ?? ? ? ? ? ??? ? ? ? ??? ?? ? ? ? ? ? ? ? ? ? ? ?? ? ? ??? ? ? ? ?? ? ? ? ?? ?? ? ? ?? ? ? ? ? ? ?              1 ! 2 @? .  I G H JK L  A  V   Y Z  X [ W BD *E F C DNM PO'Q           (*) + - .0/ " # $ %'3 & 465 7, 89: 8;<=> A 1 3 R S TU W \ ~  ? ?} ?!? ?? `CD1: n G C N8?? N U :$D N8IFD 1<8><8?J:?8C Z GQI FD < > ? I!:]a ?`CD ?8C F G I!{ ? :$D!c >eI : m ? >? D : ? n < i=? g:];=< ?< ? h ?Q> Z @ X<8I ; :I k C4w <Bp i ? C4U k ; Z k >F : Z F { D :I?:? G > ?8? FF k : I : ? : k8n c8>eCD ? Z \ : F k : :];J<8??<8?BD n a :I w : D : c ; >: m=m { D!> < ? FD < ? ? F C nZ]Z :]g:$D n U: g: ; <8><8?Q??<JFG : ? 9 :]c :g > < A C `?U G : : ? CD Z]N D c8: I n$U ^ : u C ?$G I ? I ? N8:?FeCLn m : Z D!: \ : > <Lg:];=< >< A D n Ue: M ? kj l ? xw : C D ?P??? z _y M Pb hid e_g |{ |} f? ? f? L? 0? _y t ? ???i???s???s??? ??? E_a ?? L~s ? M 0` cPdfe_g 0v monqpsrut M ] ?_^ |? z?? ?? ? } ?? ?? ?k?s? : Z C _ ?8; : F k :=u : D ` C D H p Z :I w >UG F G :+< C t@ ?? ]: m H : ? ?? ?i??? I {; : i? ?? *? ? R ? ? J: p ? ? ? w G8d D!: I a k : < N H ? :D C ` I u o:I ? ? i ;=:1F G8:1n$Z]U N @ ? i ?4D : m ? U : m c @ {8:I s?: gC4< ? :$D?F G8: o: n ? Fe> :L?ECD F G8:=` C D!: Z \ F U G :Lg;=? : D F k : :]?? : Z Ue: mVi >? : D : < ]:=?< ? : D ` C ? < Z: ? :U w :]: < ? p i W :a k { IQ`!C Z]{ I C < ?Qc8: I U :]u D!: m > Z U >eC< I w G8:D :F G8:?m ? :D!: < ]: ? : U w : :]< W p m ? ?eI?H C IEF x C <8C N < Z : m `D!C U G :?I :U w : Z C < I!> i d D!: m ? ? ? ? E ? ? ? N D : I G C w I F G8: ? <4? >c m { @ u : D!` C D!H n < ]:I UG : ; Z]G >a: Z U N DE: g :]` U I ? i : C4`?I!:]c8:$DE@ < I `C DQF G :?F \ } C ? i > D F??c : I!F:?B?4D :]m ? ]U >C < I? : eI IG ; : i G8> i i d < N8< > FeI ^DE>? G ? I K i : ?eI?I!: ?8; n a : G > i i4:< N < ?FeI ^ p i w :?c ~=h ><B: Z]G I N ? ?8oeC F G8: ? ?  ? ?   ? UD n >< ><8? W ?  E? s?:D!:L> I <8CJg ;=? : m > ?:D :]< Z : w > U G : I n  I!>A < >? ?]p U?i >? :$D :]< Z : ? :Ue?Bd :] <  p m ?  ?!CD i >z :D!:]< F c @ N8:I C4`  C w:]c :$D ^ U G ?$D  h C w W8Y >eI ? : U U d D UG p ?  st>I?D : I { gF i :]u : i I?? Z @ g? C4<=F G8: ` nZU?Uk8nU jG : i8n Fen GQnc d :$D  :  | ? ? a p @BC "j ! ?8> ? > U n o <8C > I : F G :  n > < IEC N x ]: C4` <8C I!d ? I  N I UQU G  NQn F>  n ?e> C : Cx C ? U G  Z C < c :D!U: D ? n I ???L? f? o? ? r ? ?? ?7?? ? ??i?? ? ?s? ???? ? ?*? ??? ?i?? ? ?k???x? 0? ? f? ???? ? ? ? ???L? ???x? ?*?o? ?? P{ ??s? ?L???????6? ?s? ? ???x? ?L??? ?i? t ? ? ? i? ???_? ?? v q? ?? ??? z ? ???s}f? ? z?? ? } i{ z?? v s? ? ? sv ?w?? ? s? ? ? ? L? |v P? ?? ? 0} *? ?x? k? ? z s? ? ???????i?q? v ?? |? 0? v ?} z P? ?? ? k? ? { } ? ? w { { ? ? } ? & (*),+.-0/21430&057689- ( ? :%;,<,-0= & &9> * ) S,T V & Y*&9&0(7Z\[*&^] - : ) = :@?ACB D )7E )*F*GHF -0IJ 14K.L9M N O - :P2Q*R , D UW-0X  & _) = #%$' & j,k,&9l6 80( F :mon ) = ( )k -0p ] < b -  = ="8 -9: )OW ( q J 14r = - F4U*&9=s -0= f &t E 21 F 1 - / X &O` : 6a-0/\b 1 ( 14ced TJ & 12f9[*g : h R &i= d (F [ & 9 J & 1 f U F :u0D = - R ( _o= )*6 - S,( 1 _ ),vwD P2:Hx Y*<*4F y4) ( Y*&9F2J^&9&0( ? L -9E DVz|{ ? - ?           ! " #    $&% () (* +,- .0/ 45 (( 126 / ( => (D (( (% G 9 : ' ' '3 '7 ' ' 89 ' ? ; < ' ' ' E F H IKJML N OQPSRUT X ln V W P R mo NYX m Z[P b p q R N \ ][P degf&hQiSjk k : '@ ; '? < 'B A ' C ' HI^R N O[P p R N V X _[v PR`mwN oV ZQm a b R p q N b ][c dMhYrtsMfYs h juk h k 8 xzy|{S}S~Mg?u??z??????????~z?[?u??u? ? ~ !?!?u????!??t?Q ??? ~M???[? ?Q? ???????? ?~?u? ?[??? ?!?uy ? !??[? ~M??? ? ? ? ? !???M? ? ??????????????? ?S??? ?u??? ? ?U??!?? ? ?? ?t? ? ??S????? ? ???S??????? ? ??? ? ?t???????Q??? ? ? ?u????? yQ?Q?S??? ? ?  z ??? ?M? ???u??? ? ? ???? ????? ?t??uy[?!??? ??? ?  ??? ?Q!? ? ? ? ??  ? ? ? { ? ?  ? ?  ? ?  ~??S~  ??[? ? ? ? { ?  ?  ~t? ??? ?u? ? ~ ? } ? ?S~t??Q? ? y ? { ? ???Q???  ? ? ?M? ? ? ? !?? ? ?? ? ? ???^?? ? ?u ?  ? ? } ? ?  ? ?   ? ? ?uQ ?? ? ? ? ?u ~ ? { ? ? ? y[?g? ? zy  ? ?M?? ?M?u???? ? ? ? y  ?  ??? ??? ? ? M ? ? ??Q? ? ? ? g? ? ???  ?!  ?#" ? ? ? ? ? ?Q? ?? ~M? ? #~ ? ?! ? ? Qy ?y ? !? ?  ? ? $ ? ? ? y ? ? ?  ? } ? ~ ? ? ?uy[? ?u ? } ? y ? ?&%? ?? ? ? ~&' ? !? ???~ ? ? ??y ? ? ?S? ?  ? ? y ? { ? ( ? ? ? ) ????u? y ? ??~ ? ?u?? ?  ??? ??? ?+*? ? ? ? ?Q??? , ?  ??? ? ???.- /?? y0 ? ? ? ? y ? ?1 2 ? 35 ??? ? ? y[7? 6!~M .8 ?u? ? ?u?:? 9  ;< ? ? % ? ?u ? ? ?!? ?[? ? ? ? ? ?y ? y ? ? ? ? ? ?u?= > =@?A BC D E&B?Q? ( ?  ~M4 ? ~M?M? #~ ????!Uy ? ?y[? ? ?Q ? ? ? ? ?? ?  ?  ? ? ? ~M? ? G ?!? ( ? ??~M ? ? ?S~ ?!!I? H <  ? ? ? K ? J ? ? ? ? ? ?uy ? ? ? ? ? ? F ? ? ? ? ? ?M?? ? L M N = ? A BC OQP R T? S $ ?  U??~ WV??  ? ? 0 y[? ? ?  ? ? ? ?? ?[?  ~M y ??? ? ?  ?[? ? ? ?Q ? ?y ?  ? ? ? ?u ? ? 7? XY2 ? ? y ???[y Z $ ~M!? ?? \ . ? F? ? ??? ? ?S? ?  ) ~ ? ? ?Q  ? %  ] ? H % ^ !? ?  ? ? _ `@a ?  #~ cb ? . ? d?? ( ? ? egfIhji k l m ? ? !? ? p ??~ ? } q  ? ?   rS? ? ? ? ~ ? ?  ? s ? ! ? ? u ? t?~M ? n ? w ? v? x x ( ? ? Q? yjzi { m&| n ??|??y ? o ? ?M? ? ~  ?y[? ? yQ? ? ~ ? r ? ? ? ? !? {  ? ? ? ?y k z - ? ?M?~???S~ ? ?u ~ ?y ? ?? ?t? ?|? ? ? ?  u} }u? ? ? hjk k { ? % ?  ? ? ? ? 7? ?? ? ? ? ?Q }?? ? ? { ? > ? ?? ? ?? ? ? ?? ? ?? ? ?? ? ? ? ? ? ?? ???? ?? ? ????g? w? ?S} ~#${??? ? ?  }  ? ? ? u? ? ? ??? ? ???? ? ? ?? ?? ? ???? ???? ? ? ? ?&? ? ? ? ? ?? ????? ???? ?7? ? ?? ?7? ? ???? ? ? ? ?? ? ? ?? ? ???Q?#????? ? ? ? ? ? ?  ? ? ?  ? ? ~M?? ~# ? ?S?!g? ?? ? ? $, 6 zyQ{ ?? ??    &% ' )( , ? ?W? ? ?? ?? ? ? ???? ????? ? ?? ? ? ? ? ? ? ? ??????? ??? ? ? ? ? ??? ?? ? ? ??? ? ?7? ? ????? ????? ? ? ?#???K? ?~ ?u? ?Q?  } ? ? ?  ? ? ?S~t ?? g? ? ??  ~M?? ?# ? ? ?  ? ?  ? ? ?? !? ? ?~ ?u? ; ? M~  {U? ? ? ?   h?? ? }  ? ? ? ? ? ? ? ? ~ ? ? 5  ? ? ? ? ?~ ?? ? ~ ? y ? !? ?  ? ? ? ?  ?  } ? ?Q ~ % ?  ? ? ~ ? ? ~ ?? ?  ? ? ??~ *? ?  ~M? ?? ?? ?  ? y ? ? ~M ? ??? ?  ?? r ? ??? ?~?y [?u~ N * x ?S~  } Q?Q~g?S~  ? [? ?Q? ? ? ? ? ? ? ? ? ? [??? ? ~?? ?  % ? ?~ * % ? ??~M ? } [? ? ? ?~? Su? ? / ?z??? ?  ?  ? ? ? ~ ? ? } ? ?uy ? ? ? ?S~t ? ~ ?  } ? y ? { ? ? { ?~t? ? ? ?   S?uy[? ? ?  ? ? ??~ ? ?u?? ? } ? ? ( ?? ?S?N  ??}? y ? ?  ? ? ? ~ ? % ? y ? ?  ? ? ? q ?~ ?}??!? Q? ??? ?  } y[? ?  ? ? ??? {g? ?  ?  ? ? ?? ?u? ??~ ? ? ? ? ? ??~ y ?Q? ? ?  ? ? y ? { % ? ?  ? ? / ? ?? ? ?/ ? y?  ? ? N ?  ? ? ?~ ?u?  ? ? ? y ? ?[?uN ?  ? ? ? ?? ? {  ? ??M?4~ ? ?   ? = ?? ??    ! "$# /. 7 0  *+   ' -' , 21 3 54 *( 3                "!#$ &%' (    ) *,+.-0/1 2 +34 657 8 9':0; <>=@?0AB CED'F ? GIH J6KMLONQP J ? KMLONIRTS U J KWV NIXTY[Z J KMV N \T] P ? a bQcedfgIh i jIk lemon lep n q$r s q$ton q$t u {e| n }e~ } { ps xy z wv }e~ { n se? q n&se? s ?e?T? x? n&se? s n n??? " n $ s t ? & n $ q  ? I n ? ? ? ? $ l ? ? x? ? n&?$? ? n n ? s n&?$?T? nue? ? ? ?e? } n }$?T? x? aefO??c d cI??? ? n??? n??? np z xI? n ? { n~ { } tq ?$t?? ?e?T? } ?s } ?? x? n s ?q ?$~ s x? ? ?{ ? w? ? w? j? n se??? n n h u ? n? l n np ? n n? ? n q tq n u t? n u t?? x? n uI? { n uI? ? ?$?&?I?o? n ~???I?'? ? ?'??? ???? ? ? ?I? ?? ?????? ? ???? ? ? ?? ?M? L?? P LO?>? ?E?6? ?>? ?? _?? ? ? ?? ? ? ? ?I? ?"?i ??? ? ? ? ? ???W? ? ? ?O? ? ? ? ?M? ? ?? ? ?? ? ? ? ? ? ? ??? ? ? P?? ?"?? ?1? ?1?T??O??O??? ?? ? ? ? ?  ?  ?? ?  ?    ? ?   ?  ? ? ?  ? ?    ? ? ?   ?  ?   ?     ?   "  # ?  ?  ? ? ? ? 14 *3 ?) ??? / ? ? 0 ?2 7 ?  8: 9 ? ? ? ?? ?A@ <>= ? E? D F ? R ? ? ? T ? ?? ???1? ! . GXWKY ? ? ? ? ? ? ? k Xl U $ +. GIHKJ L ^`_ ? $v w n u s$ }Iu s0? ? u s  ?>n np?? s se? s ? s$? q ? seh { l s p? { s w? ? ? ?W?o? ? ? ?? ? ? ?'? ? ? ?"? ? ?"? ? ? ?? ?"??? ???? ?? ? ??   ??   ?     ?  ?  ?    ?? ?  ?? ?    ?  ?   %'&( ?    ?? ?? , -5  ? 6  3    ?  ? ? 3(;  ? ? ? ?  ? C ?0 ? ? B ? ? M= N ? < Q 7 OP ?  ? ? ?  ? ? ?S V T G ?X] GO?K^ _ GO?Ib &? ?dZo?Kc([ e \ e fKg "? ? ? c'?4` h a i fKj ? m ?? ? ? ? ? ?Kn ? ? ? X? o*p ?(m ? ? ? ? ? ?rqI??o? ? ? ? ? ?Xs ? n ? ? ?L ?utt s 2 u P ' D K MJ eB ? \ ? ? ?? ? ? ???X? ? ? ? ?:??? ? ? p ??? ? ??? ? ? ? ? ? ? ? ??? ? ? ?? ? ????? ? ? O ? ? ? ? ? ? ? ? ? ? ?? ?? ? ? ? ? ? ? ? ? ? ? ??? ? ? ? ????? ? ? ??? ? ? ? ??? ?? ? ? ? ???'? ? ? ? ? ? ? ? ? o ? ? ? ? ? ? ? ? ? ? ?WL ? ?? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? q ???'??? ? ?'??n ? ?"P ? ? ? ? ? ? ? ?? ? ??? ? ? ? ? ?? ? ?$? ? ?? ? ? ? ? ? L ? ?o? ? ?? &? ? ? ? ? ? ? ? ? ?? ? ? ? ? ??? ??? ??? ? ? ? ? ? ? ? ? ?'? ? l ? ?I? ? ? ? n ? ? ? ????? ? ? ? ?}? ? ?  ? ? ? ? ? ? ?? ?? ??? ? ? ? ? ?  k ? ? ? ? ? ?? ? ??? ? ?? ? ? ?????? ? ??q ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ?? ? ? ?W? ? ? ? ? ?o? ? ?? ??? ? n p ?'? ? ? ? ? k ??? ? ? k ? ? ? ? ? n ? ? ? ? q ? ? ? ? ? ? ?? ? ? ? ? ? ? ? ? e? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? n?? ? ? ? ? ? ? ? ? ?2? ?,? ?? ? ? ? ? ~ ? k ? ? ? ? ? ? ? ? ? ? ?'? ? ? ? ? ? ? ? ? ? ? ? ? ??? ? ? ? ? ? ? ? ? ? ?.? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? l?? ? ? ? ? ? ? ? ? ?I? ? ? ?o? ?o? ? ? ? ? ? ~ B v :wyx{z}|:~??? |:?u? }?'x ???? |:?u? :|?|:?'? ~ ?q q ?? X? ?? d?}? ?? q p '? ??u? ?? }? q ?? '? '?'? ? k n p n n ?? q '? ??p '? Xs '? ? n (? *? I? ? ??? :? :? ? ? k Xl n ?r? n } ? ? q ? ?? n l p ? ? ???d???*?4?????4????? ? ?:? ? ? ? ? ? ? ? ? ? ? ?}? ? ? ? ? ? ? ? ??? ? ? ??? ? ? p?? ? ? ? ? ? ? ? ??L ? P ? ? ? ? ? ? ? ? ??? ? ? ? ? ? ? ? ? ? ? ?:??? ? p ? ? ? ? ? ? ? ? ? ? ? ?'? ? ? ? ? ? ? ? o? ? ? ? ? ? ?o??? ??? ? ? nI? ? ? ? ?I? ??? ?? ? ? ? ?? ? ? ? ? ? ? ? ? :? ? ? ? ? ? ? ? n ? ? ? ?I? ? ?   !" #%$'&() +* , -   / UWVYXZF\[;] / _K ^;/`G / Eba > EbcdUZJ/ >fe;g >6h0i .0/%1320465879;:<=?>A@CB'D /4 / E#FHG / E I G6/ JK 02 L %/ MN8/PO >AQSR6T P G T Bj 2 5 c I'k 4lB I G / J 2_mnJ 6 j o K B I / prq;s0t6/ j /u_p / E"v 2_462 Ibw 2 J8J n i B+> p_20Oyx_z /E V B p B|{ Ib/ j}2_~ h0i / 9nB 4s_8? / ??> @ G / K 2 h 2 > 4? ? > E?K i 2 I? > N8? 4 >? Ib/?K_? cS/ I / E 2 /I j B p B'? / m 4fj ??? 2 p i ? 4?? B J i G T  ?;2 D c M p t/ ? / T 46? 4 5 J { E L /? 7 a > ESp0?6/ ?_4;j? L 2_j { F > { s G { h I#? j > 4 >6? ? / G /? jd>6~ K i /?I G6? 20m J?T J i 2 K / J p { Eb? >? I"? TY??>? I / G T  p / i 2 j6j / 4 { 462 K?I ? ?H? J E > I I K i / h N > G ( T B j62_D c?I M s i / . / TS4 204 Dy? {6? L / I 2_4S96? I i;> N B|._TSD / E/+? > E p E B+j /? >? ? / T 4 204;5n? ? > 4n? I??;s i /8204 ?? L;20j {;? >? K_G6?6K?? p ?;U?? ? 4?? ? ? > E p_i / ? > 4;D/ I K ? / B ??K_?_c / J > 4 I 2 ?;/Eb/ j??rmn? /? I K_/ R?? / B j GE ? ?0J K 2 > 4 I?b? h?i / j62 ?'/+Eb/4 J / 2_I?c > I p;? E > 4 > { ~ J%/ Vn?? 9;: >{ p R /E ? >? I ? ? g / K Ny??4?? : U j?? ? ? 20p i 204?? :??D 2 L 2_4 D p i /?46/ ??N > E?? B ?6?;20p ? > 4 F?K z ?6Iy??I { J i z?4 > p > 4 .0? GAE /?;20JK ? 4 D K ?6/ mL /?bI h / G??Z/ B ? a > Eb/ J z p g?6p F?I > K i / 204 h / ? / j6??B K / I p_/ R I8? ? ?nI ? ??4 5 I i TS? i ? V ?6/ 4 { 4 2_p I#? Vf2_jS46> KnJ i U D / w ?;/ R / E ??>? UZJ/\O > c R;T / ? w > B I ? 4 D x / > { K0G;? p Ny? ? w ??Ib/ R;T?B p_/ ?;? ?? j / ~ { 4;20p I ? ( ??? / g ?? p J i;> 20J /?> a6? /E / O/ 46J ? N8/ ? 5i s? B R6R / T I;K >'g /? 4 K ? /Eb? 5 / > a6????Y??? ??? ?? ? ? ? ?? > E L /E#Ib? I p i / B JK { ? /? > E2 I B{ I / a { . 46/N j;2?B 5 4 > I K_??J?G TSp_? Ju?;? TS? ? ??? ? > p_p ? ??4 D w ?6/?? ( >6?Y> { s_? > a ? I#??G?_/ j;B p B a > E4 > ? I ? G E > g6. / c I? ?/ T 4 ?4 D ?4 / G / ? >? U Ju/ > a ? G E /Ib/46p2?4 h0i / ? B p_B ? ? > E K i /8. > N?4 > ? I 2 4?? I 2 4 5  / T 4 ? 4 D?> L / E 9 ? ? snR E Eb/ ? N > E?? j ? p?B|8 N i /Eb/ >AL / p_p?204 5 2 I B ? ? >c G T?? I >? p > ? :?j / G /+4 j I > 4 0p ?;/ ? > { 4 h'> a 4 > 2 I / / _p 2 c / I / ? ?0/I ?nIb/ j 204YK_?62 I?G B G;/E K i / Eb/ ? I ~;> B j6L U K? 5 /  R T?B j 5 c I > 4?4 > 2 I ? / I?/ 4K;N ? /SJ >c R TS2 4 5 K i /yp?N > ; I?/ ? 2_> ? IJu?6F  / 4;D /? !#"%$ &')(+*-,/.102 3/4 ,6587:9<; (>=+?A@CBD? 5 *<EDFHG ;<I IJK (ML>?@ FN3 ? @O3 JQP L>@<RS=+4 ')E>TVUWXJQY J @ (>, = L>?<Z 2\[ @] 5 ',/G &J L R J @ ] ,^36_ @<?-`Na J 7 R JF F;Abb ?Ac!B 5 ? T ( *<'edf,/( L ? Z ,6Wg9<3Qh+J Z 3/'ji ?k Z ] , ( E ? Z-lm J GnJo65p3 4rq Z E =+L , = LX? Zts 5 ,6@<u d ?-v mNq [ wyx 9-zp{<|} {A~-?<? ???A???  A? ????? 5pJQh>T?, @ , @ 7??/? ??6i? JQ7 T , @???? { {#?<?<?+? [?? J `r? ?<? _g,Q(/?);AW (>EMbAW JN? k (+bO; u K v ? ? [?? F(X5p,/?Q(D?6d J ;A5,6W ? J ( `N? 5D?F B?? 5 x ?<Y b ;<? 2 @ R<? 9 @ ? ` ? E+5 ] ?/??? [ b ?W ?6{ { ? ? ( x ,6?, @ ,????/{ { ?<???+? ?);WX? u , F_ x ? @-Z JQ3 = E ? @ EXF ( ?%Jo??eE ZAR ( ? ? q Z  ?<?  6?-? ? A? $ ??g?+? ??? ???? ? ??? ? 2 ? ? ? $/? ? ? [? ? Q? $ &O???? &A? ? ? ? ?X? ' ] E ( JQ7 ??? ??? x (O???<? JQ5 ??? v 9 T ?<a J @ G? ??A? ????A? ? ;5pJQ?N_-??-?^?A? ?-? WXTj, @? , Z 7 [? 9??/? J^E R 'QZ 7-?/b?? |?<?A?p? ~ {-? ? h>? F7 ,/? JA? d??Q?? ? 5 W ? , k T [ F?K ?A3 E , u J?6? l ?v [  ?  d ? LX ???$ ?6@  7  [j? ? ?  ? &e? JQ ? E R J @  ]  2 ? ?? ?6{ {A ?<? ?? ? #?+? b? $J#  o/- ? "L>@-! 6R ?? /? `? ? $3/,# W %& ? (' ? *5 )? "? + ,6,?5 K ? ? BX? ?? $$5  ?  - ?A?@? W " @ . J , 5 , m0Z J / I 5 ,Q' Z K63/F EDL F ? 3 ?2 13 qpZ ? ? [54 ? ? 5 R , 7@ 6 ,/; B T?,/@@ ? m ? 2? 86? ;<:( 9 ? @ ; ? {  = < &? > ? , ?E @-R ( A? @ 5?J 7 E 3/( ? C? B J ? J := D#? 7 F ? B ? J T b ? 5o6W ? :E E?JQ5 J @?3 JAF v F&GH  ? ? I ,  KJ  -? 4  L MON? { P ? QR ? J F , k 5 ?N? ? {A { ST< U&U @ 5 ,Q3 (>E 3 , "? V ??F; J F h @ ? J Tjb ? 5p, W ? WE E JQ5DJ @<3 JyW>JQoQ? L>@ R ? X&Y ?  ?/? ? Z [ J;\6 -? ] ^ L$_ `v ??A? ap ? b ~ c ? v [ o A? ? ? ?[ &r $6J^?  E R  J ? @A ? e] ??d ? ? d  ?  [  ? $6s ? p J m 5F JQ*<,/J 7 @ E f J rR W 7 q ? g?JQ7 F ? ? 7? { 7 h EDF ? ? *? @ i u &? 2 ? J AwF vj J  (    [  ? =? k Z ? l R ? m A? ? ?  $k ^ L ?+?  m T? n ?  ? ? [5st [ ? &eJ^E R J @<2 ~ ? 7 / J > W * ? { ? J { z , , 6 ? ? m < ; Y , ~ }  ; @ ( | 6 ? { [` E ?2x x ?@ @ J 3] ? y ?A@ ED[ F ("? ? J (X? `g? 5X_ F? ?&? F q @0 ? ?   ? ?  #6 ??? ? ? ?  ? ?  ?  &? > m? 5? JQ ? 7 / ? 3? u E$/@ ?>?R ?  - ? ?; J @ 7FbE (+?J 7( F ,6@ 7 ? ? ?<3 3 ,D F , 7 @ , R R JWX??E(? o ,/( @ J7 K ? ? ? B ? ? ? ; ? , @ K_ ? b ?A|{O ? ?r| ? ? ??J ] `N? ? 7 xNh ( ?? x 7[ ?? [ ] ? EXF ? @ ? & J FW J ? ?            
940 |@word ixx:1 km:1 r:2 n8:7 ld:1 kuf:1 bc:4 ts2:1 ka:1 ixj:1 dx:1 bd:5 obi:3 wx:3 aps:1 rts:1 tdp:1 eba:1 lr:1 lx:2 h4:1 mnc:1 xz:2 ee6:1 td:5 wbi:1 gqi:1 qw:1 kg:2 sut:1 jik:1 x5p:1 ag:2 j62:1 uk:1 j24:1 uo:1 cdp:1 io:2 api:1 iie:1 dnm:1 ap:1 kml:3 ibi:1 au:1 eb:9 p_:2 bi:2 uy:7 fah:5 yj:1 vu:2 epr:1 xr:2 ofe:1 fbh:1 oqp:2 lnr:1 fkg:1 bh:4 nb:1 knk:1 go:2 l:1 eix:1 jgo:1 hd:1 fkj:1 obr:1 qh:1 jk:1 ep:5 cro:1 rq:2 pd:3 pdj:1 ui:2 zar:1 uh:1 po:2 mh:1 jgh:13 zo:2 epi:2 q7r:1 ibh:1 mon:4 ace:1 wg:1 dab:1 ip:1 a9:1 rr:1 s6t:1 fr:1 j2:1 tu:1 hid:1 wgr:1 tvu:3 g9:1 qr:1 epx:1 p:1 zts:1 zp:1 e_a:1 iq:1 n_:13 ac:2 op:2 h3:3 sa:1 eq:1 p2:1 c:1 utt:1 f4:1 vx:1 hvb:2 ja:1 hx:2 exf:2 im:1 lma:6 k3:1 cb:2 lm:2 g8:14 sma:2 fh:3 xnx:2 gihkj:1 iw:3 puh:1 wl:1 r_:1 ej:2 hj:1 og:1 jdr:1 q3:1 pdi:2 vbo:1 kc:1 jq:4 fhg:1 iu:1 whl:1 knx:1 eik:1 h_:1 x_:1 dfg:1 lac:1 abl:3 m4:1 tq:2 ab:2 efa:2 fd:4 a5:1 lep:1 eiv:1 sh:1 xb:1 egfih:1 iv:1 rqi:1 i3j:1 wb:2 zn:1 a6:2 snr:1 iee:1 acb:1 kn:1 aw:2 sv:1 st:4 ju:3 ie:5 iy:1 na:1 nm:1 x9:1 lmi:1 dr:1 tsd:1 bx:1 yp:1 de:7 eyi:1 mv:1 ad:1 nqp:1 bg:1 ayi:1 aca:1 iwi:1 oi:1 ir:2 ni:2 qk:1 rx:1 zx:2 j6:1 ah:5 wiz:1 ed:1 ty:1 di:1 ped:1 ut:1 pjq:1 ea:1 fih:1 o6:1 oei:1 iaj:1 d:1 nyx:1 xzy:1 ei:12 su:1 xnh:1 ye:1 cdz:1 iop:1 wp:1 eg:5 ue:5 m:1 o3:2 ay:5 tjb:1 gh:2 fj:1 ef:1 fi:2 rmn:1 igj:1 ji:4 nh:3 he:1 fec:1 zil:2 uv:1 fk:1 iia:2 zia:1 hjk:1 dj:1 jg:3 hxi:1 ikjml:1 igo:2 fgs:2 wv:2 vt:1 xe:1 lnp:1 fen:1 ced:2 ey:2 jgi:1 s65:2 ii:4 xf:1 af:2 heh:1 y_:1 seh:1 qpsr:1 qi:2 df:1 bz:1 ief:1 cz:1 ion:1 zw:3 sr:1 i57:1 ee:1 wn:1 zi:6 cn:1 tus:1 jgg:1 gb:1 wo:1 gnj:1 hia:1 eai:1 se:5 gih:2 rw:1 fz:5 efe:1 rb:1 awf:1 iz:3 oop:1 zv:1 xwx:1 pb:1 opi:4 uw:1 i5:2 vn:1 i9:3 ob:2 dy:1 fl:2 ct:1 hi:2 aer:1 i4:1 k8:2 lemon:1 nib:1 fei:2 bp:1 ri:1 c8:3 o_:1 jr:1 dqf:1 h0i:3 ur:1 wi:1 lp:1 n4:11 ayf:1 s1:2 ikj:2 pr:1 xo:2 g6:3 lml:1 ln:1 hh:4 fed:1 b_:2 hii:1 wyh:1 jd:4 jn:1 cf:1 oyi:1 yx:2 xw:3 cbd:1 bl:3 jgk:1 fa:6 rt:1 prq:2 ow:1 oai:1 hq:1 oa:1 nx:4 mvi:1 y5:1 s_:2 y4:6 edl:1 ibb:1 ql:1 lg:3 sjo:1 fe:2 ba:9 fdr:4 sm:6 t:2 iui:1 lb:1 wyx:3 bk:1 fgf:2 c4:11 fab:3 wy:1 h_i:1 oj:2 p2q:1 ia:2 cdj:1 hr:1 lao:10 yq:2 xg:1 hm:1 gz:1 knj:1 gf:1 kj:2 zh:1 nxn:1 vg:1 xp:1 s0:1 edf:2 i8:1 pi:1 cd:1 lo:1 qf:1 iej:2 pzy:1 l_:1 fg:3 dip:1 xn:11 egs:2 gxz:1 fb:1 far:1 ec:3 qx:3 bb:1 sj:1 opp:1 ml:2 gae:1 xi:4 un:1 zq:1 fao:1 wia:1 mj:1 k_:5 iec:3 qz:1 ioo:1 vj:2 da:2 uzj:1 f9:3 en:1 ny:2 yjz:1 xh:1 xl:2 pe:1 ib:6 yh:1 ix:1 er:1 badc:1 ton:1 dk:2 x:2 a3:1 ih:2 a43:1 ci:1 f1f:1 j6j:1 fc:3 ez:1 bo:2 syx:1 dh:1 u8:2 jf:1 psk:1 fw:1 edr:1 egf:4 rso:1 la:3 zg:1 e6:3 d1:1 ex:1
7,179
941
Deterministic Annealing Variant of the EM Algorithm N aonori U eda Ryohei N alcano ueda@cslab.kecl.ntt.jp nakano@cslab.kecl.ntt.jp NTT Communication Science Laboratories Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-02 Japan Abstract We present a deterministic annealing variant of the EM algorithm for maximum likelihood parameter estimation problems. In our approach, the EM process is reformulated as the problem of minimizing the thermodynamic free energy by using the principle of maximum entropy and statistical mechanics analogy. Unlike simulated annealing approaches, this minimization is deterministically performed. Moreover, the derived algorithm, unlike the conventional EM algorithm, can obtain better estimates free of the initial parameter values. 1 INTRODUCTION The Expectation-Maximization (EM) algorithm (Dempster, Laird & Rubin, 1977) is an iterative statistical technique for computing maximum likelihood parameter estimates from incomplete data. It has generally been employed to a wide variety of parameter estimation problems. Recently, the EM algorithm has also been successfully employed as the learning algorithm of the hierarchical mixture of experts (Jordan & Jacobs, 1993). In addition, it has been found to have some relationship to the learning of the Boltzmann machines (Byrne, 1992). This algorithm has attractive features such as reliable global convergence, low cost per iteration, economy of storage, and ease of programming, but it is not free from problems in practice. The serious practical problem associated with the algorithm 546 Naonori Veda, Ryohei Nakano is the local maxima problem. This problem makes the performance dependent on the initial parameter value. Indeed, the EM algorithm should be performed from as wide a choice of starting values as possible according to some ad hoc criterion. To overcome this problem, we adopt the principle of statistical mechanics. Namely, by using the principle of maximum entropy, the thermodynamic free energy is defined as an effective cost function that depends on the temperature. The maximization of log-likelihood is done by minimizing the cost function. Unlike simulated annealing (Geman & Geman, 1984) where stochastic search is performed on the given energy surface, this cost function is deterministically optimized at each temperature. Such deterministic annealing (DA) approach has been successfully adopted for vector quantization or clustering problems (Rose et al., 1992; Buhmann et al., 1993; Wong, 1993). Recently, Yuile et al.(Yuile, Stolorz, & Utans, 1994) have shown that the EM algorithm can be used in conjunction with the DA. In our previous paper, independent ofYuile's work, we presented a new EM algorithm with DA for mixture density estimation problems (Ueda & Nakano, 1994). The aim of this paper is to generalize our earlier work and to derive a DA variant of the general EM algorithm. Since the EM algorithm can be used not only for the mixture estimation problems but also for other parameter estimation problems, this generalization is expected to be of value in practice. 2 GENERAL THEORY OF THE EM ALGORITHM Suppose that a measure space Y of "unobservable data" exists corresponding to a measure space X of "observable data". An observable data sample ~(E X) with density p(~; @) is called incomplete and (~, y) with joint density p(~, y;@) is called complete, where y is an unobservable data sample 1 corresponding to~. Note that ~ E'R,n and y E 'R,m. @ is parameter of the density distribution to be estimated. Given incomplete data samples X = {~klk = 1"", N}, the goal of the EM algorithm is to compute the maximum-likelihood estimate of @ that maximizes the following log-likelihood function : N L(@;X) = Llogp(~k;@)' (1) 1:=1 by using the following complete data log-likelihood function: N Le(@;X) = I: logp(zk' Yk;@) ' (2) k=l In the EM algorithm, the parameter @ is iteratively estimated. Suppose that @(t) denotes the current estimate of @ after the tth iteration of the algorithm. Then @(t+1) at the next iteration is determined by the following two steps: 1 In such unsupervised learning as mixture problems, 1/ reduces to an integer value (1/ E {I, 2, ... , C}, where C is the number of components), indicating the component from which a data sample x originates. Deterministic Annealing Variant of the EM Algorithm 547 E-step: Compute the Q-function defined by the conditional expectation of the complete data log-likelihood given X and @(t): Q(@; @(t? ~ E{Lc(@;X)IX, @(t)}. M-step: Set @(t+l) equal to @ (3) which maximizes Q(@, @(t?. It has theoretically been shown that an iterative procedure for maximizing Q over will cause the likelihood L to monotonically increase, e.g., L(@(t+l? 2: L(@(t?. Eventually, L(@(t? converges to a local maximum. The EM algorithm is especially useful when the maximization of the Q-function can be more easily performed than that of L. @ By substituting Eq. 2 into Eq. 3, we have N Q(@;@{t? ~J = ... J{logp(~j:'Yj:;@)} k=l N IIp(Yjl~j;@(t))dYl . .. dYN j=l N ~ J {logp(~k' Yk; @)}p(Ykl~k; @(t?dYk' (4) k=l @ that maximizes Q(@; @(t? should satisfy aQla@ = 0, or equivalently, N ~ k=l Ja {a@ logp( ~k, Yk; @) }P(Yk I~k; @(t?dYk = o. (5) Here, p(Ykl~k' @(t? denotes the posterior and can be computed by the following Bayes rule: (6) It can be interpreted that the missing information is estimated by using the poste- rior. However, because the reliability of the posterior highly depends on the parameter @(t), the performance of the EM algorithm is sensitive to an initial parameter value @(O). This has often caused the algorithm to become trapped by some local maxima. In the next section, we will derive a new variant of the EM algorithm as an attempt at global maximization of the Q-function in the EM process. 3 3.1 DETERMINISTIC ANNEALING APPROACH DERIVATION OF PARAMETERIZED POSTERIOR Instead of the posterior given in Eq. 6, we introduce another posterior f(Ykl~k)' The function form of f will be specified later. Using f(Ykl~k)' we consider a new function instead of Q, say E, defined as: N E ~r ~ J {-logp(~k' Yk; @)}f(Yk l~k)dYk' k=l (7) 548 Naonori Ueda. Ryohei Nakano (N ote: E is always nonnegative.) One can easily see that (-E) is also the conditional expectation of the complete data log-likelihood but it differs from Q in that the expectation is taken with respect to !(lIk I:l:k) instead of the posterior given by -Q. Eq. 6. In other words, if !(Ykl:l:k) = P(lIkl:l:k; B(t?, then E = Since we do not have a priori knowledge about !(lIk I:l:k), we apply the principle of maximum entropy to specify it. That is, by maximizing the entropy given by: L J{log !(1I k I:l:k)} !(Yk I:l:k )dllb N H=- (8) k=1 J with respect to !, under the constraints of Eq. 7 and !dYk = 1, we can obtain the following Gibbs distribution: 1 !(lIk I:l:k) = exp{ -{J( -logp(:l:k.lIk; B?)}, (9) Z:l:" J where Z:I:" = exp{ -{J( -logp(:l:k, 11k; B?)}dllb and is called the partition function. The parameter {J is the Lagrange multiplier determined by the value E. From an analogy of the annealing, 1j {J corresponds to the "temperature". By simplifying Eq. 9, we obtain a new posterior parameterized by {J, !( I) lIk:l:k = p(:l:k, 11k; B)fJ Jp(:l:k, 11k; B)fJdllk . (10) Clearly, when {J = 1, !(lIkl:l:k) reduces to the original posterior given in Eq. 6. The effect of {J will be explained later. Since :1:1, ..? ,:l:N are identically and independently distributed observations, the partition function Zp(B) for X becomes Ok Z:l:". Therefore, II JP(lIbllk; B)Pdllk ? N Zp(B) = (11) k=l Once the partition function is obtained explicitly, using statistical mechanics analogy, we can define the free energy as an effective cost function that depends on the temperature: (12) At equilibrium, it is well known that a thermodynamic system settles into a configuration that minimizes its free energy. Hence, B should satisfy oFp(B)joB = O. It follows that (13) Interestingly, We have arrived at the same equation as the result ofthe maximization of the Q-function, except that the posterior P(lIk I:l:k; B(t? in Eq. 5 is replaced by !(lIk I:l:k). Deterministic Annealing Variant of the EM Algorithm 3.2 549 ANNEALING VARIANT OF THE EM ALGORITHM Let Qf3(@; @(I? be the expectation of the complete data log-likelihood by the parameterized posterior f(y" I~,,). Then, the following deterministic annealing variant of the EM algorithm can be naturally derived to maximize -Ff3(@). [Annealing EM (AEM) algorithm] 1. Set {3 +- {3min(O < {3min <t:: 1). 2. Arbitrarily choose an initial estimate @(O). Set t +- O. 3. Iterate the following two steps until convergence2 : E-step: Compute QfJ(@;@(t? N =~ "=1 J {logp(~", y,,; @)} I p(~" y . @(t?f3 , '" dy". p(~", y";@(t? f3 dy,, (14) M-step: Set @(t+l) equal to @ which maximizes Qf3(@; @(t?. 4. Increase {3. 5. If {3 < {3max, set t +- t + 1, and repeat from step 3; otherwise stop. One can see that in the proposed algorithm, an outer loop is added to the original EM algorithm for the annealing process. An important distinction to keep in mind is that unlike simulated annealing, the optimization in step 3 is deterministically performed at each {3. Now let's consider the effect of the posterior parameterization of Eq. 10. The annealing process begins at small {3 (high temperature). Clearly, at this time, since f(y"I~,,) becomes uniform, -Ff3(@) has only one global maximum. Hence, the maximum can be easily found. Then by gradually increasing {3, the influence of each ~" is gradually localized. At {3 > 0, function Qf3 will have several local maxima. However, at each step, it can be assumed that the new global maximum is close to the previous one. Hence, by this assumption, the algorithm can track the global maximum at each {3 while increasing {3. Clearly, when {3 1 the parameterized posterior coincides with the original one. Moreover, noting that -Fl(@) L(@), {3max ought be one. = = 4 Demonstration To visualize how the proposed algorithm works, we consider a simple onedimensional, two-component normal mixture problem. The mixture is given by p(x;m},m2) = 0.3kexp{-~(x - ml?} + 0.7*exp{-~(x - m2)2}. In this case, @ = (mb m2), y" E {I, 2}, and therefore, the joint density p(x, 1; ml, m2) = 0.3$ exp{ -~(x - ml)2}, while p(x, 2; ml, m2) 0.7 $ exp{ -~(x - m2)2}. = One hundred samples in total were generated from this mixture with ml = -2 and m2 2. Figure 1 shows contour plots of the -FfJ(ml,m2)/N surface. It is interesting to see how FfJ(m}, m2) varies with {3. One can see that a finer and truer structure emerges by increasing {3. Note that as explained before, when {3 1, = = 2When the sequence converges to a saddle point (e.g., when the Hessian matrix of -Fp(9) has at least one positive eigen value), a local line search in the direction of the eigen vector corresponding to the largest eigen value should be performed to escape the solution from the saddle point. 550 Naonori Ueda, Ryohei Nakano -4 -2 0 -2 2 (a) ~ .. 0.1 -4 -2 0 2 0 -4 2 -2 (b) ~ - 0.25 -4 4 ml -2 0 2 2 (c) ~ '" 0.3 -4 4 ml -2 0 (f) ~ (e) ~ .. 0.7 (d) ~ = 0.5 0 2 =1 Figure 1: Contour plots of -Fp(mhm2)fN surface. ("+" denotes a global maximum at each JJ.) -4 -2 0 2 (a) (~),m~~ = (4, -1) 4 ml -4 -2 0 (b) (m(~),m~~ 2 =(-2,-4) Figure 2: Trajectories for EM and AEM procedures. 4 ml Deterministic Annealing Variant of the EM Algorithm 551 -Fl(ml, m2) == L(mb m2). The maximum of -Fl (or L) occurs at m1 = -2.0 and m2 1.9. As initial points, (m~O),m~O? (4,-1),(-2,-4) were tested. Note that these points are close to a local maximum. Figure 2 shows how both algorithms converge from these starting points3. = = The original EM algorithm converges to (mp2), m~12? = (2.002, -1.687) (Figure (1.999, -1.696) (Figure 2(b?. In contrast, the pro2( a? and to (m~19), m~19? posed AEM algorithm converges to (m~43), m~43? = (-2.022,1.879) (Figure 2(a? and (m~43),m~43? = (-2.022,1.879) (Figure 2(b?. Since these starting points are close to the local maximum point, the original EM algorithm becomes trapped by the local maximum point, while the proposed algorithm successfully converges to the global maximum in both cases. = 5 Discussion A new view of the EM algorithm has recently been presented (Hathaway, 1986; Neal & Hinton, 1993). This view states that the EM steps can be regarded as a grouped version of the method of coordinate ascent of the following objective function: J ~f N N I: E (lIlo) {logp(:l:l:, 111:; @)} - I: E (lIlo) {logp(lIl:)} p p I: (15) I: over P(lIl:) and @. That is, the E-step corresponds to the maximization of J with respect to P(lIl:) with fixed @, while the M-step corresponds to that of J with respect to @ with fixed P(lIl:)' Neal & Hinton mentioned that apart from the sign change, J is analogous to the "free energy" well known in statistical physics. It is worth mentioning another interpretation of the derived parameterized posterior; i.e., it plays a central role in the AEM algorithm. Taking the logarithm on both sides of Eq. 10, we have (16) Moreover, taking the conditional expectation4 given :1:1:, summing over k, and using Eq. 12, we have N FfJ(@) = 1 N I: EJ(lIlol:l:lo){ -logp(:l:I:, 111:; @)}+ pI: EJ(lIlo 1:1:10) {log 1(111: 1:l:1:)}? (17) 1:=1 1:=1 = From Eqs. 7 and 8, Eq. 17 can be rewritten as FfJ(@) E - !H. Noting that l/f3 corresponds to the "temperature", one can see that this expression exactly agrees with the free energy, while Eq. 15 is completely without the "temperature". In other words, FfJ(@) can be interpreted as an annealing variant of -J. Clearly, when f3 = 1, FfJ(@) == -J. 3 Although the AEM procedure was actually performed for successive fJ (fJnew - fJ x 1.4), the results a.re superimposed on -Fl(ml, m2)/N surfa.ce for convenience. 4Since the LHS of Eq. 16 is independent of 111" it does not cha.nge a.fter expecta.tion. 552 Naonori Veda. Ryohei Nakano The proposed algorithm is also applicable to the learning of the (Generalized) Radial Basis Function (RBF, GRBF) networks. Indeed, Nowlan (Nowlan, 1990), for instance, proposes a maximum likelihood competitive learning algorithm for the RBF networks. In his study, "soft competition" and "hard competition" are eXperimentally compared and it is shown that the soft competition can give better performance. In our algorithm, on the other hand, the soft model exactly corresponds to the case f3 = 1, while the hard model corresponds to the case f3 -+ 00. Consequently, both models can be regarded as special cases in our algorithm. Acknowledgements We would like to thank Dr. Tsukasa Kawaoka, NTT CS labs, for his encouragement. References J. Buhmann & H. Kuhnel. (1993) Complexity optimized data clustering by competitive neural networks. Neural Computation, 5:75-88. W. Byrne. (1992) Alternating minimization and Boltzmann machine learning. IEEE Trans. Neural Networks, 3:612-620. A. P. Dempster, N. M. Laird & D. B. Rubin. (1977) Maximum-likelihood from incomplete data via the EM algorithm. J. Royal Statist. Soc. Ser. B (methodological),39:1-38. S. Geman & D. Geman. (1984) Stochastic relaxation, Gibbs distribution and the Baysian restortion in images. IEEE Trans. Pattern Anal. Machine Intel/" 6,6:721741. R. J. Hathaway. (1986) Another interpretation of the EM algorithm for mixture distributions. Statistics fj Probability Leiters, 4: 53-56. M. 1. Jordan & R. A. Jacob. (1993) Hierarchical mixtures of experts and the EM algorithm. MIT Dept. of Brain and Cognitive Science preprint. R. M. Neal & G. E. Hinton. (1993) A new view of the EM algorithm that justifies incremental and other variants. submitted to Biometrika. S. J. Nowlan. (1990) Maximum likelihood competitve learning. in D. S. Touretzky et al. eds., Advances in Neural Information Systems 2, Morgan Kaufmann. 574-582. K. Rose, E. Gurewitz & G. C. Fox. (1992) Vector quantization by deterministic annealing. IEEE Trans. Information Theory, 38,4:1249-1257. N. Ueda & R. Nakano. (1994) Mixture density estimation via EM algorithm with deterministic annealing. in proc. IEEE Neural Networks for Signal Processing, 69-77. Y. Wong. (1993) Clustering data by melting. Neural Computation, 5:89-104. A. 1. Yuille, P. Stolorz & J. Utans. (1994) Statistical physics, mixtures of distributions, and the EM algorithm. Neural Computation, 6:334-340.
941 |@word version:1 cha:1 jacob:2 simplifying:1 klk:1 initial:5 configuration:1 interestingly:1 current:1 nowlan:3 fn:1 partition:3 plot:2 parameterization:1 successive:1 ryohei:5 become:1 introduce:1 theoretically:1 indeed:2 expected:1 seika:1 mechanic:3 brain:1 ote:1 increasing:3 becomes:3 begin:1 moreover:3 maximizes:4 interpreted:2 minimizes:1 ought:1 poste:1 exactly:2 biometrika:1 ser:1 originates:1 before:1 positive:1 local:8 ease:1 mentioning:1 practical:1 yj:1 practice:2 differs:1 procedure:3 word:2 radial:1 convenience:1 close:3 storage:1 influence:1 wong:2 conventional:1 deterministic:10 missing:1 maximizing:2 starting:3 stolorz:2 independently:1 m2:13 rule:1 regarded:2 his:2 coordinate:1 analogous:1 suppose:2 play:1 programming:1 geman:4 role:1 preprint:1 yk:7 rose:2 mentioned:1 dempster:2 complexity:1 yuille:1 completely:1 basis:1 easily:3 joint:2 dyl:1 llogp:1 derivation:1 effective:2 posed:1 say:1 otherwise:1 statistic:1 laird:2 hoc:1 sequence:1 mb:2 loop:1 yjl:1 competition:3 convergence:1 zp:2 incremental:1 converges:5 derive:2 job:1 eq:15 soc:1 c:1 direction:1 stochastic:2 settle:1 ja:1 generalization:1 normal:1 exp:5 equilibrium:1 visualize:1 substituting:1 adopt:1 estimation:6 proc:1 applicable:1 sensitive:1 largest:1 grouped:1 agrees:1 successfully:3 minimization:2 mit:1 clearly:4 always:1 aim:1 ej:2 conjunction:1 derived:3 methodological:1 likelihood:13 superimposed:1 contrast:1 economy:1 dependent:1 unobservable:2 priori:1 proposes:1 special:1 equal:2 once:1 f3:6 unsupervised:1 serious:1 escape:1 replaced:1 attempt:1 highly:1 mixture:11 dyn:1 naonori:4 lh:1 fox:1 incomplete:4 logarithm:1 re:1 instance:1 earlier:1 soft:3 logp:11 maximization:6 cost:5 uniform:1 hundred:1 varies:1 cho:1 density:6 physic:2 central:1 iip:1 choose:1 ykl:5 dr:1 cognitive:1 expert:2 kuhnel:1 japan:1 mp2:1 satisfy:2 caused:1 explicitly:1 ad:1 depends:3 performed:7 later:2 qfj:1 view:3 tion:1 lab:1 competitive:2 bayes:1 kaufmann:1 ofthe:1 generalize:1 trajectory:1 worth:1 finer:1 submitted:1 eda:1 touretzky:1 ed:1 energy:7 naturally:1 associated:1 stop:1 knowledge:1 emerges:1 actually:1 ok:1 specify:1 melting:1 done:1 until:1 qf3:3 hand:1 grbf:1 effect:2 multiplier:1 byrne:2 hence:3 alternating:1 laboratory:1 iteratively:1 neal:3 attractive:1 coincides:1 criterion:1 generalized:1 arrived:1 complete:5 temperature:7 fj:4 image:1 recently:3 jp:4 interpretation:2 m1:1 onedimensional:1 gibbs:2 encouragement:1 reliability:1 surface:3 posterior:13 apart:1 arbitrarily:1 aem:5 morgan:1 expecta:1 employed:2 converge:1 maximize:1 monotonically:1 signal:1 ii:1 lik:7 thermodynamic:3 kyoto:1 reduces:2 ntt:4 ofp:1 variant:11 expectation:5 iteration:3 addition:1 annealing:19 unlike:4 ascent:1 jordan:2 integer:1 noting:2 identically:1 variety:1 iterate:1 veda:2 expression:1 soraku:1 reformulated:1 hessian:1 cause:1 jj:1 useful:1 generally:1 statist:1 tth:1 utans:2 sign:1 estimated:3 trapped:2 per:1 track:1 ce:1 relaxation:1 parameterized:5 ueda:5 dy:2 fl:4 nonnegative:1 kexp:1 constraint:1 min:2 cslab:2 according:1 em:36 explained:2 gradually:2 taken:1 equation:1 eventually:1 mind:1 adopted:1 rewritten:1 apply:1 hierarchical:2 eigen:3 original:5 denotes:3 clustering:3 nakano:7 especially:1 hikaridai:1 objective:1 added:1 occurs:1 thank:1 simulated:3 outer:1 gun:1 relationship:1 minimizing:2 demonstration:1 equivalently:1 anal:1 lil:4 boltzmann:2 observation:1 hinton:3 communication:1 namely:1 specified:1 optimized:2 baysian:1 distinction:1 trans:3 pattern:1 fp:2 reliable:1 max:2 royal:1 buhmann:2 ff3:2 dyk:4 gurewitz:1 acknowledgement:1 interesting:1 analogy:3 localized:1 rubin:2 kecl:2 principle:4 pi:1 lo:1 repeat:1 free:8 rior:1 side:1 wide:2 taking:2 nge:1 distributed:1 overcome:1 contour:2 observable:2 keep:1 ml:12 global:7 summing:1 assumed:1 surfa:1 search:2 iterative:2 zk:1 da:4 intel:1 lc:1 deterministically:3 ix:1 exists:1 quantization:2 justifies:1 points3:1 entropy:4 saddle:2 lagrange:1 ffj:6 hathaway:2 corresponds:6 conditional:3 goal:1 likl:2 rbf:2 consequently:1 change:1 hard:2 experimentally:1 determined:2 except:1 called:3 total:1 indicating:1 fter:1 competitve:1 dept:1 tested:1
7,180
942
Patterns of damage in neural networks: The effects of lesion area, shape and number Eytan Ruppin and James A. Reggia ? Department of Computer Science A.V. Williams Bldg. University of Maryland College Park, MD 20742 ruppin@cs.umd.edu reggia@cs.umd.edu Abstract Current understanding of the effects of damage on neural networks is rudimentary, even though such understanding could lead to important insights concerning neurological and psychiatric disorders. Motivated by this consideration, we present a simple analytical framework for estimating the functional damage resulting from focal structural lesions to a neural network. The effects of focal lesions of varying area, shape and number on the retrieval capacities of a spatially-organized associative memory. Although our analytical results are based on some approximations, they correspond well with simulation results. This study sheds light on some important features characterizing the clinical manifestations of multi-infarct dementia, including the strong association between the number of infarcts and the prevalence of dementia after stroke, and the 'multiplicative' interaction that has been postulated to occur between Alzheimer's disease and multi-infarct dementia. *Dr. Reggia is also with the Department of Neurology and the Institute of Advanced Computer Studies at the University of Maryland. 36 1 Eytan Ruppin, James A. Reggia Introduction Understanding the response of neural nets to structural/functional damage is important for a variety of reasons, e.g., in assessing the performance of neural network hardware, and in gaining understanding of the mechanisms underlying neurological and psychiatric disorders. Recently, there has been a growing interest in constructing neural models to study how specific pathological neuroanatomical and neurophysiological changes can result in various clinical manifestations, and to investigate the functional organization of the symptoms that result from specific brain pathologies (reviewed in [1, 2]). In the area of associative memory models specifically, early studies found an increase in memory impairment with increasing lesion severity (in accordance with Lashley's classical 'mass action' principle), and showed that slowly developing lesions have less pronounced effects than equivalent acute lesions [3]. Recently, it was shown that the gradual pattern of clinical deterioration manifested in the majority of Alzheimer's patients can be accounted for, and that different synaptic compensation rates can account for the observed variation in the severity and progression rate of this disease [4]. However, this past work is limited in that model elements have no spatial relationships to one another (all elements are conceptually equidistant). Thus, as there is no way to represent focal (localized) damage in such networks, it has not been possible to study the functional effect of focal lesions and to compare them with that caused by diffuse lesions. The limitations of past work led us to use spatially-organized neural network for studying the effects of different types of lesions (we use the term lesion to mean any type of structural and functional damage inflicted on an initially intact neural network). The elements in our model, which can be thought of as representing neurons, or micro-columnar units, form a 2-dimensional array (whose edges are connected, forming a torus to eliminate edge effects), and each unit is connected primarily to its nearby neighbors, as is the case in the cortex [5]. It has recently been shown that such spatially-organized attractor networks can function reasonably well as associative memory devices [6]. This paper presents the first detailed analysis of the effects of lesions of various size, form and number on the memory performance of spatially-organized attractor neural networks. Assuming that these networks are a plausible model of some frontal and associative cortical areas (see, e.g., [7]), our results shed light on the clinical progress of disorders such as stroke and dementia. In the next section, we derive a theoretical framework that characterizes the effects of focal lesions on an associative network's performance. This framework, which is formulated in very general terms, is then examined via simulations in Section 3, which show a remarkable quantitative fit with the theoretical predictions, and are compared with simulations examining performance with diffuse damage. Finally, the clinical significance of our results is discussed in Section 4. 2 Analyzing the effects of focal lesions Our analysis pertains to the case where in the pre-damaged network, all units have an approximately similar average level of activity 1. A focal structural lesion IThis is true in general for associative memory networks, when the activity of each unit is averaged over a sufficiently long time span. Patterns of Damage ill NeuraL Networks 37 (anatomical lesion), denoting the area of damage and neuronal death, is modeled by clamping the activity of the lesioned units to zero. As a result of this primary lesion, the activity of surrounding units may be decreased, resulting in a secondary area of functional lesion, as illustrated in Figure 1. We are primarily interested in large focal lesions , where the area s of the lesion is significantly greater than the local neighborhood region from which each unit receives its inputs. Throughout our analysis we shall hold the working assumption that, traversing from the border of the lesion outwards, the activity of units gradually rises from zero until it reaches its normal, predamaged levels, at some distance d from the lesion's border (see Figure 1). As s is large and d is determined by local interactions on the borders of the structural lesion, we may reasonably assume that the value of d is independent of the lesion size, and depends primarily on the specific network characteristics , such as it architecture, dynamics, and memory load. Figure 1: A sketch of a structural (dark shading) and surrounding functional (light shading) rectangular lesion. Let the intact baseline performance level of the network be denoted as P(O), and let the network size be A. The network 's performance denotes how accurately it retrieves the correct memorized patterns given a set of input cues, and is defined formally below . A structural lesion of area s (dark shading in Figure 1), causing a functional lesion of area ~s (light shading in Figure 1), will then result In a performance level of approximately P(s) = P(O) [A - (s + ~3)l + Pt:..~3 = P(O) _ (~P~3)/(A _ s) , (1) A-s where Pt:.. denotes the average level of performance over ~3 and ~P = P(O) - Pt:.. . P( s) hence reflects the performance level over the remaining viable parts of the network , discarding the structurally damaged region. Bearing these definitions in mind, a simple analysis shows that the effect of focal lesions is governed by the following rules. = 71-r2 . ~3' the area of Consider a symmetric, circular structural lesion of size s functional damage following such a lesion is then (assuming large lesions and hence VS> d) Rule 1: (2) 38 Eytan Ruppin, James A. Reggia 1.0 .--~----r-------r------.--, 0.8 -k.1 --- k=5 k=25 0.2 0.0 '---~----'"---~-~---'--' 0.0 500.0 1000.0 1500.0 Lesion size Figure 2: Theoretically predicted network performance as a function of a single focal structural lesion's size (area): analytic curves obtained for different k values; A = 1600. and ~ P(O) - Ak..JS , (3) -s for some constant k = yi4;db..P. Thus, the area of functional damage surrounding a single focal structural lesion is proportional to the square root of the structural lesion's area. Some analytic performancejlesioning curves (for various k values) are illustrated in Figure 2. Note the different qualitative shape of these curves as a function of k. Letting x = sjA be the fraction of structural damage, we have pes) P(x) ~ P(O) _ k.JX _1_ 1-x.,jA , (4) that is, the same fraction x of damage results in less performance decrease in larger networks. This surprising result testifies to the possible protective value of having functional 'modular' cortical networks of large size. Expressions 3 and 4 are valid also when the structural lesion has a square shape. To study the effect of the structural lesion's shape, we consider the area b.. 8 [n] of a functional lesion resulting from a rectangular focal lesion of size s = a . b (see Figure 1), where, without loss of generality, n = ajb ~ 1. Then, for large n, we find that the functional damage of a rectangular structural lesion of fixed size increases as its shape is more elongated, following Rule 2: (5) and pes) ~ P(O) _ k.,foS 2(A - s) (6) Patterns of Damage in Neural Networks 39 To study the effect of the number of lesions, consider the area .6. 3 m of a functional lesion composed of m focal rectangular structural lesions (with sides a = n? b), each of area s/m. We find that the functional damage increases with the number offocal sub-lesions (while total structural lesion area is held constant), according to Rule 3: (7) and ~ P(O) _ pes) k.;mns (8) 2(A - s) While Rule 3 presents a lower bound on the functional damage which may actually be significantly larger and involves no approximations, Rule 2 presents an upper bound on the actual functional damage. As we shall show in the next section, the number of lesions actually affects the network performance significantly more than its precise shape. 3 Numerical Simulation Results We now turn to examine the effect of lesions on the performance of an associative memory network via simulations. The goal of these simulations is twofold. First, to examine how accurately the general but approximate theoretical results presented above describe the actual performance degradation in a specific associative network. Second, to compare the effects of focal lesions to those of diffuse ones, as the effect of diffuse damage cannot be described as a limiting case within the framework of our analysis. Our simulations were performed using a standard Tsodyks-Feigelman attractor neural network [8]. This is a Hopfield-like network which has several features which make it more biologically plausible [4], such as low activity and non-zero positive thresholds. In all the experiments, 20 sparse random {O, I} memory patterns (with a fraction of p ~ 1 of l's) were stored in a network of N = 1600 units, placed on a 2-dimensionallattice. The network has spatially organized connectivity, where each unit has 60 incoming connections determined randomly with a Gaussian probability <J>(z) = J1/27rexp( _z2 /2(1'2), where z is the distance between two units in the array. When (1' is small, each unit is connected primarily to its nearby neighbors. As in [4], the cue input patterns are presented via an external field of magnitude e = 0.035, and the noise level is T = 0.005. The performance of the network is measured (over the viable, non-lesioned units) by the standard overlap measure which denotes the similarity between the final state S the network converges to and the memory pattern ~!l that was cued in that trial, defined by m!l(t) = (1 P ~ N )N P I:(~r - p)Si(t) . (9) i=l In all simulations we report the average overlap achieved over 100 trials. We first studied the network's performance at various (1' values. Figure 3a displays how the performance of the network degrades when diffuse structural lesions of increasing size are inflicted upon it (i.e., randomly selected units are clamped to zero), while Figure 3b plots the performance as a function of the size of a single square-shaped focal lesion. As is evident, spatially-organized connectivity enables 40 Eytan Ruppin, James A. Reggia the network to maintain its memory retrieval capacities in face of focal lesions of considerable size. Diffuse lesions are always more detrimental than single focal lesions of identical size. Also plotted in Figure 3b is the analytical curve calculated via expression (3) (with k = 5), which shows a nice fit with the actual performance of the spatially-connected network parametrized by (J' = 1. Concentrating on the study of focal lesions in a spatially-connected network, we adhere to the values (J' = 1 and k = 5 hereafter, and compare the analytical and numerical results. With these values, the analytical curves describing the performance of the network as a function of the fraction of the network lesioned (obtained using expression 4) are similar to the corresponding numerical results. (a) (b) 1.0 r-----,----~-------, 1.0 , -- - - - , - - - - - r - -- - - - , .... ~ .... 0.8 .... ~...: ~.:..-~, - -- ",gna_1 signa =3 "'gna= 10 - - - "'gnam30 ',~''', ~-...~. ; - .. -\ 0.8 0 .6 0 .6 , \ \ \', \ \ \ \ \ \ \ ".-.\ '. \ \ . \ \ , '. \ 0 .4 0.4 0.2 0.2 \ \ \ - - - sigma =1 \ . . . , sigma _3 \ sigma -10 \ - - - sigma =30 " Analytical .osuns, k = 5 " '. \ , \ '.\ , ....................... '0.0 L-_~ 0 .0 _ _'___~_~ 500.0 _ _ _---' 1000.0 Lesion sIZe 1500.0 0.0 ':--~__;:::'::_:_-~-::-:':c_:__---= 0 .0 500.0 1000.0 1500.0 Lesion size Figure 3: Network performance as a function of lesion size: simulation results obtained in four different networks , each characterized by a distinct distribution of spatially-organized connectivity. (a) Diffuse lesions. (b) Focal lesions. To examine Rule 2, a rectangular structural lesion of area s = 300 was induced in the network. As shown in Figure 4a, as the ratio n between the sides is increased, the network's performance further decreases, but this effect is relatively mild. The markedly stronger effect of varying the lesion number (described by Rule 3) is demonstrated in figure 4b, which shows the effect of multiple lesions composed of 2,4,8 and 16 separate focal lesions. For comparison, the performance achieved with a diffuse lesion of similar size is plotted on the 20'th x-ordinate. It is interesting to note that a sufficiently large multiple focal lesion (s = 512) can cause a larger performance decrease than a diffuse lesion of similar size. That is, at some point, when the size of each individual focal lesion becomes small in relation to the spread of each unit's connectivity, our analysis looses its validity, and Rule 3 ceases to hold. Patterns of Damage in Neural Networks (a) 41 (b) 0.90 ,---'-~-"-----'----r-----, <>--<> SlrnuIeIion "'1U~s - . - Analytical reou~ 0.930 0.910 lii <>--tI Sinulation (s D 256) ...... " Sinulation (s ? 512) - - - - Analytic (s =256) - . - Analytic (s ? 512) 0.80 " ~ 0.890 " , " " 0.80 0.870 0.850 '---~-'---'----:"-~~-~-----' 0.0 1.0 2.0 3.0 4.0 5.0 6 .0 Rectangular ratio n 0.50 '----'----"------'-~--'-----' 0 .0 5.0 10.0 15.0 20.0 25.0 No. 01 sub-lesions m Figure 4: Network performance as a function of focal lesion shape (a) and number (b). Both numerical and analytical results are displayed . In Figure 4b, the x-ordinate denotes the number of separate sub-lesions (1,2,4,8,16) , and, for comparison, the performance achieved with a diffuse lesion of similar size is plotted on the 20'th x-ordinate. 4 Discussion We have presented a simple analytical framework for studying the effects of focal lesions on the functioning of spatially organized neural networks . The analysis presented is quite general and a similar approach could be adopted to investigate the effect offocallesions in other neural models. Using this analysis, specific scaling rules have been formulated describing the functional effects of structural focal lesions on memory retrieval performance in associative attractor networks. The functional lesion scales as the square root of the size of a single structural lesion, and the form of the resulting performance curve depends on the impairment span d. Surprisingly, the same fraction of damage results in significantly less performance decrease in larger networks, pointing to their relative robustness. As to the effects of shape and number, elongated structural lesions cause more damage than more symmetrical ones. However, the number of sub-lesions is the most critical factor determining the functional damage and performance decrease in the model. Numerical studies show that in some conditions multiple lesions can damage performance more than diffuse damage, even though the amount of lost innervation is always less in a multiple focal lesion than with diffuse damage. Beyond its computational interest, the study of the effects of focal damage on the performance of neural network models can lead to a better understanding of functional impairments accompanying focal brain lesions. In particular, we are interested in multi-infarct dementia, a frequent cause of dementia (chronic deterioration of cognitive and memory capacities) characterized by a series of multiple, aggregat- 42 Eytan Ruppin, James A. Reggia ing focal lesions. Our results indicate a significant role for the number of infarcts in determining the extent of functional damage and dementia in multi-infarct disease. In our model, multiple focal lesions cause a much larger deficit than their simple 'sum', i.e., a single lesion of equivalent total size. This is consistent with clinical studies that have suggested the main factors related to the prevalence of dementia after stroke to be the infarct number and site, and not the overall infarct size, which is related to the prevalence of dementia in a significantly weaker manner [9, 10]. Our model also offers a possible explanation to the 'multiplicative' interaction that has been postulated to occur between co-existing Alzheimer and multi-infarct dementia [10], and to the role of cortical atrophy in increasing the prevalence of dementia after stroke; in accordance with our model, it is hypothesized that atrophic degenerative changes will lead to an increase in the value of d (and hence of k) and increase the functional damage caused by a lesion of given structural size. This hypothesis, together with a detailed study of the effects of the various network parameters on the value of d, are currently under further investigation. Acknowledgements This research has been supported by a Rothschild Fellowship to Dr. Ruppin and by Awards NS29414 and NS16332 from NINDS. References [1] J. Reggia, R. Berndt, and L. D'Autrechy. Connectionist models in neuropsychology. In Handbook of Neuropsychology, volume 9. 1994, in press. [2] E. Ruppin. Neural modeling of psychiatric disorders. Network: Computation in Neural Systems, 1995. Invited review paper, to appear. [3] J .A. Anderson. Cognitive and psychological computation with neural models. IEEE Trans. on Systems,Man, and Cybernetics, SMC-13(5):799-815, 1983. [4] D. Horn, E. Ruppin, M. Usher, and M. Herrmann. Neural network modeling of memory deterioration in alzheimer's disease. Neural Computation, 5:736-749, 1993. [5] A. M. Thomson and J. Deuchars. Temporal and spatial properties of local circuits in the neocortex. Trends in neuroscience, 17(3):119-126, 1994. [6] J .M. Karlholm. Associative memories with short-range higher order couplings. Neural Networks, 6:409-421, 1993. [7] F.A.W. Wilson, S.P.O Scalaidhe, and P.S. Goldman-Rakic. Dissociation of object and spatial processing domains in primate prefrontal cortex. Science, 260:1955-1958,1993. [8] M.V. Tsodyks and M.V. Feigel'man. The enhanced storage capacity in neural networks with low activity level. Europhys. Lett., 6:101 - 105, 1988. [9] T.K. Tatemichi, M.A. Foulkes, J.P. Mohr, J.R. Hewitt, D. B. Hier, T .R. Price, and P.A. Wolf. Dementia in stroke survivors in the stroke data bank cohort. Stroke, 21:858-866, 1990. [10] T. K. Tatemichi. How acute brain failure becomes chronic: a view of the mechanisms of dementia related to stroke. Neurology, 40:1652-1659, 1990.
942 |@word mild:1 trial:2 stronger:1 gradual:1 simulation:9 shading:4 series:1 hereafter:1 denoting:1 past:2 existing:1 current:1 z2:1 surprising:1 si:1 numerical:5 j1:1 shape:9 analytic:4 enables:1 plot:1 v:1 cue:2 selected:1 device:1 short:1 berndt:1 viable:2 qualitative:1 manner:1 theoretically:1 examine:3 growing:1 multi:5 brain:3 goldman:1 actual:3 increasing:3 becomes:2 innervation:1 estimating:1 underlying:1 circuit:1 mass:1 feigel:1 loos:1 temporal:1 quantitative:1 ti:1 shed:2 unit:15 appear:1 positive:1 accordance:2 local:3 ak:1 analyzing:1 mohr:1 approximately:2 studied:1 examined:1 co:1 limited:1 smc:1 range:1 averaged:1 horn:1 lost:1 prevalence:4 area:18 thought:1 significantly:5 pre:1 psychiatric:3 cannot:1 storage:1 equivalent:2 elongated:2 demonstrated:1 chronic:2 williams:1 rectangular:6 disorder:4 insight:1 rule:10 array:2 inflicted:2 variation:1 testifies:1 limiting:1 pt:3 enhanced:1 damaged:2 hypothesis:1 element:3 trend:1 observed:1 role:2 tsodyks:2 region:2 connected:5 feigelman:1 decrease:5 neuropsychology:2 disease:4 lesioned:3 dynamic:1 upon:1 hopfield:1 various:5 retrieves:1 surrounding:3 distinct:1 describe:1 neighborhood:1 europhys:1 whose:1 modular:1 larger:5 plausible:2 quite:1 final:1 associative:10 analytical:9 net:1 interaction:3 frequent:1 causing:1 pronounced:1 assessing:1 converges:1 object:1 cued:1 derive:1 coupling:1 measured:1 progress:1 strong:1 c:2 predicted:1 involves:1 indicate:1 correct:1 memorized:1 ja:1 investigation:1 c_:1 hold:2 rexp:1 sufficiently:2 accompanying:1 normal:1 pointing:1 early:1 jx:1 currently:1 reflects:1 gaussian:1 always:2 varying:2 wilson:1 survivor:1 degenerative:1 rothschild:1 baseline:1 ajb:1 eliminate:1 initially:1 relation:1 interested:2 overall:1 ill:1 denoted:1 spatial:3 field:1 having:1 shaped:1 identical:1 park:1 report:1 connectionist:1 micro:1 primarily:4 pathological:1 randomly:2 composed:2 individual:1 attractor:4 maintain:1 organization:1 interest:2 investigate:2 circular:1 light:4 held:1 edge:2 traversing:1 plotted:3 theoretical:3 psychological:1 increased:1 modeling:2 examining:1 stored:1 together:1 connectivity:4 slowly:1 prefrontal:1 dr:2 external:1 lii:1 cognitive:2 account:1 protective:1 postulated:2 caused:2 depends:2 multiplicative:2 root:2 performed:1 view:1 characterizes:1 square:4 characteristic:1 dissociation:1 correspond:1 conceptually:1 accurately:2 cybernetics:1 stroke:8 reach:1 fo:1 synaptic:1 definition:1 failure:1 james:5 concentrating:1 organized:8 actually:2 higher:1 response:1 though:2 symptom:1 generality:1 anderson:1 until:1 working:1 receives:1 sketch:1 hier:1 effect:25 validity:1 hypothesized:1 true:1 functioning:1 hence:3 spatially:10 symmetric:1 death:1 illustrated:2 rakic:1 manifestation:2 evident:1 thomson:1 rudimentary:1 ruppin:9 consideration:1 recently:3 functional:23 volume:1 association:1 discussed:1 significant:1 focal:30 pathology:1 acute:2 cortex:2 similarity:1 j:1 showed:1 manifested:1 greater:1 multiple:6 ing:1 characterized:2 clinical:6 long:1 retrieval:3 offer:1 concerning:1 award:1 prediction:1 patient:1 represent:1 deterioration:3 achieved:3 fellowship:1 decreased:1 adhere:1 invited:1 umd:2 usher:1 markedly:1 induced:1 db:1 alzheimer:4 structural:23 cohort:1 variety:1 affect:1 fit:2 equidistant:1 architecture:1 motivated:1 expression:3 cause:4 action:1 impairment:3 detailed:2 amount:1 dark:2 neocortex:1 hardware:1 neuroscience:1 anatomical:1 shall:2 four:1 threshold:1 fraction:5 sum:1 bldg:1 throughout:1 ninds:1 scaling:1 bound:2 display:1 activity:7 occur:2 diffuse:12 nearby:2 span:2 relatively:1 department:2 developing:1 according:1 biologically:1 primate:1 gradually:1 turn:1 describing:2 mechanism:2 mind:1 letting:1 studying:2 adopted:1 progression:1 reggia:8 robustness:1 neuroanatomical:1 denotes:4 remaining:1 atrophy:1 classical:1 damage:29 primary:1 degrades:1 md:1 detrimental:1 distance:2 separate:2 maryland:2 deficit:1 capacity:4 majority:1 parametrized:1 extent:1 reason:1 gna:1 assuming:2 modeled:1 relationship:1 ratio:2 lashley:1 sigma:4 rise:1 upper:1 neuron:1 compensation:1 displayed:1 severity:2 precise:1 ordinate:3 connection:1 trans:1 beyond:1 suggested:1 below:1 pattern:9 including:1 memory:15 gaining:1 explanation:1 overlap:2 critical:1 advanced:1 mn:1 representing:1 nice:1 understanding:5 acknowledgement:1 review:1 determining:2 relative:1 loss:1 interesting:1 limitation:1 proportional:1 localized:1 remarkable:1 consistent:1 signa:1 principle:1 bank:1 accounted:1 placed:1 surprisingly:1 supported:1 side:2 weaker:1 institute:1 neighbor:2 characterizing:1 face:1 sparse:1 curve:6 calculated:1 cortical:3 valid:1 lett:1 herrmann:1 yi4:1 approximate:1 incoming:1 handbook:1 symmetrical:1 neurology:2 reviewed:1 reasonably:2 bearing:1 constructing:1 domain:1 foulkes:1 significance:1 spread:1 main:1 hewitt:1 border:3 noise:1 lesion:83 neuronal:1 site:1 structurally:1 sub:4 torus:1 governed:1 pe:3 clamped:1 atrophic:1 load:1 specific:5 discarding:1 dementia:13 r2:1 cease:1 magnitude:1 clamping:1 outwards:1 columnar:1 led:1 forming:1 neurophysiological:1 neurological:2 wolf:1 goal:1 formulated:2 twofold:1 price:1 man:2 considerable:1 change:2 specifically:1 determined:2 ithis:1 degradation:1 total:2 secondary:1 eytan:5 intact:2 formally:1 college:1 pertains:1 frontal:1
7,181
943
Phase-Space Learning Fu-Sheng Tsung Chung Tai Ch'an Temple 56, Yuon-fon Road, Yi-hsin Li, Pu-li Nan-tou County, Taiwan 545 Republic of China Garrison W. Cottrell? Institute for Neural Computation Computer Science & Engineering University of California, San Diego La Jolla, California 92093 Abstract Existing recurrent net learning algorithms are inadequate. We introduce the conceptual framework of viewing recurrent training as matching vector fields of dynamical systems in phase space. Phasespace reconstruction techniques make the hidden states explicit, reducing temporal learning to a feed-forward problem. In short, we propose viewing iterated prediction [LF88] as the best way of training recurrent networks on deterministic signals. Using this framework, we can train multiple trajectories, insure their stability, and design arbitrary dynamical systems. 1 INTRODUCTION Existing general-purpose recurrent algorithms are capable of rich dynamical behavior. Unfortunately, straightforward applications of these algorithms to training fully-recurrent networks on complex temporal tasks have had much less success than their feedforward counterparts. For example, to train a recurrent network to oscillate like a sine wave (the "hydrogen atom" of recurrent learning), existing techniques such as Real Time Recurrent Learning (RTRL) [WZ89] perform suboptimally. Williams & Zipser trained a two-unit network with RTRL, with one teacher signal. One unit of the resulting network showed a distorted waveform, the other only half the desired amplitude. [Pea89] needed four hidden units. However, our work demonstrates that a two-unit recurrent network with no hidden units can learn the sine wave very well [Tsu94]. Existing methods also have several other ?Correspondence should be addressed to the second author: gary@cs.ucsd.edu 482 Fu-Sheng Tsung, Garrison W. Cottrell limitations. For example, networks often fail to converge even though a solution is known to exist; teacher forcing is usually necessary to learn periodic signals; it is not clear how to train multiple trajectories at once, or how to insure that the trained trajectory is stable (an attractor). In this paper, we briefly analyze the algorithms to discover why they have such difficulties, and propose a general solution to the problem. Our solution is based on the simple idea of using the techniques of time series prediction as a methodology for recurrent network training. First, by way of introducing the appropriate concepts, consider a system of coupled autonomous l first order network equations: FI (Xl (t), X2(t), . .. , Xn (t)) F2(XI(t), X2(t),? ? ?, Xn(t)) or, in vector notation, X(t) = F(X) where XCt) = (XICt), X2(t),?? ., xn(t)) The phase space is the space of the dependent variables (X), it does not include t, while the state space incorporates t. The evolution of a trajectory X(t) traces out a phase curve, or orbit, in the n-dimensional phase space of X . For low dimensional systems (2- or 3-D), it is easy to visualize the limit sets in the phase space: a fixed point and a limit cycle become a single point and a closed orbit (closed curve), respectively. In the state space they become an infinite straight line and a spiral. F(X) defines the vector field of X, because it associates a vector with each point in the phase space of X whose direction and magnitude determines the movement of that point in the next instant of time (by definition, the tangent vector). 2 ANALYSIS OF CURRENT APPROACHES To get a better understanding of why recurrent algorithms have not been very effective, we look at what happens during training with two popular recurrent learning techniques: RTRL and back propagation through time (BPTT). With each, we illustrate a different problem, although the problems apply equally to each technique. RTRL is a forward-gradient algorithm that keeps a matrix of partial derivatives of the network activation values with respect to every weight. To train a periodic trajectory, it is necessary to teacher-force the visible units [WZ89], i.e., on every iteration, after the gradient has been calculated, the activations of the visible units are replaced by the teacher. To see why, consider learning a pair of sine waves offset by 90?. In phase space, this becomes a circle (Figure la). Initially the network 1 Autonomous means the right hand side of a differential equation does not explicitly reference t, e.g. dx/dt = 2x is autonomous, even though x is a function oft, but dx/dt = 2x+t is not. Continuous neural networks without inputs are autonomous. A nonautonomous system can always be turned into an autonomous system in a higher dimension. Phase-Space Learning 483 a b Figure 1: Learning a pair of sine waves with RTRL learning. (a) without teacher forcing, the network dynamics (solid arrows) take it far from where the teacher (dotted arrows) assumes it is, so the gradient is incorrect. (b) With teacher forcing, the network's visible units are returned to the trajectory. (thick arrows) is at position Xo and has arbitrary dynamics. After a few iterations, it wanders far away from where the teacher (dashed arrows) assumes it to be. The teacher then provides an incorrect next target from the network's current position. Teacher-forcing (Figure 1b), resets the network back on the circle, where the teacher again provides useful information. However, if the network has hidden units, then the phase space of the visible units is just a projection of the actual phase space of the network, and the teaching signal gives no information as to where the hidden units should be in this higherdimensional phase space. Hence the hidden unit states, unaltered by teacher forcing , may be entirely unrelated to what they should be. This leads to the moving targets problem. During training, every time the visible units re-visit a point, the hidden unit activations will differ, Thus the mapping changes during learning. (See [Pin88, WZ89] for other discussions of teacher forcing.) With BPTT, the network is unrolled in time (Figure 2). This unrolling reveals another problem: Suppose in the teaching signal, the visible units' next state is a non-linearly separable function of their current state. Then hidden units are needed between the visible unit layers, but there are no intermediate hidden units in the unrolled network. The network must thus take two time steps to get to the hidden units and back. One can deal with this by giving the teaching signal every other iteration, but clearly, this is not optimal (consider that the hidden units must "bide their time" on the alternate steps).2 The trajectories trained by RTRL and BPTT tend to be stable in simulations of simple tasks [Pea89, TCS90], but this stability is paradoxical. Using teacher forcing, the networks are trained to go from a point on the trajectory, to a point within the ball defined by the error criterion f (see Figure 4 (a)). However, after learning, the networks behave such that from a place near the trajectory, they head for the trajectory (Figure 4 (b)). Hence the paradox. Possible reasons are: 1) the hidden unit moving targets provide training off the desired trajectory, so that if the training is successful, the desired trajectory is stable; 2) we would never consider the training successful if the network "learns" an unstable trajectory; 3) the stable dynamics in typical situations have simpler equations than the unstable dynamics [N ak93]. To create an unstable periodic trajectory would imply the existence of stable regions both inside and outside the unstable trajectory ; dynamically this is 2 At NIPS , 0 delay connections to the hidden units were suggested, which is essentially part of the solution given here. 484 Fu-Sheng Tsung, Garrison W. Cottrell ~------------. Figure 2: A nonlinearly separable mapping must be computed by the hidden units (the leftmost unit here) every other time step. Figure 3: The network used for iterated prediction training. Dashed connections are used after learning. " a b Figure 4: The paradox of attractor learning with teacher forcing. (a) During learning, the network learns to move from the trajectory to a point near the trajectory. (b) After learning, the network moves from nearby points towards the trajectory. more complicated than a simple periodic attractor. In dynamically complex tasks, a stable trajectory may no longer be the simplest solution, and stability could be a problem. In summary, we have pointed out several problems in the RTRL (forward-gradient) and BPTT (backward-gradient) classes of training algorithms: 1. Teacher forcing with hidden units is at best an approximation, and leads to the moving targets problem. 2. Hidden units are not placed properly for some tasks. 3. Stability is paradoxical. 3 PHASE-SPACE LEARNING The inspiration for our approach is prediction training [LF88], which at first appears similar to BPTT, but is subtly different. In the standard scheme, a feedforward network is given a time window, a set of previous points on the trajectory to be learned, as inputs. The output is the next point on the trajectory. Then, the inputs are shifted left and the network is trained on the next point (see Figure 3). Once the network has learned, it can be treated as recurrent by iterating on its own predictions. The prediction network differs from BPTT in two important ways. First, the visible units encode a selected temporal history of the trajectory (the time window) . The point of this delay space embedding is to reconstruct the phase space of the underlying system. [Tak81] has shown that this can always be done for a deterministic system. Note that in the reconstructed phase space, the mapping from one Phase-Space Learning 485 YI+I -------,--........r;.-.-.-. ...... ... ... .'"'"'"'" , ??? a b Figure 5: Phase-space learning. (a) The training set is a sample of the vector field. (b) Phase-space learning network. Dashed connections are used after learning. point to the next (based on the vector field) is deterministic. Hence what originally appeared to be a recurrent network problem can be converted into an entirely feed forward problem. Essentially, the delay-space reconstruction makes hidden states visible, and recurrent hidden units unnecessary. Crucially, dynamicists have developed excellent reconstruction algorithms that not only automate the choices of delay and embedding dimension but also filter out noise or get a good reconstruction despite noise [FS91, Dav92, KBA92]. On the other hand, we clearly cannot deal with non-deterministic systems by this method. The second difference from BPTT is that the hidden units are between the visible units, allowing the network to produce nonlinearly separable transformations of the visible units in a single iteration. In the recurrent network produced by iterated prediction, the sandwiched hidden units can be considered "fast" units with delays on the input/output links summing to 1. Since we are now lear~ing a mapping in phase space, stability is easily ensured by adding additional training examples that converge towards the desired orbit. 3 We can also explicitly control convergence speed by the size and direction of the vectors. Thus, phase-space learning (Figure 5) consists of: (1) embedding the temporal signal to recover its phase space structure, (2) generating local approximations of the vector field near the desired trajectory, and (3) functional approximation of the vector field with a feedforward network. Existing methods developed for these three problems can be directly and independently applied to solve the problem. Since feedforward networks are universal approximators [HSW89], we are assured that at least locally, the trajectory can be represented. The trajectory is recovered from the iterated output of the pre-embedded portion of the visible units. Additionally, we may also extend the phase-space learning framework to also include inputs that affect the output of the system (see [Tsu94] for details).4 In this framework, training multiple attractors is just training orbits in different parts of the phase space, so they simply add more patterns to the training set. In fact, we can now create designer dynamical systems possessing the properties we want, e.g., with combinations of fixed point, periodic, or chaotic attractors. 3The common practice of adding noise to the input in prediction training is just a simple minded way of adding convergence information. 4Principe & Kuo(this volume) show that for chaotic attractors, it is better to treat this as a recurrent net and train using the predictions. 486 Fu-Sheng Tsung, Garrison W. Cottrell 0.5 -0.5 Q -0.5 0 0.5 Figure 6: Learning the van der Pol oscillator. ( a) the training set. (b) Phase space plot of network (solid curve) and teacher (dots). (c) State space plot. As an example, to store any number of arbitrary periodic attractors Zi(t) with periods 11 in a single recurrent network, create two new coordinates for each Zi(t), (Xi(t),Yi(t)) = (sin(*t),cos(*t)), where (Xi,Yi) and (Xj,Yj) are disjoint circles for i 'I j. Then (Xi, Yi, Zi) is a valid embedding of all the periodic attractors in phase space, and the network can be trained. In essence, the first two dimensions form "clocks" for their associated trajectories. 4 SIMULATION RESULTS In this section we illustrate the method by learning the van der Pol oscillator (a much more difficult problem than learning sine waves), learning four separate periodic attractors, and learning an attractor inside the basin of another attractor. 4.1 LEARNING THE VAN DER POL OSCILLATOR The van der Pol equation is defined by: We used the values 0.7, 1, 1 for the parameters a, b, and w, for which there is a global periodic attractor (Figure 6). We used a step size of 0.1, which discretizes the trajectory into 70 points. The network therefore has two visible units. We used two hidden layers with 20 units each, so that the unrolled, feedforward network has a 2-20-20-2 architecture. We generated 1500 training pairs using the vector field near the attractor. The learning rate was 0.01, scaled by the fan-in, momentum was 0.75, we trained for 15000 epochs. The order of the training pairs was randomized. The attractor learned by the network is shown in (Figure 6 (b)). Comparison of the temporal trajectories is shown in Figure 6 (c); there is a slight frequency difference. The average MSE is 0.000136. Results from a network with two layers of 5 hidden units each with 500 data pairs were similar (MSE=0.00034). 4.2 LEARNING MULTIPLE PERIODIC ATTRACTORS [Hop82] showed how to store multiple fixed-point at tractors in a recurrent net. [Bai91] can store periodic and chaotic at tractors by inverting the normal forms of these attractors into higher order recurrent networks. However, traditional recurrent training offers no obvious method of training multiple attractors. [DY89] were able Phase.Space Learning 487 _.0.1. Il.6, 0.63. 0.7 'I'---:.ru"='"""--'O~--:O-=-"---! .I'--::.ru-;---;;--;;-:Oj---! 8 A .ru 0 OJ '1'--::4U~-::-0--;Oj~-! F E 1.--------, 100 1OO D 300 400 ~~ .1 0 50 100 150 1OO ~ 300 H Figure 7: Learning mUltiple attractors. In each case, a 2-20-20-2 network using conjugate gradient is used. Learning 4 attractors: (A) Training set. (B) Eight trajectories of the trained network. (C) Induced vector field of the network. There are five unstable fixed points. (D) State space behavior as the network is "bumped" between attractors. Learning 2 attractors, one inside the other: (E) Training set. (F) Four trajectories ofthe trained network. (G) Induced vector field of the network. There is an unstable limit cycle between the two stable ones. (H) State space behavior with a "bump". to store two limit cycles by starting with fixed points stored in a Hopfield net, and training each fixed point locally to become a periodic attractor. Our approach has no difficulty with multiple attractors. Figure 7 (A-D) shows the result of training four coexisting periodic attractors, one in each quadrant of the two-dimensional phase space. The network will remain in one of the attractor basins until an external force pushes it into another attractor basin. Figure 7 (E-H) shows a network with two periodic attractors, this time one inside the other. This vector field possess an unstable limit cycle between the two stable limit cycles. This is a more difficult task, requiring 40 hidden units, whereas 20 suffice for the previous task (not shown). 5 SUMMARY We have presented a phase space view of the learning process in recurrent nets. This perspective has helped us to understand and overcome some of the problems of using traditional recurrent methods for learning periodic and chaotic attractors. Our method can learn multiple trajectories, explicitly insure their stability, and avoid overfitting; in short, we provide a practical approach to learning complicated temporal behaviors. The phase-space framework essentially breaks the problem into three sub-problems: (1) Embedding a temporal signal to recover its phase space structure, (2) generating local approximations of the vector field near the desired trajectory, and (3) functional approximation in feedforward networks. We have demonstrated that using this method, networks can learn complex oscillations and multiple periodic attractors. 488 Fu-Sheng Tsung, Garrison W. Cottrell Acknowledgements This work was supported by NIH grant R01 MH46899-01A3. Thanks for comments from Steve Biafore, Kenji Doya, Peter Rowat, Bill Hart, and especially Dave DeMers for his timely assistance with simulations. References W. Baird and F. Eeckman. Cam storage of analog patterns and continuous sequences with 3n 2 weights. In R.P. Lippmann, J .E. Moody, and D.S. Touretzky, editors, Advances in Neural Information Processing Systems, volume 3, pages 91-97, 1991. Morgan Kaufmann, San Mateo. [Dav92] M. Davies. Noise reduction by gradient descent. International Journal of Bifurcation and Chaos, 3:113-118, 1992. K. Doya and S. Yoshizawa. Memorizing oscillatory patterns in the analog [DY89] neuron network. In IJCNN, Washington D.C., 1989. IEEE. J.D. Farmer and J.J. Sidorowich. Optimal shadowing and noise reduction. [FS91] Physica D, 47:373-392, 1991. [Hop82] J.J. Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proceedings of the National Academy of Sciences, USA, 79, 1982. [HSW89] K. Hornik, M. Stinchcombe, and H. White. Multilayer feedforward networks are universal approximators. Neural Networks, 2:359-366, 1989. [KBA92] M.B. Kennel, R. Brown, and H. Abarbanel. Determining embedding dimension for phase-space reconstruction using a geometrical construction. Physical Review A, 45:3403-3411, 1992. [LF88] A. Lapedes and R. Farber. How neural nets work . In D.Z. Anderson, editor, Neural Information Processing Systems, pages 442-456, Denver 1987, 1988. American Institute of Physics, New York. [N ak93] Hiroyuki Nakajima. A paradox in learning trajectories in neural networks. Working paper, Dept. of EE II, Kyoto U., Kyoto, JAPAN, 1993. [Pea89] B.A. Pearlmutter. Learning state space trajectories in recurrent neural networks . Neural Computation, 1:263-269, 1989. [Pin88] F.J . Pineda. Dynamics and architecture for neural computation. Journal of Complexity, 4:216-245, 1988. [Tak81] F. Takens. Detecting strange attractors in turbulence. In D.A. Rand and L.-S. Young, editors, Dynamical Systems and Turbulence, volume 898 of Lecture Notes in Mathematics, pages 366-381, Warwick 1980, 1981. Springer-Verlag, Berlin. [TCS90] F-S. Tsung, G . W. Cottrell, and A. I. Selverston. Some experiments on learning stable network oscillations. In IJCNN, San Diego, 1990. IEEE. [Tsu94] F-S. Tsung. Modelling Dynamical Systems with Recurrent Neural Networks. PhD thesis, University of California, San Diego, 1994. [WZ89] R.J. Williams and D. Zipser. A learning algorithm for continually running fully recurrent neural networks . Neural Computation, 1:270-280, 1989. [Bai91]
943 |@word unaltered:1 briefly:1 bptt:7 simulation:3 crucially:1 solid:2 reduction:2 series:1 lapedes:1 existing:5 current:3 recovered:1 activation:3 dx:2 must:3 cottrell:6 visible:13 plot:2 half:1 selected:1 short:2 provides:2 detecting:1 simpler:1 five:1 become:3 differential:1 incorrect:2 consists:1 inside:4 introduce:1 behavior:4 wz89:4 actual:1 window:2 unrolling:1 becomes:1 discover:1 insure:3 notation:1 unrelated:1 underlying:1 suffice:1 what:3 developed:2 selverston:1 transformation:1 temporal:7 every:5 ensured:1 demonstrates:1 scaled:1 control:1 unit:38 grant:1 farmer:1 continually:1 engineering:1 local:2 treat:1 limit:6 despite:1 china:1 mateo:1 dynamically:2 co:1 practical:1 yj:1 practice:1 differs:1 chaotic:4 universal:2 matching:1 projection:1 pre:1 road:1 quadrant:1 davy:1 get:3 cannot:1 turbulence:2 storage:1 bill:1 deterministic:4 demonstrated:1 sidorowich:1 straightforward:1 williams:2 go:1 independently:1 starting:1 his:1 stability:6 embedding:6 autonomous:5 coordinate:1 diego:3 target:4 suppose:1 construction:1 associate:1 region:1 cycle:5 movement:1 complexity:1 pol:4 cam:1 dynamic:5 trained:9 subtly:1 f2:1 easily:1 hopfield:2 emergent:1 represented:1 train:5 fast:1 effective:1 outside:1 whose:1 solve:1 warwick:1 reconstruct:1 ability:1 pineda:1 sequence:1 net:6 reconstruction:5 propose:2 reset:1 turned:1 hop82:2 academy:1 convergence:2 produce:1 generating:2 illustrate:2 recurrent:26 oo:2 c:1 kenji:1 differ:1 direction:2 waveform:1 thick:1 bumped:1 farber:1 filter:1 viewing:2 county:1 physica:1 considered:1 normal:1 mapping:4 visualize:1 bump:1 automate:1 purpose:1 shadowing:1 kennel:1 create:3 minded:1 clearly:2 always:2 avoid:1 encode:1 properly:1 modelling:1 dependent:1 initially:1 hidden:23 takens:1 bifurcation:1 field:11 once:2 never:1 washington:1 atom:1 look:1 few:1 national:1 replaced:1 phase:31 attractor:30 coexisting:1 fu:5 capable:1 partial:1 necessary:2 desired:6 orbit:4 circle:3 re:1 temple:1 introducing:1 republic:1 delay:5 successful:2 inadequate:1 stored:1 teacher:17 periodic:16 thanks:1 international:1 randomized:1 off:1 physic:1 moody:1 again:1 thesis:1 xict:1 external:1 american:1 chung:1 derivative:1 abarbanel:1 li:2 japan:1 converted:1 baird:1 explicitly:3 tsung:7 sine:5 view:1 hsin:1 closed:2 helped:1 analyze:1 break:1 portion:1 wave:5 recover:2 complicated:2 timely:1 il:1 kaufmann:1 ofthe:1 iterated:4 produced:1 trajectory:34 straight:1 dave:1 history:1 oscillatory:1 touretzky:1 higherdimensional:1 definition:1 frequency:1 obvious:1 yoshizawa:1 associated:1 demers:1 popular:1 hiroyuki:1 amplitude:1 back:3 appears:1 feed:2 steve:1 higher:2 dt:2 originally:1 methodology:1 rand:1 done:1 though:2 anderson:1 just:3 clock:1 until:1 sheng:5 hand:2 working:1 propagation:1 defines:1 usa:1 concept:1 brown:1 requiring:1 counterpart:1 evolution:1 hence:3 inspiration:1 deal:2 white:1 sin:1 during:4 assistance:1 essence:1 criterion:1 leftmost:1 pearlmutter:1 geometrical:1 chaos:1 fi:1 possessing:1 nih:1 common:1 functional:2 physical:2 denver:1 volume:3 extend:1 slight:1 analog:2 eeckman:1 mathematics:1 teaching:3 pointed:1 had:1 dot:1 moving:3 stable:9 longer:1 pu:1 add:1 own:1 showed:2 perspective:1 jolla:1 forcing:9 store:4 verlag:1 success:1 approximators:2 yi:5 der:4 morgan:1 additional:1 converge:2 period:1 signal:8 ii:1 dashed:3 multiple:10 kyoto:2 ing:1 offer:1 hart:1 equally:1 visit:1 prediction:9 multilayer:1 essentially:3 iteration:4 nakajima:1 whereas:1 want:1 addressed:1 posse:1 comment:1 induced:2 tend:1 incorporates:1 xct:1 zipser:2 near:5 ee:1 feedforward:7 intermediate:1 easy:1 spiral:1 affect:1 xj:1 zi:3 architecture:2 idea:1 peter:1 returned:1 york:1 oscillate:1 useful:1 iterating:1 clear:1 locally:2 simplest:1 wanders:1 exist:1 shifted:1 dotted:1 designer:1 disjoint:1 four:4 backward:1 distorted:1 place:1 strange:1 doya:2 oscillation:2 entirely:2 layer:3 nan:1 correspondence:1 fan:1 ijcnn:2 x2:3 nearby:1 speed:1 separable:3 alternate:1 ball:1 combination:1 conjugate:1 remain:1 rtrl:7 happens:1 memorizing:1 tou:1 xo:1 equation:4 tai:1 fail:1 needed:2 apply:1 discretizes:1 eight:1 away:1 appropriate:1 existence:1 assumes:2 running:1 include:2 paradoxical:2 instant:1 giving:1 especially:1 sandwiched:1 r01:1 move:2 traditional:2 gradient:7 link:1 separate:1 berlin:1 unstable:7 reason:1 taiwan:1 ru:3 suboptimally:1 unrolled:3 difficult:2 unfortunately:1 trace:1 design:1 collective:1 perform:1 allowing:1 neuron:1 descent:1 behave:1 situation:1 head:1 paradox:3 ucsd:1 arbitrary:3 rowat:1 inverting:1 pair:5 nonlinearly:2 connection:3 california:3 learned:3 nip:1 able:1 suggested:1 dynamical:6 usually:1 pattern:3 appeared:1 oft:1 oj:3 stinchcombe:1 difficulty:2 force:2 treated:1 scheme:1 imply:1 coupled:1 epoch:1 understanding:1 acknowledgement:1 tangent:1 review:1 determining:1 tractor:2 embedded:1 fully:2 lecture:1 limitation:1 basin:3 editor:3 summary:2 placed:1 supported:1 side:1 understand:1 institute:2 van:4 curve:3 calculated:1 dimension:4 xn:3 valid:1 rich:1 overcome:1 forward:4 author:1 san:4 far:2 reconstructed:1 lippmann:1 keep:1 global:1 overfitting:1 reveals:1 conceptual:1 summing:1 unnecessary:1 xi:4 hydrogen:1 continuous:2 why:3 additionally:1 learn:4 hornik:1 mse:2 excellent:1 complex:3 assured:1 linearly:1 arrow:4 noise:5 garrison:5 sub:1 position:2 momentum:1 explicit:1 xl:1 learns:2 young:1 offset:1 a3:1 adding:3 phd:1 magnitude:1 push:1 simply:1 springer:1 ch:1 gary:1 determines:1 lear:1 towards:2 oscillator:3 change:1 infinite:1 typical:1 reducing:1 kuo:1 la:2 principe:1 dept:1
7,182
944
Synchrony and Desynchrony in Neural Oscillator Networks DeLiang Wang David Terman Department of Computer and Information Science and Center for Cognitive Science The Ohio State University Columbus, Ohio 43210, USA dwang@cis.ohio-state.edu Department of Mathematics The Ohio State University Columbus, Ohio 43210, USA terman@math.ohio-state.edu Abstract An novel class of locally excitatory, globally inhibitory oscillator networks is proposed. The model of each oscillator corresponds to a standard relaxation oscillator with two time scales. The network exhibits a mechanism of selective gating, whereby an oscillator jumping up to its active phase rapidly recruits the oscillators stimulated by the same pattern, while preventing others from jumping up. We show analytically that with the selective gating mechanism the network rapidly achieves both synchronization within blocks of oscillators that are stimulated by connected regions and desynchronization between different blocks. Computer simulations demonstrate the network's promising ability for segmenting multiple input patterns in real time. This model lays a physical foundation for the oscillatory correlation theory of feature binding, and may provide an effective computational framework for scene segmentation and figure/ground segregation. 1 INTRODUCTION A basic attribute of perception is its ability to group elements of a perceived scene into coherent clusters (objects). This ability underlies perceptual processes such as figure/ground segregation, identification of objects, and separation of different objects, and it is generally known as scene segmentation or perceptual organization. Despite the fact 200 DeLiang Wang. David Tennan that humans perform it with apparent ease, the general problem of scene segmentation remains unsolved in the engineering of sensory processing, such as computer vision and auditory processing. Fundamental to scene segmentation is the grouping of similar sensory features and the segregation of dissimilar ones. Theoretical investigations of brain functions and feature binding point to the mechanism of temporal correlation as a representational framework (von der Malsburg, 1981; von der Malsburg and Schneider, 1986). In particular, the correlation theory of von der Malsburg (1981) asserts that an object is represented by the temporal correlation of the fIring activities of the scattered cells coding different features of the object. A natural way of encoding temporal correlation is to use neural oscillations, whereby each oscillator encodes some feature (maybe just a pixel) of an object. In this scheme, each segment (object) is represented by a group of oscillators that shows synchrony (phase-locking) of the oscillations, and different objects are represented by different groups whose oscillations are desynchronized from each other. Let us refer to this form of temporal correlation as oscillatory correlation. The theory of oscillatory correlation has received direct experimental support from the cell recordings in the cat visual cortex (Eckhorn et aI., 1988; Gray et aI., 1989) and other brain regions. The discovery of synchronous oscillations in the visual cortex has triggered much interest from the theoretical community in simulating the experimental results and in exploring oscillatory correlation to solve the problems of scene segmentation. While several demonstrate synchronization in a group of oscillators using local (lateral) connections (Konig and Schillen, 1991; Somers and Kopell, 1993; Wang, 1993, 1995), most of these models rely on long range connections to achieve phase synchrony. It has been pointed out that local connections in reaching synchrony may play a fundamental role in scene segmentation since long-range connections would lead to indiscriminate segmentation (Sporns et aI., 1991; Wang, 1993). There are two aspects in the theory of oscillatory correlation: (1) synchronization within the same object; and (2) desynchronization between different objects. Despite intensive studies on the subject, the question of desynchronization has been hardly addressed. The lack of an effIcient mechanism for de synchronization greatly limits the utility of oscillatory correlation to perceptual organization. In this paper, we propose a new class of oscillatory networks and show that it can rapidly achieve both synchronization within each object and desynchronization between a number of simultaneously presented objects. The network is composed of the following elements: (I) A new model of a basic oscillator; (2) Local excitatory connections to produce phase synchrony within each object; (3) A global inhibitor that receives inputs from the entire network and feeds back with inhibition to produce desynchronization of the oscillator groups representing different objects. In other words, the mechanism of the network consists of local cooperation and global competition. This surprisingly simple neural architecture may provide an elementary approach to scene segmentation and a computational framework for perceptual organization. 2 MODEL DESCRIPTION The building block of this network, a single oscillator i, is defined in the simplest form as a feedback loop between an excitatory unit Xi and an inhibitory unit y( Synchrony and Desynchrony in Neural Oscillator Networks 8 201 , , dxldt= 0 ,, " o. ,, " " ":' . o x -2 2 Figure 1: Nullclines and periodic orbit of a single oscillator as shown in the phase plane. When the oscillator starts at a randomly generated point in the phase plane, it quickly converged to a stable trajectory of a limit cycle. dXi -=3x?dt l dy ? 3 X? I +2-y ,? +p+I?+S? " dt' = e (y(1 + tanh(xl/3) - Yi) Figure 2: Architecture of a two dimensional network with nearest neighbor coupling. The global inhibitor is indicated by the black circle. (la) (lb) where p denotes the amplitude of a Gaussian noise term. Ii represents external stimulation to the oscillator, and Si denotes coupling from other oscillators in the network. The noise term is introduced both to test the robustness of the system and to actively desynchronize different input patterns. The parameter e is chosen to be small. In this case (1), without any coupling or noise, corresponds to a standard relaxation oscillator. The x-nullcline of (1) is a cubic curve, while the y-nullc1ine is a sigmoid function, as shown in Fig. 1. If I > 0, these curves intersect along the middle branch of the cubic, and (1) is oscillatory. The periodic solution alternates between the silent and active phases of near steady state behavior. The parameter yis introduced to control the relative times that the solution spends in these two phases. If I < 0, then the nullc1ines of (1) intersect at a stable fixed point along the left branch of the cubic. In this case the system produces no oscillation. The oscillator model (1) may be interpreted as a model of spiking behavior of a single neuron, or a mean field approximation to a network of excitatory and inhibitory neurons. The network we study here in particular is two dimensional. However, the results can easily be extended to other dimensions. Each oscillator in the network is connected to only its four nearest neighbors, thus forming a 2-D grid. This is the simplest form of local connections. The global inhibitor receives excitation from each oscillator of the grid, and in turn inhibits each oscillator. This architecture is shown in Fig. 2. The intuitive reason why the network gives rise to scene segmentation is the following. When multiple connected objects are mapped onto the grid, local connectivity on the grid will group together the oscillators covered by each object. This grouping will be reflected 202 DeLiang Wang, David Tennan by phase synchrony within each object. The global inhibitor is introduced for desynchronizing the oscillatory responses to different objects. We assume that the coupling term Si in (1) is given by Si = L W ik Soo(xk, 9x ) - W z Soo(z, 9xz ) (2) kEN(i) S (x 9' 00 ,J = __-=-1_ _ 1+ exp[-K(x-e,] (3) where Wik is a connection (synaptic) weight from oscillator k to oscillator i, and N(i) is the set of the neighoring oscillators that connect to i. In this model, NO) is the four immediate neighbors on the 2-D grid, except on the boundaries where N(i) may be either 2 or 3 immediate neighbors. 9x is a threshold (see the sigmoid function of Eq. 3) above which an oscillator can affect its neighbors. W z (positive) is the weight of inhibition from the global inhibitor z, whose activity is defined as (4) where Goo = 0 if Xi < 9zx for every oscillator, and Goo = 1 if Xi ~ 9zx for at least one oscillator i. Hence 9zx represents a threshold. If the activity of every oscillator is below this threshold, then the global inhibitor will not receive any input. In this case z ~ 0 and the oscillators will not receive any inhibition. If, on the other hand, the activity of at least one oscillator is above the threshold 9zx then, the global inhibitor will receive input. In this case z ~ 1, and each oscillator feels inhibition when z is above the threshold 9zx. The parameter l/J determines the rate at which the inhibitor reacts to such stimulation. In summary, once an oscillator is active, it triggers the global inhibitor. This then inhibits the entire network as described in Eq. 1. On the other hand, an active oscillator spreads its activation to its nearest neighbors, again through (1), and from them to its further neighbors. In the next section, we give a number of properties of this system. Besides boundaries, the oscillators on the grid are basically symmetrical. Boundary conditions may cause certain distortions to the stability of synchrous oscillations. Recently, Wang (1993) proposed a mechanism called dynamic normalization to ensure that each oscillator, whether it is in the interior or on a boundary, has equal overall connection weights from its neighbors. The dynamic normalization mechanism is adopted in the present model to form effective connections. For binary images (each pixel being either 0 or 1), the outcome of dynamic normalization is that an effective connection is established between two oscillators if and only if they are neighbors and both of them are activated by external stimulation. The network defined above can readily be applied for segmentation of binary images. For gray-level images (each pixel being in a certain value range), the following slight modification suffices to make the network applicable. An effective connection is established between two oscillators if and only if they are neighbors and the difference of their corresponding pixel values is below a certain threshold. Synchrony and Desynchrony in Neural Oscillator Networks 203 3 ANALYTICAL RESULTS We have formally analyzed the network. Due to space limitations, we can only list the major conclusions without proofs. The interested reader can find the details in Terman and Wang (1994). Let us refer to a pattern as a connected region, and a block be a subset of oscillators stimulated by a given pattern. The following results are about singular solutions in the sense that we formally set E = O. However, as shown in (Terman and Wang, 1994), the results extend to the case E> 0 sufficiently small. Theorem 1. (Synchronization). The parameters of the system can be chosen so that all of the oscillators in a block always jump up simultaneously (synchronize). Moreover, the rate of synchronization is exponential. Theorem 2. (Multiple Patterns) The parameters of the system and a constant T can be chosen to satisfy the following. If at the beginning all the oscillators of the same block synchronize with each other and the temporal distance between any two oscillators belonging to two different blocks is greater than T, then (1) Synchronization within each block is maintained; (2) The blocks activate with a fixed ordering; (3) At most one block is in its active phase at any time. Theorem 3. (Desynchronization) If at the beginning all the oscillators of the system lie not too far away from each other, then the condition of Theorem 2 will be satisfied after some time. Moreover, the time it takes to satisfy the condition is no greater than N cycles, where N is the number of patterns. The above results are true with arbitrary number of oscillators. In summary, the network exhibits a mechanism, referred to as selective gating, which can be intuitively interpreted as follows. An oscillator jumping to its active phase opens a gate to quickly recruit the oscillators of the same block due to local connections. At the same time, it closes the gate to the oscillators of different blocks. Moreover, segmentation of different patterns is achieved very rapidly in terms of oscillation cycles. 4 COMPUTER SIMULATION To illustrate how this network is used for scene segmentation, we have simulated a 2Ox20 oscillator network as defined by (1)-(4). We arbitrarily selected four objects (patterns): two O's, one H, and one I ; and they form the word OHIO . These patterns were simultaneously presented to the system as shown in Figure 3A. Each pattern is a connected region, but no two patterns are connected to each other. All the oscillators stimulated (covered) by the objects received an external input 1=0.2, while the others have 1=-0.02. The amplitude p of the Gaussian noise is set to 0.02. Thus, compared to the external input, a 10% noise is included in every oscillator. Dynamic normalization results in that only two neighboring oscillators stimulated by a single pattern have an effective connection. The differential equations were solved numerically with the following parameter values: E = 0.02, l/J = 3.0; Y= 6.0, f3 = 0.1, K = 50, Ox = -0.5, and 0zx = 0xz = 0.1. The total effective connections were normalized to 6.0. The results described below were robust to considerable changes in the parameters. The phases of all the oscillators on the grid were randomly initialized. 204 DeLiang Wang, David Terman Fig. 3B-3F shows the instantaneous activity (snapshot) of the network at various stages of dynamic evolution. The diameter of each black circle represents the normalized x activity of the corresponding oscillator. Fig. 3B shows a snapshot of the network a few steps after the beginning of the simulation. In Fig. 3B, the activities of the oscillators were largely random. Fig. 3C shows a snapshot after the system had evolved for a short time period. One can clearly seethe effect of grouping and segmentation: all the oscillators belonging to the left 0 were entrained and had large activities. At the same time, the oscillators stimulated by the other three patterns had very small activities. Thus the left 0 was segmented from the rest of the input. A short time later, as shown in Fig. 3D, the oscillators stimulated by the right 0 reached high values and were separated from the rest of the input. Fig. 3E shows another snapshot after Fig. 3D. At this time, pattern I had its turn to be activated and separated from the rest of the input. Finally in Fig. 3F, the oscillators representing H were active and the rest of the input remained silent. This successive "pop-out" of the objects continued in a stable periodic fashion. To provide a complete picture of dynamic evolution, Fig. 30 shows the temporal evolution of each oscillator. Since the oscillators receiving no external input were inactive during the entire simulation process, they were excluded from the display in Fig. 30. The activities of the oscillators stimulated by each object are combined together in the figure. Thus, if they are synchronized, they appear like a single oscillator. In Fig. 30, the four upper traces represent the activities of the four oscillator blocks, and the bottom trace represents the activity of the global inhibitor. The synchronized oscillations within each object are clearly shown within just three cycles of dynamic evolution. 5 DISCUSSION Besides neural plausibility, oscillatory correlation has a unique feature as an computational approach to the engineering of scene segmentation and figure/ground segregation. Due to the nature of oscillations, no single -object can dominate and suppress the perception of the rest of the image permanently. The current dominant object has to give way to other objects being suppressed, and let them have a chance to be spotted. Although at most one object can dominant at any time instant, due to rapid oscillations, a number of objects can be activated over a short time period. This intrinsic dynamic process provides a natural and reliable representation of multiple segmented patterns. The basic principles of selective gating are established for the network with lateral connections beyond nearest neighbors. Indeed, in terms of synchronization, more distant connections even help expedite phase entrainment. In this sense, synchronization with all-to-all connections is an extreme case of our system. With nearest-neighbor connectivity (Fig. 2), any isolated part of an image is considered as a segment. In an noisy image with many tiny regions, segmentation would result in too many small fragments. More distant connections would also provide a solution to this problem. Lateral connections typically take on the form of Gaussian distribution, with the connection strength between two oscillators falling off exponentially. Since global inhibition is superimposed to local excitation, two oscillators positively coupled may be desynchronized if global inhibition is strong enough. Thus, it is unlikely that all objects in an image form a single segment as the result of extended connections. Synchrony and Desynchrony in Neural Oscillator Networks 205 Due to its critical importance for computer vision, scene segmentation has been studied quite extensively. Many techniques have been proposed in the past (Haralick and Shapiro, 1985; Sarkar and Boyer, 1993). Despite these techniques, as pointed out by Haralick and Shapiro (1985), there is no underlying theory of image segmentation, and the techniques tend to be adhoc and emphasize some aspects while ignoring others. Compared to the traditional techniques for scene segmentation, the oscillatory correlation approach offers many unique advantages. The dynamical process is inherently parallel. While conventional computer vision algorithms are based on descriptive criteria and many adhoc heuristics, the network as exemplified in this paper performs computations based on only connections and oscillatory dynamics. The organizational simplicity renders the network particularly feasible for VLSI implementation. Also, continuous-time dynamics allows real time processing, desired by many engineering applications. Acknowledgments DLW is supported in part by the NSF grant IRI-9211419 and the ONR grant NOOOI4-931-0335. DT is supported in part by the NSF grant DMS-9203299LE. References R. Eckhorn, et aI., "Coherent oscillations: A mechanism of feature linking in the visual cortex?" Bioi. Cybem., vol. 60, pp. 121-130, 1988. C.M. Gray, P. Konig, A.K. Engel, and W. Singer, "Oscillatory responses in cat visual cortex exhibit inter-columnar synchronization which reflects global stimulus properties," Nature, vol. 338, pp. 334-337, 1989. R.M. Haralick and L.G. Shapiro, "Image segmentation techniques," Comput. Graphics Image Process., vol. 29, pp. 100-132, 1985. P. Konig and T.B. Schillen, "Stimulus-dependent assembly formation of oscillatory responses: I. Synchronization," Neural Comput., vol. 3, pp. 155-166, 1991. S. Sarkar and K.L. Boyer, "Perceptual organization in computer vision: a review and a proposal for a classificatory structure," IEEE Trans. Syst. Man Cybern., vol. 23, 382-399, 1993. D. Somers, and N. Kopell, "Rapid synchronization through fast threshold modulation," BioI. Cybern, vol. 68, pp. 393-407, 1993. O. Sporns, G. Tononi, and G.M. Edelman, "Modeling perceptual grouping and figureground segregation by means of active reentrant connections," Proc. Natl. A cad. Sci. USA, vol. 88, pp. 129-133, 1991. D. Terman and D.L. Wang, "Global competition and local cooperation in a network of neural oscillators," Physica D, in press, 1994. C. von der Malsburg, "The correlation theory of brain functions," Internal Report 81-2, Max-Planck-Institut for Biophysical Chemistry, Gottingen, FRG, 1981. C. von der Malsburg and W. Schneider, "A neural cocktail-party processor," Bioi. Cybern., vol. 54, pp. 29-40, 1986. D.L. Wang, "Modeling global synchrony in the visual cortex by locally coupled neural oscillators," Proc. 15th Ann. Conf. Cognit. Sci. Soc., pp. 1058-1063, 1993. D.L. Wang, "Emergent synchrony in locally coupled neural oscillators," IEEE Trans. on Neural Networks, in press, 1995. ? 206 DeLiang Wang, David Terman B A c ???????????????????? .... ..... ????????????? ?????? .... ?....??.?.?.... ??????????????????? , ., ???????????????????? ? ????????????????? ~ ??????? ????????????? ??????????????????? ..... ...?...... .......?.... ...... ???????????????????? ???????? ????? ??? ??? ??????????????????? ???????????????????? . . . .?.. . ... ....... ...?...?... . ?????????? ? ????????? .?.. ? ??????????????? ???.?. ?. .?. ?.?. ? ? ??? ??e? ? ? . D E .... " . ........ . ... . . .... ... .. .... ... ... F ? ....... .. ... . ..... . . .. ....... .... . . .. . . G Figure 3. A An image composed of four patterns which were presented (mapped) to a 20x20 grid of oscillators. B A snapshot of the activities of the oscillator grid at the beginning of dynamic evolution. C A snapshot taken shortly after the beginning. D Shortly after C. E Shortly after D. F Shortly after E. G The upper four traces show the combined temporal activities of the oscillator blocks representing the four patterns, respectively, and the bottom trace shows the temporal activity of the global inhibitor. The simulation took 8,000 integration steps.
944 |@word middle:1 indiscriminate:1 open:1 simulation:5 reentrant:1 fragment:1 past:1 current:1 cad:1 si:3 activation:1 readily:1 distant:2 selected:1 plane:2 xk:1 beginning:5 short:3 cognit:1 provides:1 math:1 successive:1 along:2 direct:1 differential:1 ik:1 edelman:1 consists:1 inter:1 indeed:1 rapid:2 behavior:2 xz:2 brain:3 globally:1 nullcline:1 moreover:3 underlying:1 evolved:1 interpreted:2 spends:1 recruit:2 gottingen:1 temporal:8 every:3 control:1 unit:2 grant:3 appear:1 planck:1 segmenting:1 positive:1 engineering:3 local:9 limit:2 despite:3 encoding:1 firing:1 modulation:1 black:2 studied:1 ease:1 range:3 unique:2 acknowledgment:1 classificatory:1 block:14 intersect:2 word:2 onto:1 interior:1 close:1 cybern:3 conventional:1 center:1 iri:1 simplicity:1 continued:1 dominate:1 stability:1 feel:1 play:1 trigger:1 element:2 particularly:1 lay:1 bottom:2 role:1 wang:13 solved:1 region:5 connected:6 cycle:4 ordering:1 goo:2 locking:1 dynamic:11 segment:3 easily:1 emergent:1 represented:3 cat:2 various:1 separated:2 fast:1 effective:6 activate:1 formation:1 outcome:1 apparent:1 whose:2 quite:1 solve:1 heuristic:1 distortion:1 ability:3 noisy:1 triggered:1 advantage:1 descriptive:1 analytical:1 biophysical:1 took:1 propose:1 neighboring:1 loop:1 rapidly:4 achieve:2 representational:1 description:1 asserts:1 intuitive:1 competition:2 konig:3 cluster:1 produce:3 object:29 help:1 coupling:4 illustrate:1 nearest:5 received:2 eq:2 strong:1 soc:1 synchronized:2 attribute:1 human:1 frg:1 suffices:1 investigation:1 elementary:1 exploring:1 physica:1 sufficiently:1 considered:1 ground:3 exp:1 major:1 achieves:1 perceived:1 proc:2 applicable:1 tanh:1 engel:1 reflects:1 clearly:2 inhibitor:11 gaussian:3 always:1 reaching:1 nullclines:1 haralick:3 superimposed:1 greatly:1 sense:2 dependent:1 entire:3 typically:1 unlikely:1 vlsi:1 boyer:2 selective:4 interested:1 pixel:4 overall:1 integration:1 field:1 once:1 equal:1 f3:1 represents:4 others:3 stimulus:2 terman:7 report:1 few:1 randomly:2 composed:2 simultaneously:3 phase:13 organization:4 interest:1 tononi:1 analyzed:1 extreme:1 activated:3 natl:1 jumping:3 institut:1 initialized:1 orbit:1 circle:2 desired:1 isolated:1 theoretical:2 modeling:2 organizational:1 subset:1 too:2 graphic:1 connect:1 periodic:3 combined:2 fundamental:2 off:1 receiving:1 together:2 quickly:2 connectivity:2 von:5 again:1 satisfied:1 cognitive:1 external:5 conf:1 actively:1 syst:1 de:1 chemistry:1 coding:1 satisfy:2 later:1 reached:1 start:1 parallel:1 synchrony:11 largely:1 identification:1 schillen:2 basically:1 trajectory:1 zx:6 kopell:2 processor:1 converged:1 oscillatory:14 synaptic:1 pp:8 dm:1 proof:1 dxi:1 unsolved:1 auditory:1 noooi4:1 dlw:1 segmentation:19 amplitude:2 back:1 feed:1 dt:3 reflected:1 response:3 ox:1 just:2 stage:1 correlation:14 hand:2 receives:2 lack:1 columbus:2 gray:3 indicated:1 usa:3 effect:1 building:1 normalized:2 true:1 evolution:5 analytically:1 hence:1 excluded:1 during:1 maintained:1 whereby:2 steady:1 excitation:2 criterion:1 complete:1 demonstrate:2 performs:1 image:11 instantaneous:1 ohio:7 novel:1 recently:1 sigmoid:2 stimulation:3 spiking:1 physical:1 exponentially:1 extend:1 slight:1 linking:1 numerically:1 refer:2 ai:4 eckhorn:2 mathematics:1 grid:9 pointed:2 had:4 stable:3 cortex:5 inhibition:6 deliang:5 dominant:2 certain:3 binary:2 arbitrarily:1 onr:1 der:5 yi:2 greater:2 schneider:2 period:2 ii:1 branch:2 multiple:4 expedite:1 segmented:2 plausibility:1 offer:1 long:2 spotted:1 underlies:1 basic:3 vision:4 normalization:4 represent:1 achieved:1 cell:2 receive:3 proposal:1 addressed:1 singular:1 desynchronized:2 figureground:1 rest:5 recording:1 subject:1 tend:1 entrained:1 near:1 enough:1 reacts:1 affect:1 architecture:3 silent:2 intensive:1 synchronous:1 whether:1 inactive:1 utility:1 render:1 cause:1 hardly:1 cocktail:1 generally:1 covered:2 maybe:1 locally:3 extensively:1 ken:1 simplest:2 diameter:1 shapiro:3 nsf:2 inhibitory:3 vol:8 group:6 four:8 threshold:7 falling:1 relaxation:2 somers:2 reader:1 separation:1 oscillation:11 dy:1 display:1 activity:15 strength:1 scene:13 encodes:1 aspect:2 inhibits:2 department:2 alternate:1 belonging:2 suppressed:1 dwang:1 modification:1 intuitively:1 taken:1 segregation:5 equation:1 remains:1 turn:2 desynchronize:1 mechanism:9 singer:1 adopted:1 away:1 simulating:1 robustness:1 shortly:4 gate:2 permanently:1 denotes:2 ensure:1 assembly:1 malsburg:5 instant:1 question:1 traditional:1 exhibit:3 distance:1 mapped:2 lateral:3 simulated:1 sci:2 reason:1 besides:2 x20:1 trace:4 rise:1 suppress:1 implementation:1 perform:1 upper:2 neuron:2 snapshot:6 immediate:2 extended:2 lb:1 arbitrary:1 community:1 sarkar:2 david:5 introduced:3 connection:23 adhoc:2 coherent:2 established:3 pop:1 trans:2 beyond:1 below:3 pattern:18 perception:2 dynamical:1 exemplified:1 reliable:1 soo:2 max:1 sporns:2 critical:1 natural:2 rely:1 synchronize:2 representing:3 scheme:1 wik:1 picture:1 coupled:3 review:1 discovery:1 relative:1 synchronization:13 limitation:1 foundation:1 principle:1 tiny:1 excitatory:4 cooperation:2 summary:2 surprisingly:1 supported:2 neighbor:12 feedback:1 curve:2 dimension:1 boundary:4 preventing:1 sensory:2 jump:1 far:1 party:1 emphasize:1 global:16 active:8 cybem:1 symmetrical:1 xi:3 continuous:1 why:1 stimulated:8 promising:1 nature:2 robust:1 inherently:1 ignoring:1 spread:1 noise:5 positively:1 fig:14 referred:1 scattered:1 cubic:3 fashion:1 exponential:1 xl:1 lie:1 comput:2 perceptual:6 theorem:4 remained:1 gating:4 desynchronization:6 list:1 grouping:4 intrinsic:1 importance:1 ci:1 columnar:1 forming:1 visual:5 binding:2 corresponds:2 determines:1 chance:1 bioi:3 ann:1 oscillator:75 man:1 considerable:1 change:1 feasible:1 included:1 except:1 entrainment:1 called:1 total:1 experimental:2 la:1 formally:2 internal:1 support:1 dissimilar:1
7,183
945
The Electrotonic Transformation: a Tool for Relating Neuronal Form to Function Nicholas T. Carnevale Department of Psychology Yale University New Haven, CT 06520 Kenneth Y. Tsai Department of Psychology Yale University New Haven, CT 06520 Brenda J. Claiborne Division of Life Sciences University of Texas San Antonio, TX 79285 Thomas H. Brown Department of Psychology Yale University New Haven, CT 06520 Abstract The spatial distribution and time course of electrical signals in neurons have important theoretical and practical consequences. Because it is difficult to infer how neuronal form affects electrical signaling, we have developed a quantitative yet intuitive approach to the analysis of electrotonus. This approach transforms the architecture of the cell from anatomical to electrotonic space, using the logarithm of voltage attenuation as the distance metric. We describe the theory behind this approach and illustrate its use. 1 INTRODUCTION The fields of computational neuroscience and artificial neural nets have enjoyed a mutually beneficial exchange of ideas. This has been most evident at the network level, where concepts such as massive parallelism, lateral inhibition, and recurrent excitation have inspired both the analysis of brain circuits and the design of artificial neural net architectures. Less attention has been given to how properties of the individual neurons or processing elements contribute to network function. Biological neurons and brain circuits have 70 Nicholas Carnevale, Kenneth Y. Tsai, Brenda J. Claiborne, Thomas H. Brown been simultaneously subject to eons of evolutionary pressure. This suggests an essential interdependence between neuronal form and function, on the one hand, and the overall architecture and operation of biological neural nets, on the other. Therefore reverseengineering the circuits of the brain appears likely to reveal design principles that rely upon neuronal properties. These principles may have maximum utility in the design of artificial neural nets that are constructed of processing elements with greater similarity to biological neurons than those which are used in contemporary designs. Spatiotemporal extent is perhaps the most obvious difference between real neurons and processing elements. The processing element of most artificial neural nets is essentially a point in time and space. Its activation level is the instantaneous sum of its synaptic inputs. Of particular relevance to Hebbian learning rules, all synapses are exposed to the same activation level. These simplifications may insure analytical and implementational simplicity, but they deviate sharply from biological reality. Membrane potential, the biological counterpart of activation level, is neither instantaneous nor spatially uniform. Every cell has finite membrane capacitance, and all ionic currents are finite, so membrane potential must lag behind synaptic inputs. Furthermore, membrane capacitance and cytoplasmic resistance dictate that membrane potential will almost never be uniform throughout a living neuron embedded in the circuitry of the brain. The combination of ever-changing synaptic inputs with cellular anatomical and biophysical properties guarantees the existence of fluctuating electrical gradients. Consider the task of building a massively parallel neural net from processing elements with such "nonideal" characteristics. Imagine moreover that the input surface of each processing element is an extensive, highly branched structure over which approximately 10,000 synaptic inputs are distributed. It might be tempting to try to minimize or work around the limitations imposed by device physics. However, a better strategy might be to exploit the computational consequences of these properties by making them part of the design, thereby turning these apparent weaknesses into strengths. To facilitate an understanding of the spatiotemporal dynamics of electrical signaling in neurons, we have developed a new theoretical approach to linear electrotonus and a new way to make practical use of this theory. We present this method and illustrate its application to the analysis of synaptic interactions in hippocampal pyramidal cells. 2 THEORETICAL BACKGROUND Our method draws upon and extends the results of two prior approaches: cable theory and two-port analysis. 2.1 CABLE THEORY The modern use of cable theory in neuroscience began almost four decades ago with the work of RaIl (1977). Much of the attraction of cable theory derives from the conceptual simplicity of the steady-state decay of voltage with distance along an infinite cylindrical cable: V(x) = Voe- xl). where x is physical distance and .4 is the length constant. This exponential relationship makes it useful to define the electrotonic distance X as the The Electronic Transfonnation: A Tool for Relating Neuronal Fonn to Function 7l logarithm of the signal attenuation: X = lnVo/V(x). In an infinite cylindrical cable, electrotonic distance is directly proportional to physical distance: X = x/2 . However, cable theory is difficult to apply to real neurons since dendritic trees are neither infinite nor cylindrical. Because of their anatomical complexity and irregular variations of branch diameter and length, attenuation in neurons is not an exponential function of distance. Even if a cell met the criteria that would allow its dendrites to be reduced to a finite equivalent cylinder (RaIl 1977), voltage attenuation would not bear a simple exponential relationship to X but instead would vary inversely with a hyperbolic function (Jack et a!. 1983). 2.2 TWO-PORT THEORY Because of the limitations and restrictions of cable theory, Carnevale and Johnston (1982) turned to two-port analysis. Among their conclusions, three are most relevant to this discussion. Figure 1: Attenuation is direction-dependent. The first is that signal attenuation depends on the direction of signal propagation. Suppose that i and J are two points in a cell where i is "upstream" from J (voltage is spreading from i to J), and define the voltage attenuation from i to j: A~ = ~ IV). Next suppose that the direction of signal propagation is reversed, so that j is now upstream from i, and define the voltage attenuation A ~ =Vj I~. In general these two *- AV. attenuations will not be equal: A~ 1) )1 They also showed that voltage attenuation in one direction is identical to current attenuation in the opposite direction (Carnevale and Johnston 1982). Suppose current Ii enters the cell at i, and the current that is captured by a voltage clamp at J is Ii' and define the current attenuation A; = Ii j ' Because of the directional reciprocity 11 Similarly, if we interchange the between current and voltage attenuation, A!1) = AV. j1 current entry and voltage clamp sites, the current attenuation ratio would be A )1l = A 1)~ . Finally, they found that charge and DC current attenuation in the same direction are identical (Carnevale and Johnston 1982). Therefore the spread of electrical signals between any two points is completely characterized by the voltage attenuations in both directions. 72 Nicholas Carnevale, Kenneth Y. Tsai, Brenda 1. Claiborne, Thomas H. Brown 2.3 THE ELECTROTONIC TRANSFORMATION The basic idea of the electrotonic transformation is to remap the cell from anatomical space into "electrotonic space," where the distance between points reflects the attenuation of an electrical signal spreading between them. Because of the critical role of membrane potential in neuronal function, it is usually most appropriate to deal with voltage attenuations. 2.3.1 The Distance Metric We use the logarithm of attenuation between points as the distance metric in electrotonic space: Li; = InAij (Brown et a1. 1992, Zador et a1. 1991). To appreciate the utility of this definition, consider voltage spreading from point i to point j , and suppose that k is on the direct path between i and j. The voltage attenuations are AI~ = V,/~ , At = Vk I~, and A& = V, IV; =A~ At . This last equation and our definition of L establish the additive property of electrotonic distance Llj = Lik + LIg" That is, electrotonic distances are additive over a path that has a consistent directIon of signal propagation. This justifies using the logarithm of attenuation as a metric for the electrical separation between points in a cell. At this point several important facts should be noted. First, unlike the electrotonic distance X of classical cable theOI)" our new definition of electrotonic distance L always bears a simple and direct logarithmic relationship to attenuation. Second, because of membrane capacitance, attenuation increases with frequency. Since both steady-state and transient signals are of interest, we evaluate attenuations at several different frequencies . Third, attenuation is direction-dependent and usually asymmetric. Therefore at every frequency of interest, each branch of the cell has two different representations in electrotonic space depending on the direction of signal flow. 2.3.2 Representing a Neuron in Electrotonic Space Since attenuation depends on direction, it is necessary to construct transforms in pairs for each frequency of interest, one for signal spread away from a reference point (Vout ) and the other for spread toward it (Vin ). The soma is often a good choice for the reference point, but any point in the cell could be used, and a different vantage point might be more appropriate for particular analyses. The only difference between using one point i as the reference instead of any other point j is in the direction of signal propagation along the direct path between i and j (dashed arrows in Figure 2). where Vout relative to i is the same as Vin relative to j and vice versa. The directions of signal flow and therefore the attenuations along all other branches of the cell are unchanged. Thus the transforms relative to i and j differ only along the direct path ij, and once the T/~ut and Vin transforms have been created for one reference i, it is easy to assemble the transforms with respect to any other reference j. The Electronic Transformation: A Tool for Relating Neuronal Form to Function 73 .,. .,. Figure 2: Effect of reference point location on direction of signal propagation. We have found two graphical representations of the transform to be particularly useful. "Neuromorphic figures," in which the cell is redrawn so that the relative orientation of branches is preserved (Figures 3 and 4), can be readily compared to the original anatomy for quick, ''big picture" insights regarding synaptic integration and interactions. For more quantitative analyses, it is helpful to plot electrotonic distance from the reference point as a function of anatomical distance (Tsai et al. 1993). 3 COMPUTATIONAL METHODS The voltage attenuations along each segment of the cell are calculated from detailed, accurate morphometric data and the best available experimental estimates of the biophysical properties of membrane and cytoplasm. Any neural simulator like NEURON (Hines 1989) could be used to find the attenuations for the DC Vout transform. The DC Vin attenuations are more time consuming because a separate run must be performed for each of the dendritic terminations. However, the AC attenuations impose a severe computational burden on time-domain simulators because many cycles are required for the response to settle. For example, calculating the DC Vout attenuations in a hippocampal pyramidal cell relative to the soma took only a few iterations on a SUN Sparc 10-40, but more than 20 hours were required for 40 Hz (Tsai et al. 1994). Finding the full set of attenuations for a Vin transform at 40 Hz would have taken almost three months. Therefore we designed an O(N) algorithm that achieves high computational efficiency by operating in the frequency domain and taking advantage of the branched architecture of the cell. In a series of recursive walks through the cell, the algorithm applies Kirchhoff's laws to compute the attenuations in each branch. The electrical characteristics of each segment of the cell are represented by an equivalent T circuit. Rather than "lump" the properties of cytoplasm and membrane into discrete resistances and capacitances, we determine the elements of these equivalent T circuits directly from complex impedance functions that we derived from the impulse response of a finite cable (Tsai et a1. 1994). Since each segment is treated as a cable rather than an isopotential compartment, the resolution of the spatial grid does not affect accuracy, and there is no need to increase the resolution of the spatial grid in order to preserve accuracy as frequency increases. This is an important consideration for hippocampal neurons, which have long membrane time constants and begin to show increased attenuations at frequencies as low as 2 - 5 Hz (Tsai et al. 1994). It also allows us to treat a long unbranched neurite of nearly constant diameter as a single cylinder. Thus runtimes scale linearly with the number of grid points, are independent of frequency, and we can even reduce the number of grid points if the diameters of adjacent 74 Nicholas Carnevale. Kenneth Y. Tsai. Brenda 1. Claiborne. Thomas H. Brown unbranched segments are similar enough. A benchmark of a program that uses our algorithm with NEURON showed a speedup of more than four orders of magnitude without sacrificing accuracy (2 seconds vs. 20 hours to calculate the Vout attenuations at 40 Hz in a CAl pyramidal neuron model with 2951 grid points) (Tsai et al. 1994). 4 RESULTS 4.1 DC TRANSFORMS OF A CAl PYRAMIDAL CELL Figure 3 shows a two-dimensional projection of the anatomy of a rat CA 1 pyramidal neuron (cell 524, left) with neuromorphic renderings of its DC Vout and Vin transforms (middle and right) relative to the soma. The three-dimensional anatomical data were obtained from HRP-filled cells with a computer microscope system as described elsewhere (Rihn and Claiborne 1992, Claiborne 1992). The passive electrical properties used to compute the attenuations were Ri = 200 Ocm, em = 1 J.lF/cm2 (for nonzero frequencies, not shown here) and Rm = 30 kncm2 (Spruston and Johnston 1992). 524 1 Figure 3: CAl pyramidal cell anatomy (cell 524, left) with neuromorphic renderings of Vout (middle) and Vin (right) transforms at DC. The Vout transform is very compact, indicating that voltage propagates from the soma into the dendrites with relatively little attenuation. The basilar dendrites and the terminal branches of the primary apical dendrite are almost invisible, since they are nearly isopotential along their lengths. Despite the fact that the primary apical dendrite has a larger diameter than any of its daughter branches, most of the voltage drop for somatofugal signaling is in the primary apical. Therefore it accounts for almost all of the electrotonic length of the cell in the Vout transform. The Vin transform is far more extensive, but most of the electrotonic length of the cell is now in the basilar and terminal apical branches. This reflects the loading effect of downstream membrane on these thin dendritic branches. 4.2 SYNAPTIC INTERACTIONS The transform can also give clues to possible effects of electrotonic architecture on voltage-dependent forms of associative synaptic plasticity and other kinds of synaptic interactions. Suppose the cell of Figure 3 receives a weak or "student" synaptic input The Electronic Transformation: A Tool for Relating Neuronal Form to Function 75 located 400 J.lm from the soma on the primary apical dendrite, and a strong or "teacher" input is situated 300 J.lm from the soma on the same dendrite. [!] student @ teacher A. cell 524 B. cell 503 Figure 4: Analysis of synaptic interactions. The anatomical arrangement is depicted on the left in Figure 4A ("student" = square, "teacher" = circle). The Vin transform with respect to the student (right figure of this pair) shows that voltage spreads from the teacher to the student synapse with little attenuation, which would favor voltage-dependent associative interactions. Figure 4B shows a different CAl pyramidal cell in which the apical dendrite bifurcates shortly after arising from the soma. Two teacher synapses are indicated, one on the same branch as the student and the other on the opposite branch. The Vin transform with respect to the student (right figure of this pair) shows clearly that the teacher synapse on the same branch is closely coupled to the student, but the other is electrically much more remote and less likely to influence the student synapse. 5. SUMMARY The electrotonic transformation is based on a logical, internally consistent conceptual approach to understanding the propagation of electrical signals in neurons. In this paper we described two methods for graphically presenting the results of the transformation: neuromorphic rendering, and plots of electrotonic distance vs. anatomical distance. Using neuromorphic renderings, we illustrated the electrotonic properties of a previously unreported hippocampal CAl pyramidal neuron as viewed from the soma (cell 524, Figure 3). We also extended the use of the transformation to the study of associative interactions between "teacher" and "student" synapses by analyzing this cell from the viewpoint of a "student" synapse located in the apical dendrites, contrasting this result with a different cell that had a bifurcated primary apical dendrite (cell 503, Figure 4). This demonstrates the versatility of the electrotonic transformation, and shows how it can convey the electrical signaling properties of neurons in ways that are quickly and easily comprehended. This understanding is important for several reasons. First, electrotonus affects the integration and interaction of synaptic inputs, regulates voltage-dependent mechanisms of synaptic plasticity, and influences the interpretation of intracellular recordings. In addition, phylogeny, development, aging, and response to injury and disease are all accompanied by alterations of neuronal morphology, some subtle and some profound. 76 Nicholas Carnevale, Kenneth Y. Tsai, Brenda J. Claiborne, Thomas H. Brown The significance of these changes for brain function becomes clear only if their effects on neuronal signaling are reckoned. Finally, there is good reason to expect that neuronal electrotonus is highly relevant to the design of artificial neural networks. Acknowledgments We thank R.B. Gonzales and M.P. O'Boyle for their contributions to the morphometric analysis, and Z.F. Mainen for assisting in the initial development of graphical rendering. This work was supported in part by ONR, ARPA, and the Yale Center for Theoretical and Applied Neuroscience (CTAN). References Brown, T.H. , Zador, A.M. , Mainen, Z.F. and Claiborne, BJ. Hebbian computations in hippocampal dendrites and spines. In: Single Neuron Computation, eds. McKenna, T., Davis, J. and Zornetzer, S.F., New York, Academic Press, 1992, pp. 81-116. Carnevale, N. T. and Johnston, D.. Electrophysiological characterization of remote chemical synapses. J. Neurophysiol. 47:606-621 , 1982. Claiborne, BJ. The use of computers for the quantitative, three-dimensional analysis of dendritic trees. In: Methods in Neuroscience. Vol. 10: Computers and Computation in the Neurosciences, ed. Conn, P.M., New York, Academic Press, 1992, pp. 315-330. Hines, M. A program for simulation of nerve equations with branching geometries. Internat. J. Bio-Med Computat. 24:55-68, 1989. Rall, W. . Core conductor theory and cable properties of neurons. In: Handbook of Physiology, The Nervous System , ed. Kandel, E.R., Bethesda, MD, Am. Physiol. Soc. , 1977, pp.39-98. Rihn, L.L. and Claiborne, BJ. Dendritic growth and regression in rat dentate granule cells during late postnatal development. Brain Res. Dev. 54(1): 115-24, 1990. Spruston, N. and Johnston, D. Perforated patch-clamp analysis of the passive membrane properties of three classes of hippocampal neurons. J. Neurophysiol. 67 :508529, 1992. Tsai, K.Y., Carnevale, N.T. , Claiborne, BJ. and Brown, T.H. Morphoelectrotonic transforms in three classes of rat hippocampal neurons. Soc. Neurosci. Abst. 19: 1522, 1993. Tsai, K.Y. , Carnevale, N.T. , Claiborne, BJ. and Brown, T.H. Efficient mapping from neuroanatomical to electrotonic space. Network 5:21-46, 1994. Zador, A.M. , Claiborne, BJ. and Brown, T.H. Electrotonic transforms of hippocampal neurons. Soc. Neurosci. Abst. 17: 1515, 1991.
945 |@word cylindrical:3 middle:2 loading:1 termination:1 cm2:1 simulation:1 pressure:1 fonn:1 thereby:1 initial:1 series:1 mainen:2 current:9 activation:3 yet:1 must:2 readily:1 physiol:1 additive:2 j1:1 plasticity:2 plot:2 designed:1 drop:1 v:2 device:1 nervous:1 postnatal:1 core:1 characterization:1 contribute:1 location:1 along:6 constructed:1 direct:4 profound:1 interdependence:1 spine:1 nor:2 simulator:2 brain:6 terminal:2 morphology:1 inspired:1 ctan:1 rall:1 little:2 becomes:1 begin:1 insure:1 moreover:1 circuit:5 kind:1 developed:2 contrasting:1 finding:1 transformation:9 guarantee:1 quantitative:3 every:2 attenuation:39 charge:1 growth:1 rm:1 demonstrates:1 bio:1 internally:1 treat:1 aging:1 consequence:2 despite:1 analyzing:1 path:4 approximately:1 might:3 suggests:1 practical:2 acknowledgment:1 recursive:1 lf:1 signaling:5 hyperbolic:1 dictate:1 physiology:1 projection:1 vantage:1 cal:5 influence:2 restriction:1 equivalent:3 imposed:1 quick:1 center:1 graphically:1 attention:1 zador:3 resolution:2 simplicity:2 boyle:1 rule:1 attraction:1 insight:1 variation:1 imagine:1 suppose:5 massive:1 us:1 element:7 particularly:1 located:2 asymmetric:1 role:1 electrical:11 enters:1 calculate:1 cycle:1 sun:1 remote:2 contemporary:1 disease:1 complexity:1 dynamic:1 segment:4 exposed:1 upon:2 division:1 efficiency:1 completely:1 neurophysiol:2 easily:1 kirchhoff:1 represented:1 tx:1 describe:1 artificial:5 bifurcated:1 cytoplasmic:1 apparent:1 lag:1 larger:1 favor:1 transform:9 associative:3 advantage:1 abst:2 biophysical:2 net:6 analytical:1 took:1 bifurcates:1 clamp:3 interaction:8 turned:1 relevant:2 intuitive:1 lig:1 illustrate:2 recurrent:1 depending:1 ac:1 basilar:2 ij:1 strong:1 soc:3 met:1 differ:1 direction:14 anatomy:3 closely:1 redrawn:1 transient:1 settle:1 exchange:1 biological:5 dendritic:5 around:1 mapping:1 bj:6 dentate:1 lm:2 circuitry:1 vary:1 achieves:1 spreading:3 vice:1 tool:4 reflects:2 clearly:1 always:1 rather:2 voltage:21 derived:1 vk:1 am:1 helpful:1 dependent:5 overall:1 among:1 orientation:1 development:3 spatial:3 integration:2 field:1 equal:1 never:1 construct:1 once:1 runtimes:1 identical:2 nearly:2 thin:1 haven:3 few:1 modern:1 simultaneously:1 preserve:1 individual:1 hrp:1 geometry:1 versatility:1 cylinder:2 interest:3 highly:2 severe:1 weakness:1 behind:2 accurate:1 necessary:1 tree:2 iv:2 filled:1 logarithm:4 walk:1 spruston:2 circle:1 sacrificing:1 re:1 theoretical:4 arpa:1 increased:1 dev:1 injury:1 implementational:1 neuromorphic:5 apical:8 entry:1 uniform:2 teacher:7 spatiotemporal:2 physic:1 quickly:1 li:1 account:1 potential:4 accompanied:1 alteration:1 student:11 depends:2 performed:1 try:1 parallel:1 vin:10 contribution:1 minimize:1 square:1 compartment:1 accuracy:3 characteristic:2 directional:1 vout:9 weak:1 ionic:1 ago:1 synapsis:4 synaptic:13 ed:3 definition:3 frequency:9 pp:3 obvious:1 logical:1 ut:1 electrophysiological:1 subtle:1 nerve:1 appears:1 response:3 synapse:4 furthermore:1 hand:1 receives:1 propagation:6 indicated:1 impulse:1 reveal:1 perhaps:1 facilitate:1 effect:4 building:1 brown:10 concept:1 counterpart:1 chemical:1 spatially:1 nonzero:1 illustrated:1 deal:1 adjacent:1 during:1 branching:1 davis:1 noted:1 excitation:1 steady:2 rat:3 criterion:1 hippocampal:8 presenting:1 evident:1 invisible:1 passive:2 instantaneous:2 jack:1 consideration:1 began:1 physical:2 regulates:1 interpretation:1 relating:4 versa:1 ai:1 enjoyed:1 grid:5 similarly:1 ocm:1 unreported:1 had:1 neurite:1 similarity:1 surface:1 inhibition:1 operating:1 internat:1 showed:2 rihn:2 sparc:1 massively:1 onr:1 life:1 captured:1 greater:1 impose:1 determine:1 tempting:1 signal:15 assisting:1 living:1 branch:12 ii:3 lik:1 infer:1 dashed:1 hebbian:2 full:1 characterized:1 academic:2 long:2 a1:3 basic:1 regression:1 essentially:1 metric:4 iteration:1 cell:33 irregular:1 preserved:1 background:1 microscope:1 addition:1 johnston:6 pyramidal:8 unlike:1 subject:1 hz:4 recording:1 med:1 transfonnation:1 flow:2 lump:1 easy:1 enough:1 rendering:5 affect:3 psychology:3 architecture:5 opposite:2 reduce:1 idea:2 regarding:1 texas:1 utility:2 remap:1 resistance:2 york:2 electrotonic:24 antonio:1 useful:2 detailed:1 clear:1 transforms:10 situated:1 diameter:4 reduced:1 computat:1 neuroscience:5 arising:1 anatomical:8 discrete:1 vol:1 conn:1 four:2 soma:8 changing:1 neither:2 branched:2 kenneth:5 downstream:1 sum:1 run:1 extends:1 almost:5 throughout:1 electronic:3 separation:1 patch:1 draw:1 ct:3 simplification:1 yale:4 assemble:1 strength:1 sharply:1 ri:1 relatively:1 llj:1 department:3 speedup:1 combination:1 electrically:1 membrane:12 beneficial:1 em:1 bethesda:1 cable:12 making:1 taken:1 equation:2 mutually:1 previously:1 mechanism:1 carnevale:11 available:1 operation:1 apply:1 fluctuating:1 appropriate:2 away:1 nicholas:5 shortly:1 existence:1 thomas:5 original:1 neuroanatomical:1 graphical:2 calculating:1 exploit:1 eon:1 establish:1 granule:1 classical:1 appreciate:1 unchanged:1 capacitance:4 arrangement:1 strategy:1 primary:5 md:1 evolutionary:1 gradient:1 distance:18 reversed:1 separate:1 lateral:1 thank:1 voe:1 gonzales:1 extent:1 cellular:1 toward:1 reason:2 length:5 reciprocity:1 relationship:3 ratio:1 difficult:2 daughter:1 design:6 av:2 neuron:23 benchmark:1 finite:4 extended:1 ever:1 dc:7 pair:3 required:2 extensive:2 hour:2 parallelism:1 usually:2 inaij:1 program:2 critical:1 treated:1 rely:1 turning:1 representing:1 inversely:1 picture:1 created:1 coupled:1 brenda:5 perforated:1 deviate:1 prior:1 understanding:3 relative:6 law:1 embedded:1 expect:1 bear:2 limitation:2 proportional:1 consistent:2 propagates:1 principle:2 port:3 morphometric:2 viewpoint:1 course:1 elsewhere:1 summary:1 supported:1 last:1 allow:1 taking:1 distributed:1 calculated:1 interchange:1 clue:1 san:1 far:1 cytoplasm:2 compact:1 claiborne:13 handbook:1 conceptual:2 consuming:1 zornetzer:1 decade:1 reality:1 impedance:1 ca:1 dendrite:11 upstream:2 complex:1 domain:2 vj:1 significance:1 spread:4 linearly:1 arrow:1 big:1 intracellular:1 neurosci:2 convey:1 neuronal:11 site:1 exponential:3 xl:1 kandel:1 rail:2 third:1 late:1 mckenna:1 decay:1 derives:1 essential:1 burden:1 magnitude:1 nonideal:1 justifies:1 depicted:1 logarithmic:1 likely:2 applies:1 hines:2 month:1 viewed:1 change:1 infinite:3 conductor:1 isopotential:2 experimental:1 indicating:1 phylogeny:1 relevance:1 tsai:12 evaluate:1
7,184
946
Boosting the Performance of RBF Networks with Dynamic Decay Adjustment Michael R. Berthold Forschungszentrum Informatik Gruppe ACID (Prof. D. Schmid) Haid-und-Neu-Strasse 10-14 76131 Karlsruhe, Germany eMail: berthold@fzLde Jay Diamond Intel Corporation 2200 Mission College Blvd. Santa Clara, CA, USA 95052 MS:SC9-15 eMail: jdiamond@mipos3.intel.com Abstract Radial Basis Function (RBF) Networks, also known as networks of locally-tuned processing units (see [6]) are well known for their ease of use. Most algorithms used to train these types of networks, however, require a fixed architecture, in which the number of units in the hidden layer must be determined before training starts. The RCE training algorithm, introduced by Reilly, Cooper and Elbaum (see [8]), and its probabilistic extension, the P-RCE algorithm, take advantage of a growing structure in which hidden units are only introduced when necessary. The nature of these algorithms allows training to reach stability much faster than is the case for gradient-descent based methods. Unfortunately P-RCE networks do not adjust the standard deviation of their prototypes individually, using only one global value for this parameter. This paper introduces the Dynamic Decay Adjustment (DDA) algorithm which utilizes the constructive nature of the P-RCE algorithm together with independent adaptation of each prototype's decay factor. In addition, this radial adjustment is class dependent and distinguishes between different neighbours. It is shown that networks trained with the presented algorithm perform substantially better than common RBF networks. 522 1 Michael R. Berthold, Jay Diamond Introduction Moody and Darken proposed Networks with locally-tuned processing units, which are also known as Radial Basis Functions (RBFs, see [6]). Networks of this type have a single layer of units with a selective response for some range of the input variables. Earn unit has an overall response function, possibly a Gaussian: D_("') (11x - 2riW) X = exp - .J.t,i (1) (Ii Here x is the input to the network, ri denotes the center of the i-th RBF and (Ii determines its standard deviation. The second layer computes the output function for each class as follows: m (2) i=l with m indicating the number of RBFs and Ai being the weight for each RBF. Moody and Darken propose a hybrid training, a combination of unsupervised clustering for the centers and radii of the RBFs and supervised training of the weights. Unfortunately their algorithm requires a fixed network topology, which means that the number of RBFs must be determined in advance. The same problem applies to the Generalized Radial Basis Functions (GRBF), proposed in [12]. Here a gradient descent technique is used to implement a supervised training of the center locations, which has the disadvantage of long training times. In contrast RCE (Restricted Coulomb Energy) Networks construct their architecture dynamically during training (see [7] for an overview). This algorithm was inspired by systems of charged particles in a three-dimensional space and is analogous to the Liapunov equation: 1 m Qi (3) ~ = - L i=l II'"X _ T,...?II 2L l: where ~ is the electrostatic potential induced by fixed particles with charges -Qi and locations ri. One variation of this type of networks is the so called P-RCE network, which attempts to classify data using a probabilistic distribution derived from the training set. The underlying training algorithm for P-RCE is identical to RCE training with gaussian activation functions used in the forward pass to resemble a Probabilistic Neural Network (PNN [10]). PNNs are not suitable for large databases because they commit one new prototype for each training pattern they encounter, effectively becoming a referential memory scheme. In contrast, the P-RCE algorithm introduces a new prototype only when necessary. This occurs when the prototype of a conflicting class misclassifies the new pattern during the training phase. The probabilistic extension is modelled by incrementing the a-priori rate of occurrence for prototypes of the same class as the input vector, therefore weights are only connecting RBFs and an output node of the same class. The recall phase of the P-RCE network is similar to RBFs, except that it uses one global radius for all prototypes and scales each gaussian by the a-priori rate of occurrence: (4) Boosting the Performance of RBF Networks with Dynamic Decay Adjustment 523 Figure 1: This picture shows how a new pattern results in a slightly higher activity for a prototype of the right class than for the conflicting prototype. Using only one threshold, no new prototype would be introduced in this case. where c denotes the class for which the activation is computed, me is the number of prototypes for class c, and R is the constant radius of the gaussian activation functions. The global radius of this method and the inability to recognize areas of conflict, leads to confusion in some areas of the feature space, and therefore non-optimal recognition performance. The Dynamic Decay Adjustment (DDA) algorithm presented in this paper was developed to solve the inherent problems associated with these methods. The constructive part of the P-RCE algorithm is used to build a network with an appropriate number of RBF units, for which the decay factor is computed based on information about neighbours. This technique increases the recognition accuracy in areas of conflict. The following sections explain the algorithm, compare it with others, and examine some simulation results. 2 The Algorithm Since the P-RCE training algorithm already uses an independent area of influence for each RBF, it is relatively straightforward to extract an individual radius. This results, however, in the problem illustrated in figure 1. The new pattern p of class B is properly covered by the right prototype of the same class. However, the left prototype of conflicting class A results in almost the same activation and this leads to a very low confidence when the network must classify the pattern p. To solve this dilemma, two different radii, or thresholds 1 are introduced: a so-called positive threshold (0+), which must be overtaken by an activation of a prototype of the same class so that no new prototype is added, and a negative threshold (0-), which is the upper limit for the activation of conflicting classes. Figure 2 shows an example in which the new pattern correctly results in activations above the positive threshold for the correct class B and below the negative threshold for conflicting class A . This results in better classification-confidence in areas where training IThe conversion from the threshold to the radius is straightforward as long as the activation function is invertible. 524 Michael R. Berthold, Jay Diamond new input pattern (class B) x Figure 2: The proposed algorithm distinguishes between prototypes of correct and conflicting classes and uses different thresholds. Here the level of confidence is higher for the correct classification of the new pattern. patterns did not result in new prototypes. The network is required to hold the following two equations for every pattern x of class c from the training data: 3i : Rf(x) 2:: 8+ Vk :/; c, 1 ~ j ~ mk : Rj(x) (5) < 8- (6) The algorithm to construct a classifier can be extracted partly from the ReE algorithm. The following pseudo code shows what the training for one new pattern x of class c looks like: / / reset weights: FORALL prototypes pf DO Af = 0.0 END FOR / / train one complete epoch FORALL training pattern (x,c) DO: IF 3pi : Ri( x) 2:: 8+ THEN Ai+ = 1.0 ELSE / / "commit": introduce new prototype add new prototype P~c+1 with: ~c+1 =x O'~ +1 = maJ:C {O' : R~ +1 (r7) < 8-} c A~c+1 k#cl\l~J::;mk = 1.0 c mc+= 1 ENDIF / / "shrink": adjust conflicting prototypes FORALL k :/; c, 1 ~ j ~ mk DO O'j = max{O' : Rj(x) < 8-} ENDFOR First, all weights are set to zero because otherwise they would accumulate duplicate information about training patterns. Next all training patterns are presented to the Boos,;ng the Peiformance of RBF Networks with Dynamic Decay Adjustment 525 pIx) pattern class A (1) +2 (2) pIx) pattern class B (3) pattern class A x (4) Figure 3: An example of the DDA- algorithm: (1) a pattern of class A is encountered and a new RBF is created; (2) a training pattern of class B leads to a new prototype for class B and shrinks the radius of the existing RBF of class A; (3) another pattern of class B is classified correctly and shrinks again the prototype of class A; (4) a new pattern of class A introduces another prototype of that class. network. If the new pattern is classified correctly, the weight of the closest prototype is increased; otherwise a new protoype is introduced with the new pattern defining its center. The last step of the algorithm shrinks all prototypes of conflicting classes if their activations are too high for this specific pattern. Running this algorithm over the training data until no further changes are required ensures that equations (5) and (6) hold. The choice of the two new parameters, (J+ and (J- are not as critical as it would initially appear2. For all of the experiments reported, the settings (J+ = 0.4 and (J- = 0.1 were used, and no major correlations of the results to these values were noted. Note that when choosing (J+ = (J- one ends up with an algorithm having the problem mentioned in figure l. Figure 3 shows an example that illustrates the first few training steps of the DDAalgorithm. 3 Results Several well-known databases were chosen to evaluate this algorithm (some can be found in the eMU Neural Network Benchmark Databases (see [13])). The DDA2Theoretically one would expect the dimensionality of the input- space to playa major role for the choice of those parameters 526 Michael R. Berthold, Jay Diamond algorithm was compared against PNN, RCE and P-RCE as well as a classic Multi Layer Perceptron which was trained using a modified Backpropagation algorithm (Rprop, see [9]). The number of hidden nodes of the MLP was optimized manually. In addition an RBF-network with a fixed number of hidden nodes was trained using unsupervised clustering for the center positions and a gradient descent to determine the weights (see [6] for more details). The number of hidden nodes was again optimized manually. ? Vowel Recognition: Speaker independent recognition of the eleven steady state vowels of British English using a specified training set of Linear Predictive Coding (LPC) derived log area ratios (see [3]) resulting in 10 inputs and 11 classes to distinguish. The training set consisted of 528 tokens, with 462 different tokens used to test the network. II performance I #units I #epochs I algorithm Nearest Neighbour 56% 1 ..... 200 MLP (RPRUP) 57% 5 528 PNN 61% ..... 100 RBF 59% 70 27% 125 RCE 3 P-RCE 125 59% 3 DDA-RBF 204 4 65_% ? Sonar Database: Discriminate between sonar signals bounced off a metal cylinder and those bounced off a roughly cylindrical rock (see [4] for more details) . The data has 60 continuous inputs and is separated into two classes. For training and testing 104 samples each were used. algorithm MLP (RPROP) PNN RBF RCE P-RCE DDA-RBF I performance 90.4% 91.3% 90.7% 77.9% 90.4% 93.3% I #units I #epochs I 50 104 80 68 68 68 ..... 250 - ..... 150 3 3 3 ? Two Spirals: This well-known problem is often used to demonstrate the generalization capability of a network (see [5]). The required task involves discriminating between two intertwined spirals. For this paper the spirals were changed slightly to make the problem more demanding. The original spirals radius declines linearly and can be correctly classified by RBF networks with one global radius. To demonstrate the ability of the DDAalgorithm to adjust the radii of each RBF individually, a quadratic decline was chosen for the radius of both spirals (see figure 4) . The training set consisted of 194 points, and the spirals made three complete revolutions. Figure 4 shows both the results of an RBF Network trained with the DDA technique and the same problem solved with a Multi-Layer Perceptron (2-20-20-1) trained using a modified Error Back Propagation algorithm (Rprop, see [9]). Note that in both cases all training points are classified correctly. Boosting the Peifonnance of RBF Networks with Dynamic Decay Adjustment 527 Figure 4: The (quadratic) "two spirals problem" solved by a MLP (left) using Error Back Propagation (after 40000 epochs) and an RBF network (right) trained with the proposed DDA-algorithm (after 4 epochs). Note that all training patterns (indicated by squares vs. crosses) are classified correctly. In addition to these tasks, the BDG-database was used to compare the DDA algorithm to other approaches. This database was used by Waibel et al (see [11]) to introduce the Time Delay Neural Network (TDNN). Previously it has been shown that RBF networks perform equivalently (when using a similar architecture, [1], [2]) with the DDA technique used for training of the RBF units. The BDG task involves distinguishing the three stop consonants "B", "D" and "G". While 783 training sets were used, 749 data sets were used for testing. Each of these contains 15 frames of melscale coefficients, computed from a 10kHz, 12bit converted signal. The final frame frequency was 100Hz. algorithm TDNN TDRBF (P-RCE) TDRBF (DDA) 4 II performance 98.5% 85.2% 98.3% I #epochs I ",50 5 6 Conclusions It has been shown that Radial Basis Function Networks can boost their performance by using the dynamic decay adjustment technique. The algorithm necessary to construct RBF networks based on the RCE method was described and a method to distinguish between conflicting and matching prototypes at the training phase was proposed. An increase in performance was noted, especially in areas of conflict, where standard (P-)RCE did not commit new prototypes. Four different datasets were used to show the performance of the proposed DDAalgorithm. In three of the cases, RBF networks trained with dynamic decay adjustment outperformed known RBF training methods and MLPs. For the fourth task, the BDG-recognition dataset, the TDRBF was able to reach the same level 528 Michael R. Berthold. Jay Diamond of performance as a TDNN. In addition, the new algorithm trains very quickly. Fewer than 6 epochs were sufficient to reach stability for all problems presented. Acknowledgements Thanks go to our supervisors Prof. D. Schmid and Mark Holler for their support and the opportunity to work on this project. References [1] M. R. Berthold: "A Time Delay Radial Basis FUnction Network for Phoneme Recognition" in Proc. of the IEEE International Conference on Neural Networks, 7, p.447D---4473, 1994. [2] M. R. Berthold: "The TDRBF: A Shift Invariant Radial Basis Function Network" in Proc. of the Irish Neural Network Conference, p.7-12, 1994. [3] D. Deterding: "Speaker Normalization for Automatic Speech Recognition" , PhD Thesis, University of Cambridge, 1989. [4] R. Gorman, T. Sejnowski: "Analysis of Hidden Units in a Layered Network Trained to Classify Sonar Targets" in Neural Networks 1, pp.75. [5] K. Lang, M. Witbrock: "Learning to Tell Two Spirals Apart", in Proc. of Connectionist Models Summer School, 1988. [6] J . Moody, C.J. Darken: "Fast Learning in Networks of Locally-Tuned Processing Units" in Neural Computation 1, p.281-294, 1989. [7] M.J. Hudak: "RCE Classifiers: Theory and Practice" in Cybernetics and Systems 23, p.483-515, 1992. [8] D.L. Reilly, L.N. Cooper, C. Elbaum: "A Neural Model for Category Learning" in BioI. Cybernet. 45, p.35-41, 1982. [9] M. Riedmiller, H. Braun: "A Direct Adaptive Method for Faster Backpropagation Learning: The Rprop Algorithm" in Proc. of the IEEE International Conference on Neural Networks, 1, p.586-591 , 1993. [10] D.F. Specht: "Probabilistic Neural Networks" in Neural Networks 3, p.109118,1990. [11] A. Waibel, T. Hanazawa, G. Hinton, K. Shikano, K. Lang: " Phoneme Recognition Using Time-Delay Neural Networks" in IEEE Trans. in Acoustics, Speech and Signal Processing Vol. 37, No. 3, 1989. [12] D. Wettschereck, T. Dietterich: "Improving the Performance of Radial Basis Function Networks by Learning Center Locations" in Advances in Neural Information Processing Systems 4, p.1133- 1140, 1991. [13] S. Fahlman, M. White: "The Carnegie Mellon University Collection of Neural Net Benchmarks" from ftp.cs.cmu .edu in /afs/cs/project/connect/bench.
946 |@word bounced:2 cylindrical:1 simulation:1 contains:1 tuned:3 existing:1 com:1 clara:1 activation:9 lang:2 must:4 eleven:1 v:1 fewer:1 liapunov:1 boosting:3 node:4 location:3 direct:1 introduce:2 roughly:1 examine:1 growing:1 multi:2 inspired:1 pf:1 project:2 underlying:1 what:1 substantially:1 developed:1 elbaum:2 corporation:1 pseudo:1 every:1 charge:1 braun:1 classifier:2 unit:12 before:1 positive:2 limit:1 ree:1 becoming:1 blvd:1 dynamically:1 ease:1 range:1 testing:2 practice:1 implement:1 backpropagation:2 area:7 riedmiller:1 strasse:1 reilly:2 matching:1 confidence:3 radial:8 layered:1 influence:1 charged:1 center:6 peiformance:1 straightforward:2 go:1 stability:2 classic:1 variation:1 analogous:1 target:1 us:3 distinguishing:1 recognition:8 database:6 role:1 solved:2 ensures:1 mentioned:1 und:1 dynamic:8 trained:8 ithe:1 predictive:1 dilemma:1 basis:7 train:3 separated:1 fast:1 sejnowski:1 tell:1 choosing:1 solve:2 otherwise:2 ability:1 commit:3 hanazawa:1 final:1 advantage:1 net:1 rock:1 propose:1 mission:1 reset:1 adaptation:1 ftp:1 nearest:1 school:1 c:2 resemble:1 involves:2 radius:12 correct:3 require:1 generalization:1 d_:1 extension:2 hold:2 exp:1 major:2 proc:4 endfor:1 outperformed:1 individually:2 dda:10 gaussian:4 modified:2 forall:3 derived:2 properly:1 vk:1 contrast:2 dependent:1 initially:1 hidden:6 selective:1 germany:1 overall:1 classification:2 priori:2 overtaken:1 misclassifies:1 construct:3 having:1 ng:1 irish:1 manually:2 identical:1 look:1 unsupervised:2 r7:1 others:1 connectionist:1 inherent:1 duplicate:1 distinguishes:2 few:1 neighbour:3 recognize:1 individual:1 phase:3 bdg:3 vowel:2 attempt:1 cylinder:1 mlp:4 adjust:3 introduces:3 melscale:1 necessary:3 mk:3 increased:1 classify:3 disadvantage:1 witbrock:1 deviation:2 delay:3 supervisor:1 too:1 reported:1 connect:1 thanks:1 international:2 discriminating:1 probabilistic:5 off:2 holler:1 invertible:1 michael:5 together:1 connecting:1 moody:3 quickly:1 earn:1 again:2 thesis:1 possibly:1 potential:1 converted:1 wettschereck:1 coding:1 coefficient:1 start:1 capability:1 rbfs:6 mlps:1 square:1 accuracy:1 acid:1 phoneme:2 modelled:1 informatik:1 mc:1 cybernetics:1 classified:5 explain:1 reach:3 neu:1 email:2 rprop:4 against:1 energy:1 frequency:1 pp:1 associated:1 stop:1 dataset:1 recall:1 dimensionality:1 back:2 higher:2 supervised:2 response:2 shrink:4 until:1 correlation:1 grbf:1 propagation:2 indicated:1 karlsruhe:1 usa:1 dietterich:1 consisted:2 illustrated:1 white:1 during:2 noted:2 speaker:2 steady:1 m:1 generalized:1 complete:2 demonstrate:2 confusion:1 common:1 overview:1 khz:1 accumulate:1 mellon:1 cambridge:1 ai:2 automatic:1 particle:2 electrostatic:1 add:1 playa:1 closest:1 apart:1 determine:1 riw:1 signal:3 ii:6 rj:2 faster:2 af:1 cross:1 long:2 qi:2 cmu:1 normalization:1 pnn:4 addition:4 else:1 induced:1 hz:1 spiral:8 boo:1 architecture:3 topology:1 decline:2 prototype:28 shift:1 speech:2 santa:1 covered:1 referential:1 locally:3 category:1 correctly:6 intertwined:1 carnegie:1 endif:1 vol:1 four:1 threshold:8 pix:2 fourth:1 almost:1 utilizes:1 bit:1 layer:5 summer:1 distinguish:2 quadratic:2 encountered:1 activity:1 ri:3 relatively:1 waibel:2 combination:1 slightly:2 restricted:1 invariant:1 equation:3 previously:1 specht:1 end:2 appropriate:1 coulomb:1 occurrence:2 encounter:1 original:1 denotes:2 clustering:2 running:1 opportunity:1 rce:22 prof:2 build:1 especially:1 already:1 added:1 occurs:1 gradient:3 me:1 code:1 ratio:1 equivalently:1 unfortunately:2 negative:2 diamond:5 perform:2 upper:1 conversion:1 darken:3 datasets:1 benchmark:2 descent:3 defining:1 hinton:1 frame:2 introduced:5 required:3 specified:1 optimized:2 conflict:3 acoustic:1 conflicting:9 emu:1 boost:1 trans:1 able:1 below:1 pattern:25 lpc:1 rf:1 max:1 memory:1 suitable:1 critical:1 demanding:1 hybrid:1 afs:1 scheme:1 picture:1 maj:1 created:1 tdnn:3 extract:1 schmid:2 epoch:7 acknowledgement:1 expect:1 metal:1 sufficient:1 pi:1 changed:1 token:2 last:1 fahlman:1 english:1 haid:1 perceptron:2 berthold:8 computes:1 forward:1 made:1 adaptive:1 collection:1 global:4 forschungszentrum:1 consonant:1 shikano:1 continuous:1 sonar:3 nature:2 ca:1 improving:1 cl:1 did:2 linearly:1 incrementing:1 intel:2 cooper:2 position:1 jay:5 british:1 specific:1 revolution:1 decay:10 effectively:1 phd:1 illustrates:1 gorman:1 adjustment:9 applies:1 determines:1 extracted:1 bioi:1 rbf:26 change:1 determined:2 except:1 called:2 pas:1 partly:1 discriminate:1 indicating:1 college:1 mark:1 support:1 inability:1 constructive:2 evaluate:1 bench:1
7,185
947
A Charge-Based CMOS Parallel Analog Vector Quantizer Gert Cauwenberghs Johns Hopkins University ECE Department 3400 N. Charles St. Baltimore, MD 21218-2686 gert@jhunix.hcf.jhu.edu Volnei Pedroni California Institute of Technology EE Department Mail Code 128-95 Pasadena, CA 91125 pedroni@romeo.caltech.edu Abstract We present an analog VLSI chip for parallel analog vector quantization. The MOSIS 2.0 J..Lm double-poly CMOS Tiny chip contains an array of 16 x 16 charge-based distance estimation cells, implementing a mean absolute difference (MAD) metric operating on a 16-input analog vector field and 16 analog template vectors. The distance cell including dynamic template storage measures 60 x 78 J..Lm2 ? Additionally, the chip features a winner-take-all (WTA) output circuit of linear complexity, with global positive feedback for fast and decisive settling of a single winner output. Experimental results on the complete 16 x 16 VQ system demonstrate correct operation with 34 dB analog input dynamic range and 3 J..Lsec cycle time at 0.7 mW power dissipation. 1 Introduction Vector quantization (VQ) [1] is a common ingredient in signal processing, for applications of pattern recognition and data compression in vision, speech and beyond. Certain neural network models for pattern recognition, such as Kohonen feature map classifiers [2], are closely related to VQ as well. The implementation of VQ, in its basic form, involves a search among a set of vector templates for the one which best matches the input vector, whereby the degree of matching is quantified by a given vector distance metric. Effi- 780 Gert Cauwenberghs, Volnei Pedroni cient hardware implementation requires a parallel search over the template set and a fast selection and encoding of the "winning" template. The chip presented here implements a parallel synchronous analog vector quantizer with 16 analog input vector components and 16 dynamically stored analog template vectors, producing a 4-bit digital output word encoding the winning template upon presentation of an input vector. The architecture is fully scalable as in previous implementations of analog vector quantizers, e.g. [3,4,5,6], and can be readily expanded toward a larger number of vector components and template vectors without structural modification of the layout. Distinct features of the present implementation include a linear winner-take-all (WTA) structure with globalized positive feedback for fast selection of the winning template, and a mean absolute difference (MAD) metric for the distance estimations, both realized with a minimum amount of circuitry. Using a linear charge-based circuit topology for MAD distance accumulation, a wide voltage range for the analog inputs and templates is achieved at relatively low energy consumption per computation cycle. 2 System Architecture The core of the VQ consists of a 16 x 16 2-D array of distance estimation cells, configured to interconnect columns and rows according to the vector input components and template outputs. Each cell computes in parallel the absolute difference distance between one component x} of the input vector x and the corresponding component yi} of one of the template vectors y, (1) The mean absolute difference (MAD) distance between input and template vectors is accumulated along rows A ? d(X,y') = 1 ~ . 16 ~ Ix} - y'}I, i = 1 ... 16 (2) J=l and presented to the WTA, which selects the single winner kWTA = arg min d(x, f) . i (3) Additional parts are included in the architecture for binary encoding of the winning output, and for address selection to write and refresh the template vectors. 3 VLSI Circuit Implementation The circuit implementation of the major components of the VQ, for MAD distance estimation and WTA selection, is described below. Both MAD distance and WTA cells operate in clocked synchronous mode using a precharge/evaluate scheme in the voltage domain. The approach followed here offers a wide analog voltage range of inputs and templates at low power weak-inversion MOS operation, and a fast and decisive settling of the winning output using a single communication line for global positive feedback. The output encoding and address decoding circuitry are implemented using standard CMOS logic. 781 A Charge-Based CMOS Parallel Analog Vector Quantizer VVRi -r--------------------~-----? VVRi 812 Xj Vref PRE -1 4/2 -r----------~----------------Zi (b) (a) Figure 1: Schematic of distance estimation circuitry. Output precharge circuitry. 3.1 (a) Absolute distance cell. (b) Distance Estimation Cell The schematic of the distance estimation cell, replicated along rows and columns of the VQ array, is shown in Figure 1 (a). The cell contains two source followers, which buffer the input voltage x j and the template voltage yi j . The template voltage is stored dynamically onto Cstore, written or refreshed by activating WRi while the y' j value is presented on the Xj input line. The WRi and WRi signal levels along rows of the VQ array are driven by the address decoder, which selects a single template vector yi to be written to with data presented at the input x when WR is active. Additional lateral transistors connect symmetrically to the source follower outputs x / and yi /- By means of resistive division, the lateral transistors construct the maximum and minimum of xl' and yi / on Zi i HI and Zi j LO, respectively. In particular, when x j is . I HI , . LO muc h Iarger than y' j' the vo tage Z' j approaches x j and the voltage Z' j approaches yi /- By symmetry, the complementary argument holds in case Xj is much smaller than yi j. Therefore, the differential compo~ent of Zi j HI and Zi JLO approximately represents the absolute difference value of xJ and y' j : i Z j with K HI - i LO z j , yi j ') - ~ max(xj, - IXj , - Yi j 'I ~ K . mm(xj IXj - ' i ') ,y Yi j j (4) I, the MOS back gate effect coefficient [7] . The mean absolute difference (MAD) distances (2) are obtained by accumulating con- Gert Cauwenberghs, Volnei Pedroni 782 tributions (4) along rows of cells through capacitive coupling, using the well known technique of correlated double sampling. To this purpose, a coupling capacitor Cc is provided in every cell, coupling its differential output to the corresponding output row line. In the precharge phase, the maximum values Zi j HI are coupled to the output by activating HI, and the output lines are preset to reference voltage Vref by activating PRE, Figure 1 (b). In the evaluate phase, PRE is de-activated, and the minimum values Zi j LO are coupled to the output by activating LO. From (4), the resulting voltage outputs on the floating row lines are given by V, re f - 1 L16 16. - . (Zl . J HI - . LO Zl . J ) (5) J=1 1 Vref- K 16 16~ IXj-/jl. J=1 The last term in (5) corresponds directly to the distance measure d(x, yi) in (2). Notice that the negative sign in (5) could be reversed by interchanging clocks HI and LO, if needed. Since the subsequent WTA stage searches for maximum Zi, the inverted distance metric is in the form needed for VQ. Characteristics of the MAD distance estimation (5), measured directly on the VQ array with uniform inputs x j and templates yi j' are shown in Figure 2. The magnified view in Figure 2 (b) clearly illustrates the effective smoothing of the absolute difference function (4) near the origin, x j ~ yi j' The smoothing is caused by the shift in x/and yi j' due to the conductance of the lateral coupling transistors connected to the source follower outputs in Figure 1 (a), and extends over a voltage range comparable to the thermal voltage kT /q depending on the relative geometry of the transistors and current bias level of the source followers. The observed width of the flat region in Figure 2 spans roughly 60 mY, and shows little variation for bias current settings below 0.5 f-I,A. Tuning of the bias current allows to balance speed and power dissipation requirements, since the output response is slew-rate limited by the source followers. 3.2 Winner-Take-All Circuitry The circuit implementation of the winner-take-all (WTA) function combines the compact sizing and modularity of a linear architecture as in [4,8,9] with positive feedback for fast and decisive output settling independent of signal levels, as in [6,3]. Typical positive feedback structures for WTA operation use a logarithmic tree [6] or a fully interconnected network [3], with implementation complexities of order O(n log n) and O(n2) respectively, n being the number of WTA inputs. The present implementation features an O(n) complexity in a linear structure by means of globalized positive feedback, communicated over a single line. The schematic of the WTA cell, receiving the input Zi and constructing the digital output di through global competition communicated over the COMM line, is shown in Figure 3. The global COMM line is source connected to input transistor Mi and positive feedback transistor Mf, and receives a constant bias current lb (WTA) from Mbl. Locally, the WTA operation is governed by the dynamics of d;' on (parasitic) capacitor Cp' A high pulse A Charge-Based CMOS Parallel Analog Vector Quantizer 783 0.0 -0.12 -0.2 ~ :> -0.14 - -0.4 ..... -0.16 ~ ~I -018 ? l ~I -0.6 ' -N -0.20 -0.8 -0.22 -1.0 ~--~----~------~--~ 1.5 2.0 2.5 -0.2 -0.1 0.0 ! Xj (V) Xj-Yj (V) (a) (b) 0.1 0.2 Figure 2: Distance estimation characteristics, (a) for various values of yi j; (b) magnified view. on RST, resetting d/ to zero, marks the beginning of the WTA cycle. With Mf initially inactive, the total bias current n Ib (WTA) through COMM is divided over all competing WTA cells, according to the relative Zi voltage levels, and each cell fraction is locally mirrored by the Mml-Mm2 pair onto d;', charging C p' The cell with the highest Zi input voltage receives the largest fraction of bias current, and charges Cp at the highest rate. The winning output is detennined by the first d;' reaching the threshold to turn on the corresponding Mf feedback transistor, say i = k. This threshold voltage is given by the source voltage on COMM, common for all cells. The positive feedback of the state dk ' through Mf, which eventually claims the entire fraction of the bias current, enhances and latches the winning output level dk ' to the positive supply and shuts off the remaining losing outputs d;' to zero, i :f:. k. The additional circuitry at the output stage of the cell serves to buffer the binary d;' value at the d; output tenninal. No more than one winner can practically co-exist at equilibrium, by nature of the combined positive feedback and global renonnalization in the WTA competition. Moreover, the output settling times of the winner and losers are fairly independent of the input signal levels, and are given mainly by the bias current level Ib (WTA) and the parasitic capacitance Cpo Tests conducted on a separate 16-element WTA array, identical to the one used on the VQ chip, have demonstrated single-winner WTA operation with response time below 0.5 /-Lsec at less than 2 /-L W power dissipation per cell. 4 Functionality Test To characterize the performance of the entire VQ system under typical real-time conditions, the chip was presented a periodic sequence of 16 distinct input vectors x(i), stored and refreshed dynamically in the 16 template locations y; by circularly incrementing the template address and activating WR at the beginning of every cycle. The test vectors rep- 784 Gert Cauwenberghs, Volnei Pedroni Mm2 2212 Zi -1 8/2 MI 11/2 11/2 di T COMM Mb1 VbWfA 11/2 812 RST Figure 3: Circuit schematic of winner-take-all cell. resent a single triangular pattern rotated over the 16 component indices with single index increments in sequence. The fundamental component xo(i) is illustrated on the top trace of the scope plot in Figure 4. The other components are uniformly displaced in time over one period, by a number of cycles equal to the index, x j (i) = xo(i - j mod 16). Figure 4 also displays the VQ output waveforms in response to the triangular input sequence, with the desired parabolic profile for the analog distance output ZO and the expected alternating bit pattern of the WTA least significant output bit. l The triangle test performed correctly at speeds limited by the instrumentation equipment, and the dissipated power on the chip measures 0.7 mW at 3 f.Lsec cycle time2 and 5 V supply voltage. An estimate for the dynamic range of analog input and template voltages was obtained directly by observing the smallest and largest absolute voltage difference still resolved correctly by the VQ output, uniformly over all components. By tuning the voltage range of the triangular test vectors, the recorded minimum and maximum voltage amplitudes for 5 V supply voltage are Ymin = 87.5 mV and YlII3lt = 4 V, respectively. The estimated analog dynamic range YlII3lt /Ymin is thus 45.7, or roughly 34 dB, per cell. The value obtained for Ymin indicates that the dynamic range is limited mainly by the smoothing of the absolute distance measure characteristic (1) near the origin in Figure 2 (b). We notice that a similar limitation of dynamic range applies to other distance metrics with vanishing slope near the origin as well, the popular mean square error (MSE) formulation in particular. The MSE metric is frequently adopted in VQ implementations using strong inversion MOS circuitry, and offers a dynamic range typically worse than obtained here regardless of implementation accuracy, due to the relatively wide flat region of the MSE distance function near the origin. ITbe voltages on the scope plot are inverted as a consequence of the chip test setup. 2including template write operations A "Charge-Based CMOS Parallel Analog Vector Quantizer 785 Figure 4: Scope plot of VQ waveforms. Top: Analog input Xo. Center: Analog distance output zO. Bottom: Least significant bit of encoded output. Table 1: Technology Supply voltage Power dissipation VQchip Dynamic range inputs, templates Area Features of the VQ chip 2 ~ p-well double-poly CMOS +5V 0.7 mW (3 J.lSec cycle time) 34 dB VQchip 2.2 mm X 2.25 mm distance cell WTAcell 6O~X78~ 76~X80~ 786 Gert Cauwenberghs, Volnei Pedroni 5 Conclusion We proposed and demonstrated a synchronous charge-based CMOS VLSI system for parallel analog vector quantization, featuring a mean absolute difference (MAD) metric, and a linear winner-take-all (WTA) structure with globalized positive feedback. By virtue of the MAD metric, a fairly large (34 dB) analog dynamic range of inputs and templates has been obtained in the distance computations through simple charge-based circuitry. Likewise, fast and unambiguous settling of the WTA outputs, using global competition communicated over a single wire, has been obtained by adopting a compact linear circuit structure to implement the positive feedback WTA function. The resulting structure of the VQ chip is highly modular, and the functional characteristics are fairly consistent over a wide range of bias levels, including the MOS weak inversion and subthreshold regions. This allows the circuitry to be tuned to accommodate various speed and power requirements. A summary of the chip features of the 16 x 16 vector quantizer is presented in Table I. Acknowledgments Fabrication of the CMOS chip was provided through the DARPAINSF MOSIS service. The authors thank Amnon Yariv for stimulating discussions and encouragement. References [1] A. Gersho and RM. Gray, Vector Quantization and Signal Compression, Norwell, MA: Kluwer, 1992. [2] T. Kohonen, Self-Organisation and Associative Memory, Berlin: Springer-Verlag, 1984. [3] Y. He and U. Cilingiroglu, "A Charge-Based On-Chip Adaptation Kohonen Neural Network," IEEE Transactions on Neural Networks, vol. 4 (3), pp 462-469, 1993. [4] J.e. Lee, B.J. Sheu, and W.e. Fang, "VLSI Neuroprocessors for Video Motion Detection," IEEE Transactions on Neural Networks, vol. 4 (2), pp 78-191, 1993. [5] R Tawel, "Real-Time Focal-Plane Image Compression," in Proceedings Data Compression Conference" Snowbird, Utah, IEEE Computer Society Press, pp 401-409, 1993. [6] G.T. Tuttle, S. Fallahi, and A.A. Abidi, "An 8b CMOS Vector NO Converter," in ISSCC Technical Digest, IEEE Press, vol. 36, pp 38-39, 1993. [7] C.A. Mead, Analog VLSI and Neural Systems, Reading, MA: Addison-Wesley, 1989. [8] J. Lazzaro, S. Ryckebusch, M.A. Mahowald, and C.A. Mead, "Winner-Take-All Networks of O(n) Complexity," in Advances in Neural Information Processing Systems, San Mateo, CA: Morgan Kaufman, vol. 1, pp 703-711, 1989. [9] A.G. Andreou, K.A. Boahen, P.O. Pouliquen, A. Pavasovic, RE. Jenkins, and K. Strohbehn, "Current-Mode Subthreshold MOS Circuits for Analog VLSI Neural Systems," IEEE Transactions on Neural Networks, vol. 2 (2), pp 205-213, 1991.
947 |@word inversion:3 compression:4 mb1:1 pulse:1 accommodate:1 contains:2 tuned:1 current:9 ixj:3 follower:5 written:2 readily:1 john:1 refresh:1 subsequent:1 resent:1 plot:3 globalized:3 plane:1 beginning:2 vanishing:1 core:1 compo:1 quantizer:6 location:1 along:4 differential:2 supply:4 consists:1 resistive:1 isscc:1 combine:1 expected:1 roughly:2 frequently:1 little:1 provided:2 moreover:1 circuit:8 vref:3 kaufman:1 shuts:1 magnified:2 every:2 charge:10 classifier:1 rm:1 zl:2 producing:1 positive:12 service:1 consequence:1 encoding:4 mead:2 approximately:1 mateo:1 quantified:1 dynamically:3 wri:3 co:1 limited:3 range:13 acknowledgment:1 yj:1 yariv:1 implement:2 communicated:3 area:1 jhu:1 matching:1 word:1 pre:3 onto:2 selection:4 storage:1 accumulating:1 accumulation:1 map:1 demonstrated:2 center:1 layout:1 regardless:1 array:6 fang:1 time2:1 gert:6 variation:1 increment:1 losing:1 origin:4 element:1 recognition:2 tributions:1 observed:1 bottom:1 abidi:1 region:3 cycle:7 connected:2 highest:2 boahen:1 comm:5 complexity:4 dynamic:10 upon:1 division:1 triangle:1 resolved:1 chip:13 various:2 zo:2 distinct:2 fast:6 effective:1 fallahi:1 encoded:1 larger:1 modular:1 say:1 triangular:3 associative:1 sequence:3 transistor:7 interconnected:1 adaptation:1 kohonen:3 loser:1 detennined:1 competition:3 ent:1 rst:2 double:3 requirement:2 cmos:10 rotated:1 coupling:4 depending:1 snowbird:1 measured:1 strong:1 implemented:1 involves:1 waveform:2 closely:1 correct:1 functionality:1 implementing:1 activating:5 sizing:1 hold:1 mm:3 practically:1 equilibrium:1 scope:3 mo:5 lm:1 circuitry:9 claim:1 major:1 smallest:1 purpose:1 estimation:9 largest:2 clearly:1 reaching:1 voltage:23 romeo:1 indicates:1 mainly:2 equipment:1 interconnect:1 accumulated:1 entire:2 typically:1 initially:1 pasadena:1 vlsi:6 selects:2 arg:1 among:1 smoothing:3 fairly:3 field:1 construct:1 equal:1 sampling:1 mm2:2 represents:1 identical:1 interchanging:1 floating:1 phase:2 geometry:1 conductance:1 detection:1 highly:1 activated:1 kt:1 norwell:1 tree:1 x80:1 re:2 desired:1 column:2 tuttle:1 mahowald:1 uniform:1 fabrication:1 conducted:1 kwta:1 characterize:1 stored:3 connect:1 periodic:1 my:1 combined:1 st:1 muc:1 fundamental:1 lee:1 off:1 receiving:1 decoding:1 hopkins:1 recorded:1 worse:1 de:1 coefficient:1 configured:1 caused:1 mv:1 decisive:3 performed:1 view:2 observing:1 cauwenberghs:5 parallel:9 slope:1 square:1 accuracy:1 characteristic:4 resetting:1 likewise:1 subthreshold:2 weak:2 tenninal:1 cc:1 energy:1 pp:6 refreshed:2 di:2 mi:2 con:1 popular:1 amplitude:1 back:1 wesley:1 response:3 formulation:1 stage:2 clock:1 receives:2 mode:2 gray:1 utah:1 effect:1 alternating:1 illustrated:1 latch:1 width:1 self:1 unambiguous:1 whereby:1 clocked:1 complete:1 demonstrate:1 vo:1 dissipation:4 cp:2 motion:1 image:1 charles:1 common:2 functional:1 winner:12 sheu:1 jl:1 analog:24 he:1 kluwer:1 significant:2 encouragement:1 tuning:2 focal:1 operating:1 driven:1 instrumentation:1 certain:1 buffer:2 slew:1 verlag:1 binary:2 rep:1 yi:15 caltech:1 inverted:2 morgan:1 minimum:4 additional:3 period:1 signal:5 technical:1 match:1 offer:2 divided:1 schematic:4 scalable:1 basic:1 vision:1 metric:8 pavasovic:1 adopting:1 achieved:1 cell:21 baltimore:1 source:7 operate:1 db:4 capacitor:2 mod:1 ee:1 mw:3 structural:1 symmetrically:1 near:4 xj:8 zi:12 architecture:4 topology:1 competing:1 converter:1 shift:1 synchronous:3 inactive:1 amnon:1 speech:1 lazzaro:1 amount:1 locally:2 hardware:1 l16:1 exist:1 mirrored:1 notice:2 sign:1 estimated:1 per:3 wr:2 correctly:2 write:2 vol:5 threshold:2 mosis:2 fraction:3 extends:1 parabolic:1 comparable:1 bit:4 hi:8 followed:1 display:1 flat:2 lsec:4 speed:3 argument:1 min:1 span:1 expanded:1 relatively:2 department:2 according:2 smaller:1 wta:23 modification:1 xo:3 vq:18 pouliquen:1 turn:1 eventually:1 needed:2 mml:1 addison:1 gersho:1 serf:1 lm2:1 adopted:1 operation:6 jenkins:1 hcf:1 gate:1 capacitive:1 top:2 remaining:1 include:1 effi:1 society:1 capacitance:1 realized:1 digest:1 ryckebusch:1 md:1 enhances:1 distance:26 reversed:1 separate:1 lateral:3 thank:1 decoder:1 berlin:1 consumption:1 mail:1 tage:1 toward:1 mad:10 code:1 index:3 balance:1 setup:1 quantizers:1 trace:1 negative:1 implementation:11 wire:1 displaced:1 thermal:1 communication:1 lb:1 jhunix:1 pair:1 andreou:1 california:1 address:4 beyond:1 renonnalization:1 below:3 pattern:4 reading:1 including:3 max:1 memory:1 video:1 charging:1 power:7 settling:5 scheme:1 technology:2 ymin:3 dissipated:1 coupled:2 relative:2 cpo:1 fully:2 limitation:1 pedroni:6 ingredient:1 digital:2 degree:1 mbl:1 consistent:1 tiny:1 row:7 lo:7 featuring:1 summary:1 last:1 bias:9 institute:1 wide:4 template:25 absolute:11 feedback:12 computes:1 author:1 replicated:1 san:1 transaction:3 compact:2 logic:1 global:6 active:1 search:3 modularity:1 table:2 additionally:1 nature:1 ca:2 symmetry:1 mse:3 poly:2 constructing:1 domain:1 incrementing:1 profile:1 n2:1 complementary:1 cient:1 winning:7 xl:1 governed:1 ib:2 ix:1 precharge:3 dk:2 virtue:1 organisation:1 quantization:4 cstore:1 circularly:1 illustrates:1 mf:4 logarithmic:1 strohbehn:1 applies:1 springer:1 corresponds:1 ma:2 stimulating:1 presentation:1 included:1 typical:2 uniformly:2 preset:1 tawel:1 total:1 ece:1 experimental:1 jlo:1 parasitic:2 mark:1 evaluate:2 correlated:1
7,186
948
Computational structure of coordinate transformations: A generalization study Zoubin Ghahramani zoubin@psyche.mit.edu Daniel M. Wolpert wolpert@psyche.mit.edu Michael I. Jordan jordan@psyche.mit.edu Department of Brain & Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 Abstract One of the fundamental properties that both neural networks and the central nervous system share is the ability to learn and generalize from examples. While this property has been studied extensively in the neural network literature it has not been thoroughly explored in human perceptual and motor learning. We have chosen a coordinate transformation system-the visuomotor map which transforms visual coordinates into motor coordinates-to study the generalization effects of learning new input-output pairs. Using a paradigm of computer controlled altered visual feedback, we have studied the generalization of the visuomotor map subsequent to both local and context-dependent remappings. A local remapping of one or two input-output pairs induced a significant global, yet decaying, change in the visuomotor map, suggesting a representation for the map composed of units with large functional receptive fields. Our study of context-dependent remappings indicated that a single point in visual space can be mapped to two different finger locations depending on a context variable-the starting point of the movement. Furthermore, as the context is varied there is a gradual shift between the two remappings, consistent with two visuomotor modules being learned and gated smoothly with the context. 1 Introduction The human central nervous system (CNS) receives sensory inputs from a multitude of modalities, each tuned to extract different forms of information from the 1126 Zoubin Ghahramani, Daniel M. Wolpert, Michael 1. Jordan environment. These sensory signals are initially represented in disparate coordinate systems, for example visual information is represented retinotopically whereas auditory information is represented tonotopically. The ability to transform information between coordinate systems is necessary for both perception and action. When we reach to a visually perceived object in space, for example, the location of the object in visual coordinates must be converted into a representation appropriate for movement, such as the configuration of the arm required to reach the object. The computational structure of this coordinate transformation, known as the visuomotor map, is the focus of this paper. By examining the change in visuomotor coordination under prismatically induced displacement and rotation, Helmholtz (1867/1925) and Stratton (1897a,1897b) pioneered the systematic study of the representation and plasticity of the visuomotor map. Their studies demonstrate both the fine balance between the visual and motor coordinate systems, which is disrupted by such perturbations, and the ability of the visuomotor map to adapt to the displacements induced by the prisms. Subsequently, many studies have further demonstrated the remarkable plasticity of the map in response to a wide variety of alterations in the relationship between the visual and motor system (for reviews see Howard, 1982 and Welch, 1986)-the single prerequisite for adaptation seems to be that the remapping be stable (Welch, 1986). Much less is known, however, about the topological properties of this map. A coordinate transformation such as the visuomotor map can be regarded as a function relating one set of variables (inputs) to another (outputs). For the visuomotor map the inputs are visual coordinates of a desired target and the outputs are the corresponding motor coordinates representing the arm's configuration (e.g. joint angles). The problem of learning a sensorimotor remapping can then be regarded as a function approximation problem. Using the theory of function approximation one can make explicit the correspondence between the representation used and the patterns of generalization that will emerge. Function approximators can predict patterns of generalization ranging from local (look-up tables), through intermediate (CMACs, Albus, 1975; and radial basis functions, Moody and Darken, 1989 ) to global (parametric models). In this paper we examine the representational structure of the visuomotor map through the study of its spatial and contextual generalization properties. In the spatial generalization study we address the question of how pointing changes over the reaching workspace after exposure to a highly localized remapping. Previous work on spatial generalization, in a study restricted to one dimension, has led to the conclusion that the visuomotor map is constrained to generalize linearly (Bedford, 1989). We test this conclusion by mapping out the pattern of generalization induced by one and two remapped points in two dimensions. In the contextual generalization study we examine the question of whether a single point in visual space can be mapped into two different finger locations depending on the context of a movement-the start point. If this context-dependent remapping occurs, the question arises as to how the mapping will generalize as the context is varied. Studies of contextual remapping have previously shown that variables such as eye position (Kohler, 1950; Hay and Pick, 1966; Shelhamer et al., 1991), the feel of prisms (Kravitz, 1972; Welch, 1971) or an auditory tone (Kravitz and Yaffe, 1972), can induce context-dependent aftereffects. The question of how these context- 1127 Computational Structure of Coordinate Transformations dependent maps generalize-which has not been previously explored-reflects on the possible representation of multiple visuomotor maps and their mixing with a context variable. 2 Spatial Generalization To examine the spatial generalization of the visuomotor map we measured the change in pointing behavior subsequent to one- and two-point remappings. In order to measure pointing behavior and to confine subjects to learn limited input-output pairs we used a virtual visual feedback setup consisting of a digitizing tablet to record the finger position on-line and a projection/mirror system to generate a cursor spot image representing the finger position (Figure 1a). By controlling the presence of the cursor spot and its relationship to the finger position, we could both restrict visual feedback of finger position to localized regions of space and introduce perturbations of this feedback. b) a) VGAScreen PnljecIor P_ Ii \ \ /, /0: \ /, /, / ,/ ,l 0 0 0 Finger Position ~~ . Angerf_ o image \ ;(..'.. ~"'" . .. \ \) Rear profection acreen -, ' >'"'!'f'----Vi-rtual-lma-II"Seml-sliverad mirTOr Dl!Jtizjng Tablel Actual Anger Position Actual FIngerPosI1Ion .......... o a 150m '.. -" C)[J' ....., , ? . . d)Q'. ? 1 ? T ? ? e)lZj' " ,t ,, , , , Figure 1. a) Experimental setup. The subjects view the reflected image of the rear projection screen by looking down at the mirror. By matching the screenmirror distance to the mirror-tablet distance all projected images appeared to be in the plane of the finger (when viewed in the mirror) independent of head position. b) The position of the grid of targets relative to the subject. Also shown, for the x-shift condition, is the perceived and actual finger position when pointing to the central training target. The visually perceived finger position is indicated by a cursor spot which is displaced from the actual finger position. c) A schematic showing the perturbation for the x-shift group. To see the cursor spot on the central target the subjects had to place their finger at the position indicated by the tip of the arrow. d) & e) Schematics similar to c) showing the perturbation for the y-shift and two point groups, respectively. In the tradition of adaptation studies (e.g. Welch, 1986), each experimental session consisted of three phases: pre-exposure, exposure, and post-exposure. During the pre- and post-exposure phases, designed to assess the visuomotor map, the subject pointed repeatedly, without visual feedback of his finger position, to a grid of targets over the workspace. As no visual input of finger location was given, no learning of the visuomotor map could occur . During the exposure phase subjects pointed repeatedly to one or two visual target locations, at which we introduced a discrep- 1128 Zoubin Ghahramani, Daniel M. Wolpert, Michael!. Jordan ancy between the actual and visually displayed finger location. No visual feedback of finger position was given except when within 0.5 cm of the target, thereby confining any learning to the chosen input-output pairs. Three local perturbations of the visuomotor map were examined: a 10 cm rightward displacement (x-shift group, Figure lc), 10 cm displacement towards the body (y-shift group, Figure Id), and a displacement at two points, one 10 cm away from, and one 10 cm towards the body (two point group , Figure Ie). For example, for the x-shift displacement the subject had to place his finger 10 cm to the right of the target to visually perceive his finger as being on target (Figure Ib). Separate control subjects, in which the relationship between the actual and visually displayed finger position was left unaltered, were run for both the one- and two-point displacements, resulting in a total of 5 groups with 8 subjects each. 50 -?) 45 .. ,. 40 50 __ .......................... " 4' __ .......................... ..... 40 _ _ ............ _ ..';1.' ] JS >30 30 ........ _ _ _ ............ _ _ ........ --..... ~~ .............. 25 20 1~ 10 45 Cl CJ 41) ~ 45 \ 40 \ I \ 25 t9 20 20 -10 -5 10 15 -10 20 10 ".'i 55 55 50 5(J , , , 45 45 " e 35 ~ >- 30 E- I 40 0. '0 15 20 -10 () x (em) X (I.:m ) c X (em) 35 >- 30 25 20 X (Lm) X ?(,:m) ~ 30 25 25 I' I \ t t 2() t !() 2() t tt tit J f I 2(J G 20 ~ ~ , 10 X (em) I I I I 15 15 -15 ? 10 -5 0 5 10 x (em) 15 20 2.' -15 -10?5 I) 5 11) X (em) 15 211 25 -10 () X (em) Figure 2. Results of the spatial generalization study. The first column shows the mean change in pointing, along with 95% confidence ellipses, for the x-shift, y-shift and two point groups. The second column displays a vector field of changes obtained from the data by Gaussian kernel smoothing. The third column plots the proportion adaptation in the direction of the perturbation. Note that whereas for the x- and yshift groups the lighter shading corresponds to greater adaptation, for the two point group lighter shades correspond to adaptation in the positive y direction and darker shades in the negative y direction. Computational Structure of Coordinate Transfonnations 1129 The patterns of spatial generalization subsequent to exposure to the three local remappings are shown in Figure 2. All three perturbation groups displayed both significant adaptation at the trained points, and significant, through decremented, generalization of this learning to other targets. As expected, the control groups (not shown) did not show any significant changes. The extent of spatial generalization, best depicted by the shaded contour plots in Figure 2, shows a pattern of generalization that decreases with distance away from the trained points. Rather than inducing a single global change in the map, such as a rotation or shear, the two point exposure appears to induce two opposite fields of decaying generalization at the intersection of which there is no change in the visuomotor map. 3 Contextual Generalization The goal of this experiment was first to explore the possibility that multiple visuomotor maps, or modules, could be learned, and if so, to determine how the overall system behaves as the context used in training each module is varied. To achieve this goal, we exposed subjects to context-dependent remappings in which a single visual target location was mapped to two different finger positions depending on the start point of the movement. Pointing to the target from seven different starting points (Figure 3) was assessed before and after an exposure phase. During this exposure phase subjects made repeated movements to the target from starting points 2 and 6 with a different perturbation of the visual feedback depending on the starting point . The form of these context-dependent remappings is shown in Figure 3. For example, for the open x-shift group (Figure 3c), the visual feedback of the finger was displaced to the right for movements from point 2 and to the left from point 6. Therefore the same visual target was mapped to two different finger positions depending on the context of the movement. To test learning of the remapping and generalization to other start points we examined the change in pointing, without visual feedback, to the target from the 7 start points. b) crossed x-shift a) control ~' .... ~ ....'e'- ?t -'? -' t, o 1234567 c) open x-shift ..' ." ........... ~ ...?.?. ......? dlYL\ 1234567 :,., .... o o 1234567 1234567 0 Figure 3. Schematic of the exposure phase in the contextual generalization experiment. Shown are the actual finger path (solid line), the visually displayed finger path (dotted line), the seven start points and the target used in the pre- and postexposure phases. The perturbation introduced depended on whether the movement started form start point 2 or 6. Note that for the three perturbation groups, although the subjects saw a triangle being traced out, the finger took a different path. 1130 Zoubin Ghahramani, Daniel M. Wolpert, Michael I. Jordan The results are shown in Figure 4. Whereas the controls did not show any significant pattern of change, the three other groups showed adaptive, start point dependent, changes in the direction opposite to the perturbation. Thus, for example, the xopen group displayed a pattern of change in the leftward (negative x) direction for movements from the left start points and rightwards for movements from the right start points. Furthermore, as the start point was varied, the change in pointing varied gradually. a) U a control 0--0 croned x?,bltt - - 0.' b) 2.0 a 1.0 !!. !!. " .;J " .S! ;; 0 -0.' ..... >< c. ." -u ....... c. ~v~ -2.' ." , 4 Start point 0.0 , -1.0 2 4 Start point Figure 4. a) Adaptation in the x direction plotted as a function of starting point for the control, crossed x-shift and open x-shift groups (mean and 1 s.e.). b) Adaptation in the y direction for the control and y-shift groups. 4 Discussion Clearly, from the perspective of function approximation theory, the problem of relearning the visuomotor mapping from exposure to one or two input-output pairs is ill-posed. The mapping learned, as measured by the pattern of generalization to novel inputs, therefore reflects intrinsic constraints on the internal representation used. The results from the spatial generalization study suggest that the visuomotor coordinate transformation is internally represented with units with large but localized receptive fields. For example, a neural network model with Gaussian radial basis function units (Moody and Darken, 1989), which can be derived by assuming that the internal constraint in the visuomotor system is a smoothness constraint (Poggio and Girosi, 1989), predicts a pattern of generalization very similar the one experimentally observed (e.g. see Figure 5 for a simulation of the two point generalization experiment).1 In contrast, previously proposed models for the representation of the visuomotor map based on global parametric representations in terms of felt direction of gaze and head position (e.g. Harris, 1965) or linear constraints (Bedford, 1989) do not predict the decaying patterns of Cartesian generalization found. 1 See also Pouget & Sejnowski (this volume) who, based on a related analysis of neurophysiological data from parietal cortex, suggest that a basis function representation may be used in this visuomotor area. Computational Structure of Coordinate Transformations ~ >- 50 , \ 45 \ \ 40 \ \ \ 35 30 25 \ \ ~ \ \ t t ~! I ~ ? 10 ?5 0 X (em) t }~e t t I I I' I' 10 15 20 20 1131 ~ >- I X (em) Figure 5. Simulation of the two point spatial generalization experiment using a radial basis function network with 64 units with 5 cm Gaussian receptive fields. The inputs to the network were the visual coordinates of the target and the outputs were the joint angles for a two-link planar arm to reach the target. The network was first trained to point accurately to the targets , and then, after exposure to the perturbation, its pattern of generalization was assessed. The results from the second study suggest that multiple visuomotor maps can be learned and modulated by a context. A suggestive computational model for how such separate modules can be learned and combined is the mixture-of-experts neural network architecture (Jacobs et al. , 1991). Interpreted in this framework, the gradual effect of varying the context seen in Figure 4 could reflect the output of a gating network which uses context to modulate between two visuomotor maps. However, our results do not rule out models in which a single visuomotor map is parametrized by starting location, such as one based on the coding of locations via movement vectors (Georgopoulos, 1990) . 5 Conclusions The goal of these studies has been to infer the internal constraints in the visuomotor system through the study of its patterns of generalization to local remappings. We have found that local perturbations of the visuomotor map produce global changes, suggesting a distributed representation with large receptive fields. Furthermore, context-dependent perturbations induce changes in pointing consistent with a model of visuomotor learning in which separate maps are learned and gated by the context. The approach taken in this paper provides a strong link between neural network theory and the study of learning in biological systems. Acknowledgements This project was supported in part by a grant from the McDonnell-Pew Foundation , by a grant from ATR Human Information Processing Research Laboratories, by a grant from Siemens Corporation, and by grant N00014-94-1-0777 from the Office of Naval Research. Zoubin Ghahramani and Daniel M. Wolpert are McDonnell-Pew Fellows in Cognitive Neuroscience. Michael I. Jordan is a NSF Presidential Young Investigator. 1132 Zoubin Ghahramani, Daniel M. Wolpert, Michael I. Jordan References Albus, J. (1975) . A new approach to manipulator control: The cerebellar model articulation controller (CMAC). J. of Dynamic Systems, Measurement, and Control, 97:220-227. Bedford, F. (1989). Constraints on learning new mappings between perceptual dimensions. J. of Experimental Psychology: Human Perception and Performance, 15(2):232-248. Georgopoulos, A. (1990). Neurophysiology of reaching. In Jeannerod, M., editor, Attention and performance XIII, pages 227-263 . Lawrence Erlbaum, Hillsdale. Harris, C. (1965). Perceptual adaptation to inverted, reversed, and displaced vision. Psychological Review, 72:419-444. Hay, J. and Pick, H. (1966). Gaze-contingent prism adaptation: Optical and motor factors. J. of Experimental Psychology, 72 :640-648. Howard, 1. (1982). Human visual orientation. Wiley, Chichester, England. Jacobs, R., Jordan, M. , Nowlan, S. , and Hinton, G. (1991) . Adaptive mixture of local experts. Neural Computation, 3:79-87. Kohler, 1. (1950). Development and alterations of the perceptual world: conditioned sensations. Proceedings of the Austrian Academy of Sciences, 227. Kravitz, J. (1972). Conditioned adaptation to prismatic displacement. Perception and Psychophysics, 11:38-42. Kravitz, J . and Yaffe, F . (1972) . Conditioned adaptation to prismatic displacement with a tone as the conditional stimulus. Perception and Psychophysics, 12:305-308. Moody, J. and Darken, C. (1989) . Fast learning in networks of locally-tuned processing units. Neural Computation, 1(2) :281-294. Poggio, T. and Girosi, F. (1989). A theory of networks for approximation and learning. AI Lab Memo 1140, MIT. Pouget, A. and Sejnowski, T. (1994) . Spatial representation in the parietal cortex may use basis functions. In Tesauro, G., Touretzky, D. , and Alspector, J., editors, Advances in Neural Information Processing Systems 7. Morgan Kaufmann. Shelhamer, M., Robinson, D., and Tan, H. (1991) . Context-specific gain switching in the human vestibuloocular reflex. In Cohen, B., Tomko, D., and Guedry, F., editors, Annals Of The New York Academy Of Sciences, volume 656, pages 889-891. New York Academy Of Sciences, New York. Stratton, G. (1897a). Upright vision and the retinal image. Psychological Review, 4:182187. Stra.tton, G. (1897b) . Vision without inversion of the retinal image. Psychological Review, 4:341-360, 463-481. von Helmholtz, H. (1925). Treatise on physiological optics (1867). Optical Society of America, Rochester, New York. Welch, R. (1971). Discriminative conditioning of prism adaptation. Perception and Psychophysics, 10:90-92 . Welch, R. (1986). Adaptation to space perception. In Boff, K., Kaufman, L. , and Thomas, J ., editors, Handbook of perception and performance, volume 1, pages 24- 1-24-45. Wiley-Interscience, New York.
948 |@word neurophysiology:1 unaltered:1 inversion:1 seems:1 proportion:1 open:3 gradual:2 simulation:2 jacob:2 pick:2 thereby:1 solid:1 shading:1 configuration:2 daniel:6 tuned:2 contextual:5 nowlan:1 yet:1 must:1 subsequent:3 plasticity:2 girosi:2 motor:6 designed:1 plot:2 nervous:2 tone:2 plane:1 record:1 provides:1 location:9 along:1 interscience:1 introduce:1 expected:1 alspector:1 behavior:2 examine:3 brain:1 actual:7 project:1 remapping:7 cm:7 interpreted:1 kaufman:1 transformation:7 corporation:1 fellow:1 transfonnations:1 control:9 unit:5 internally:1 grant:4 positive:1 before:1 local:8 depended:1 switching:1 id:1 path:3 studied:2 examined:2 shaded:1 limited:1 p_:1 spot:4 displacement:9 cmac:1 area:1 projection:2 matching:1 pre:3 radial:3 induce:3 confidence:1 zoubin:7 suggest:3 context:21 map:28 demonstrated:1 exposure:13 attention:1 starting:6 welch:6 perceive:1 pouget:2 rule:1 regarded:2 his:3 coordinate:17 feel:1 annals:1 target:19 controlling:1 tan:1 pioneered:1 lighter:2 us:1 tablet:2 helmholtz:2 predicts:1 observed:1 module:4 region:1 remappings:8 movement:11 decrease:1 environment:1 dynamic:1 trained:3 tit:1 exposed:1 basis:5 triangle:1 rightward:1 joint:2 represented:4 america:1 finger:25 fast:1 sejnowski:2 visuomotor:30 posed:1 presidential:1 ability:3 transform:1 took:1 adaptation:14 stra:1 mixing:1 achieve:1 representational:1 academy:3 albus:2 boff:1 inducing:1 produce:1 object:3 depending:5 measured:2 strong:1 direction:8 sensation:1 subsequently:1 human:6 virtual:1 hillsdale:1 generalization:29 biological:1 lma:1 confine:1 visually:6 lawrence:1 mapping:5 predict:2 lm:1 pointing:9 perceived:3 coordination:1 saw:1 reflects:2 mit:4 clearly:1 gaussian:3 reaching:2 rather:1 vestibuloocular:1 varying:1 office:1 derived:1 focus:1 naval:1 contrast:1 tradition:1 dependent:9 rear:2 initially:1 overall:1 ill:1 orientation:1 development:1 spatial:11 constrained:1 smoothing:1 psychophysics:3 field:6 look:1 anger:1 decremented:1 stimulus:1 xiii:1 composed:1 phase:7 consisting:1 cns:1 highly:1 possibility:1 chichester:1 mixture:2 necessary:1 poggio:2 desired:1 plotted:1 psychological:3 column:3 bedford:3 examining:1 erlbaum:1 lzj:1 combined:1 thoroughly:1 disrupted:1 fundamental:1 ie:1 workspace:2 systematic:1 michael:6 tip:1 gaze:2 moody:3 von:1 central:4 reflect:1 cognitive:2 expert:2 suggesting:2 converted:1 retinal:2 alteration:2 coding:1 vi:1 crossed:2 view:1 lab:1 start:12 decaying:3 rochester:1 ass:1 kaufmann:1 who:1 correspond:1 generalize:4 accurately:1 reach:3 touretzky:1 sensorimotor:1 gain:1 auditory:2 massachusetts:1 cj:1 appears:1 reflected:1 response:1 planar:1 furthermore:3 receives:1 indicated:3 manipulator:1 effect:2 consisted:1 laboratory:1 during:3 tt:1 demonstrate:1 ranging:1 image:6 novel:1 rotation:2 behaves:1 shear:1 functional:1 cohen:1 retinotopically:1 conditioning:1 volume:3 digitizing:1 relating:1 significant:5 measurement:1 cambridge:1 ai:1 smoothness:1 pew:2 grid:2 session:1 pointed:2 had:2 stable:1 treatise:1 cortex:2 j:1 showed:1 leftward:1 perspective:1 tesauro:1 hay:2 n00014:1 prism:4 approximators:1 inverted:1 seen:1 morgan:1 greater:1 contingent:1 determine:1 paradigm:1 signal:1 ii:2 multiple:3 infer:1 england:1 adapt:1 post:2 ellipsis:1 controlled:1 schematic:3 controller:1 vision:3 austrian:1 kernel:1 cerebellar:1 whereas:3 fine:1 modality:1 induced:4 subject:12 jeannerod:1 jordan:8 presence:1 intermediate:1 variety:1 psychology:2 architecture:1 restrict:1 opposite:2 shift:15 whether:2 york:5 action:1 repeatedly:2 transforms:1 extensively:1 locally:1 generate:1 nsf:1 dotted:1 neuroscience:1 group:17 traced:1 tonotopically:1 prismatic:2 run:1 angle:2 place:2 display:1 correspondence:1 topological:1 occur:1 optic:1 constraint:6 georgopoulos:2 felt:1 optical:2 department:1 mcdonnell:2 em:8 psyche:3 restricted:1 gradually:1 taken:1 aftereffect:1 previously:3 prerequisite:1 away:2 appropriate:1 thomas:1 ghahramani:6 society:1 question:4 occurs:1 receptive:4 parametric:2 distance:3 separate:3 mapped:4 link:2 atr:1 parametrized:1 reversed:1 seven:2 extent:1 assuming:1 relationship:3 balance:1 setup:2 negative:2 disparate:1 memo:1 gated:2 confining:1 darken:3 displaced:3 howard:2 displayed:5 parietal:2 hinton:1 looking:1 head:2 varied:5 perturbation:14 introduced:2 pair:5 required:1 learned:6 robinson:1 address:1 perception:7 pattern:12 articulation:1 remapped:1 appeared:1 arm:3 representing:2 altered:1 technology:1 eye:1 started:1 extract:1 review:4 literature:1 acknowledgement:1 relative:1 remarkable:1 localized:3 shelhamer:2 foundation:1 consistent:2 kravitz:4 editor:4 share:1 supported:1 institute:1 wide:1 emerge:1 distributed:1 feedback:9 dimension:3 world:1 contour:1 sensory:2 made:1 adaptive:2 projected:1 tomko:1 global:5 suggestive:1 handbook:1 discriminative:1 table:1 learn:2 cl:1 did:2 linearly:1 arrow:1 repeated:1 body:2 screen:1 darker:1 wiley:2 lc:1 position:19 explicit:1 perceptual:4 ib:1 third:1 young:1 down:1 shade:2 specific:1 showing:2 gating:1 explored:2 physiological:1 multitude:1 t9:1 dl:1 intrinsic:1 mirror:4 conditioned:3 cartesian:1 cursor:4 relearning:1 wolpert:7 smoothly:1 led:1 depicted:1 intersection:1 explore:1 neurophysiological:1 visual:22 reflex:1 corresponds:1 harris:2 ma:1 conditional:1 modulate:1 viewed:1 cmacs:1 goal:3 towards:2 change:16 experimentally:1 upright:1 except:1 total:1 experimental:4 siemens:1 internal:3 arises:1 assessed:2 modulated:1 investigator:1 kohler:2 rightwards:1
7,187
949
Inferring Ground Truth from Subjective Labelling of Venus Images Padhraic Smyth, Usama Fayyad Jet Propulsion Laboratory 525-3660, Caltech, 4800 Oak Grove Drive, Pasadena, CA 91109 Michael Burl, Pietro Perona Department of Electrical Engineering Caltech, MS 116-81, Pasadena, CA 91125 Pierre Baldi* Jet Propulsion Laboratory 303-310, Caltech, 4800 Oak Grove Drive, Pasadena, CA 91109 Abstract In remote sensing applications "ground-truth" data is often used as the basis for training pattern recognition algorithms to generate thematic maps or to detect objects of interest. In practical situations, experts may visually examine the images and provide a subjective noisy estimate of the truth. Calibrating the reliability and bias of expert labellers is a non-trivial problem. In this paper we discuss some of our recent work on this topic in the context of detecting small volcanoes in Magellan SAR images of Venus. Empirical results (using the Expectation-Maximization procedure) suggest that accounting for subjective noise can be quite significant in terms of quantifying both human and algorithm detection performance. 1 Introduction In certain pattern recognition applications, particularly in remote-sensing and medical diagnosis, the standard assumption that the labelling of the data has been * and Division of Biology, California Institute of Technology 1086 Padhraic Smyth. Usama Fayyad. Michael Burl. Pietro Perona. Pierre Baldi carried out in a reasonably objective and reliable manner may not be appropriate. Instead of "ground truth" one may only have the subjective opinion(s) of one or more experts. For example, medical data or image data may be collected off-line and some time later a set of experts analyze the data and produce a set of class labels. The central problem is that of trying to infer the "ground truth" given the noisy subjective estimates of the experts. When one wishes to apply a supervised learning algorithm to the data, the problem is primarily twofold: (i) how to evaluate the relative performance of experts and algorithms, and (ii) how to train a pattern recognition system in the absence of absolute ground truth. In this paper we focus on problem (i), namely the performance evaluation issue, and in particular we discuss the application of a particular modelling technique to the problem of counting volcanoes on the surface of Venus. For problem (ii), in previous work we have shown that when the inferred labels have a probabilistic interpretation, a simple mixture model argument leads to straightforward modifications of various learning algorithms [1]. It should be noted that the issue of inferring ground truth from subjective labels has appeared in the literature under various guises. French [2] provides a Bayesian perspective on the problem of combining multiple opinions. In the field of medical diagnosis there is a significant body of work on latent variable models for inferring hidden "truth" from subjective diagnoses (e.g., see Uebersax [3]). More abstract theoretical models have also been developed under assumptions of specific labelling patterns (e.g., Lugosi [4] and references therein). The contribution of this paper is twofold: (i) this is the first application of latent-variable subjective-rating models to a large-scale image analysis problem as far as we are aware, and (ii) the focus of our work is on the pattern recognition aspect of the problem, i.e., comparing human and algorithmic performance as opposed to simply comparing humans to each other. 2 Background: Automated Detection of Volcanoes in Radar Images of Venus Although modern remote-sensing and sky-telescope technology has made rapid recent advances in terms of data collection capabilities, image analysis often remains a strictly manual process and much investigative work is carried out using hardcopy photographs. The Magellan Venus data set is a typical example: between 1991 and 1994 the Magellan spacecraft transmitted back to earth a data set consisting of over 30,000 high resolution (75m per pixel) synthetic aperture radar (SAR) images of the Venusian surface [5]. This data set is greater than that gathered by all previous planetary missions combined - planetary scientists are literally swamped by data. There are estimated to be on the order of 106 small (less than 15km in diameter) vl.sible volcanoes scattered throughout the 30,000 images [6]. It has been estimated that manually locating all of these volcanoes would require on the order of 10 man-years of a planetary geologist's time to carry out - our experience has been that even a few hours of image analysis severely taxes the concentration abilities of human labellers. From a scientific viewpoint the ability to accurately locate and characterize the Inferring Ground Truth from Subjective Labelling of Venus Images 1087 many volcanoes is a necessary requirement before more advanced planetary geology studies can be carried out: analysis of spatial clustering patterns, correlation with other geologic features, and so forth. From an engineering viewpoint, automation of the volcano detection task presents a significant challenge to current capabilities in computer vision and pattern recognition due to the variability of the volcanoes and the significant background "clutter" present in most of the images. Figure 1 shows a Magellan subimage of size 30km square containing at least 10 small volcanoes. Volcanoes on Venus Figure 1: A 30km x 30km region from the Magellan SAR data, which contains a number of small volcanoes. The purpose of this paper is not to describe pattern recognition methods for volcano detection but rather to discuss some of the issues involved in collecting and calibrating labelled training data from experts. Details of a volcano detection method using matched filtering, SVD projections and a Gaussian classifier are provided in [7]. 3 Volcano Labelling Training examples are collected by having the planetary geologists examine an image on the computer screen and then using a mouse to indicate where they think the volcanoes (if any) are located. Typically it can take from 15 minutes to 1 hour to label an image (depending on how many volcanoes are present), where each image represents a 75km square patch on the surface of Venus. An image may contain on the order of 100 volcanoes, although a more typical number is between 30 and 40. There can be considerable ambiguity in volcano labelling: for the same image, different scientists can produce different label lists, and even the same scientist can produce different lists over time. To address this problem we introduced the notion of having the scientists label training examples into quantized probability 1088 Padhraic Smyth, Usama Fayyad, Michael Burl, Pierro Perona, Pierre Baldi bins or "types", where the probability bins correspond to visually distinguishable sub-categories of volcanoes. In particular, we have used 5 types: (1) summit pits, bright-dark radar pair, and apparent topographic slope, all clearly visible, probability 0.98, (2) only 2 of the 3 criteria of type 1 are visible, probability 0.80, (3) no summit pit visible, evidence of flanks or circular outline, probability 0.60, (4) only a summit pit visible, probability 0.50, (5) no volcano-like features visible, probability 0.0. These subjective probabilities correspond to the mean probability that a volcano exists at a particular location given that it belongs to a particular type and were elicited after considerable discussions with the planetary geologists. Thus, the observed data for each RDI consists of labels l, which are noisy estimates of true "type" t, which in turn is probabilistically related to the hidden event of interest, v, the presence of a volcano: T (1) p(vlD = LP(vlt)p(tID t=1 where T is the number of types (and labels). The subjective probabilities described above correspond to p(vlt): to be able to infer the probability of a volcano given a set of labels l it remains to estimate the p(tlD terms. 4 Inferring the Label-Type Parameters via the EM Procedure We follow a general model for subjective labelling originally proposed by Dawid and Skene [8] and apply it to the image labelling problem: more details on this overall approach are provided in [9]. Let N be the number of local regions of interest (RDl's) in the database (these are 15 pixel square image patches for the volcano application) . For simplicity we consider the case of just a single labeller who labels a given set of regions of interest (RDIs) a number of times - the extension to multiple labellers is straightforward assuming conditional independence of the labellings given the true type. Let nil be the number of times that RDI i is labelled with labell. Let lit denote a binary variable which takes value 1 if the true type of volcano i is t* , and is 0 otherwise. We assume that labels are assigned independently to a given RDI from one labelling to the next, given that the type is known. If the true type t* is known then T p(observed labelslt*, i) ex: II p(lltti!? (2) 1=1 Thus, unconditionally, we have p(observed labels, t*li) ex: )Yit gT(T pet) gP(lltt il , (3) where lit = 1 if t = t* and 0 otherwise. Assuming that each RDI is labelled independently of the others (no spatial correlation in terms of labels), Ng T(T )Yit pet) gp(lltt'l p(observed labels, t;) ex: ~ (4) Inferring Ground Truth from Subjective Labelling of Venus Images 1089 Still assuming that the types t for each ROI are known (the Yit)' the maximum likelihood estimators of p(llt) and p(t) are p(llt) = ~i Yitnil EI Ei Yitnil (5) and (6) From Bayes' rule one can then show that 1 T p(Yit = llobserved data) = C IIp(llt)nilp(t) (7) I where C is a normalization constant. Thus, given the observed data nil and the parameters p(llt) and p(t), one can infer the posterior probabilities of each type via Equation 7. However, without knowing the Yit values we can not infer the parameters p(llt) and p(t). One can treat the Yit as hidden and thus apply the well-known ExpectationMaximization (EM) procedure to find a local maximum of the likelihood function: 1. Obtain some initial estimates of the expected values of Yit, e.g., E[Yitl = ~ Elnil (8) 2. M-step: choose the values of p(llt) and p(t) which maximize the likelihood function (according to Equations 5 and 6), using E[Yitl in place of Yit. 3. E-step: calculate the conditional expectation of Yit, E[Yitldatal = p(Yit = Ildata) (Equation 7). 4. Repeat Steps 2 and 3 until convergence. 5 5.1 Experimental Results Combining Multiple Expert Opinions Labellings from 4 geologists on the 4 images resulted in 269 possible volcanoes (ROIs) being identified. Application of the EM procedure resulted in label-type probability matrices as shown in Table 1 for Labeller C. The diagonal elements provide an indication of the reliability of the labeller. There is significant miscalibration for label 3's: according to the model, a label 3 from Labeller C is most likely to correspond to type 2. The label-type matrices of all 4 labellers (not shown) indicated that the model placed more weight on the conservative labellers (C and D) than the aggressive ones (A and B). The determination of posterior probabilities for each of the ROIs is a fundamental step in any quantitative analysis of the volcano data: p(vlD = E;=1 p(vlt)p(tID where the p(tlD terms are the posterior probabilities of type given labels provided 1090 Padhraic Smyth, Usama Fayyad, Michael Burl, Pietro Perona, Pierre Baldi Table 1: Type-Label Probabilities for Individual Labellers as estimated via the EM Procedure Labell Label 2 Label 3 Label 4 Label 5 Probability(typellabel), Labeller C Type 1 Type 2 Type 3 Type 4 1.000 0.000 0.000 0.000 0.019 0.977 0.004 0.000 0.000 0.667 0.175 0.065 0.000 0.725 0.000 0.042 0.000 0.000 0.389 0.000 Type 5 0.000 0.000 0.094 0.233 0.611 Table 2: 10 ROIs from the database: original scientist labels shown with posterior probabilities estimated via the EM procedure Scientist Labels JD ROI 1 2 3 4 5 6 7 8 9 10 A B 4 1 1 3 3 2 3 2 3 4 4 4 1 1 1 2 1 1 2 4 C 4 4 2 5 3 2 5 4 5 4 D 5 2 2 3 3 4 5 4 3 4 Posterior Probabilities (EM), p(tlD Type 1 Type 2 Type 3 Type 4 Type 5 0.000 0.000 0.000 0.816 0.184 0.000 0.000 0.000 0.991 0.009 0.023 0.000 0.000 0.977 0.000 0.000 0.000 1.000 0.000 0.000 0.000 0.536 0.452 0.012 0.000 0.000 1.000 0.000 0.000 0.000 0.000 0.000 0.000 1.000 0.000 0.000 0.000 0.000 0.999 0.000 0.000 0.000 0.992 0.000 0.008 0.000 0.000 0.000 0.996 0.004 p(vlO 0.408 0.496 0.804 0.600 0.706 0.800 0.600 0.500 0.595 0.498 by the EM procedure, and the p(vlt) terms are the subjective volcano-type probabilities discussed in Section 3.2. As shown in Table 2, posterior probabilities for the volcanoes generally are in agreement with intuition and often correspond to taking the majority vote or the "average" of the C and D labels (the conservative labellers). However some p(vlD estimates could not easily be derived by any simple averaging or voting scheme, e.g., see ROIs 3, 5 and 7 in the table. 5.2 Experiment on Comparing Human and Algorithm Performance The standard receiver operating characteristic (ROC) plots detections versus false alarms [10] . The ROCs shown here differ in two significant ways [11] : (1) the false alarm axis is normalized relative to the number of true positives (necessary since the total number of possible false alarms is not well defined for object detection in images), and (2) the reference labels used in scoring are probabilistic: a detection "scores" p(v) on the detection axis and 1 - p( v) on the false alarm axis. Inferring Ground Truth from Subjective Labelling of Venus Images 100 100 Mean Detection Rate [%) 80 Mean Detection Rate [%) 80 60 60 40 40 20 1091 -SVD Algorithm -t( Scientist A -0 Scientist B -- Scientist C .+ ScientistD 0~_ _~_ _~~======7 o 20 40 60 80 Mean False Alarm Rate (expressed as % of total number of volcanoes) -SVD Algorithm -t( Scientist A -0 Scientist B -- Scientist C .+ ScientistD 20 40 60 80 Mean False Alarm Rate (expressed as % of total number of volcanoes) Figure 2: Modified ROCs for both scientists and algorithms: (a) without the labelling or type uncertainty, (b) with full uncertainty model factored in. As before, data came from 4 images, and there were 269 labelled local regions. The SVD-Gaussian algorithm was evaluated in cross-validation mode (train on 3 images, test on the 4th) and the results combined. The first ROC (Figure 2(a)) does not take into account either label-type or type-volcano probabilities, i.e., the reference list (for algorithm training and overall evaluation) is a consensus list (2 scientists working together) where labels 1,2,3,4 are ignored and all labelled items are counted equally as volcanoes. The individual labellers and algorithm are then scored in the standard "non-weighted" ROC fashion. This curve is optimistic in terms of depicting the accuracy of the detectors since it ignores the underlying probabilistic nature of the labels. Even with this optimistic curve, volcano labelling is relatively inaccurate by either man or machine. Figure 2(b) shows a weighted ROC: for each of 4 scientists the probabilistic "reference labels" were derived via the EM procedure as in Table 2 from the other 3 scientists, and the detections of each scientist were scored according to each such reference set. Performance of the algorithm (the SVD-Gaussian method) was evaluated relative to the EM-derived label estimates of all 4 scientists. Accounting for all of the uncertainty in the data results in a more realistic, if less flattering, set of performance characteristics. The algorithm's performance degrades more than the scientist's performance (for low false alarms rates compared to Figure 2(a)) when the full noise model is used. The algorithm is estimating the posterior probabilities of volcanoes rather poorly and the complete uncertainty model is more sensitive to this fact. This is a function of the SVD feature space rather than the Gaussian classification model. 1092 6 Padhraic Smyth, Usama Fayyad, Michael Burl, Pietro Perona, Pierre Baldi Conclusion Ignoring subjective uncertainty in image labelling can lead to significant overconfidence in terms of performance estimation (for both humans and machines). For the volcano detection task a simple model for uncertainty in the class labels provided insight into the performance of both human and algorithmic detectors. An obvious extension of the maximum likelihood framework outlined here is a Bayesian approach [12]: accounting for parameter uncertainty in the model given the limited amount of training data available is worth investigating. Acknowledgements The research described in this paper was carried out by the Jet Propulsion Laboratory, California Institute of Technology, under a contract with the National Aeronautics and Space Administration and was supported in part by ARPA under grant number NOOOl4-92-J-1860. References 1. P. Smyth, "Learning with probabilistic supervision," in Computational Learning Theory and Natural Learning Systems 3, T. Petcshe, M. Kearns, S . Hanson, R. Rivest (eds), Cambridge, MA: MIT Press, to appear. 2. S. French, "Group consensus probability distributions: a critical survey," in Bayesian Statistics 2, J. M. Bernardo, M . H. DeGroot, D. V. Lindley, A. F. M. Smith (eds.), Elsevier Science Publishers, North-Holland, pp.183-202, 1985. 3. J . S. Uebersax, "Statistical modeling of expert ratings on medical treatment appropriateness," J. Amer. Statist. Assoc., vol.88, no.422, pp.421-427, 1993. 4. G. Lugosi, "Learning with an unreliable teacher," Pattern Recognition, vol. 25, no.1, pp.79-87. 1992. 5. Science, special issue on Magellan data, April 12, 1991. 6. J . C. Aubele and E . N. Slyuta, "Small domes on Venus: characteristics and origins," in Earth, Moon and Planets, 50/51, 493-532, 1990. 7. M. C. Burl, U. M . Fayyad, P . Perona, P . Smyth, and M. P . Burl, "Automating the hunt for volcanoes on Venus," in Proceedings of the 1994 Computer Vision and Pattern Recognition Conference: CVPR-94, Los Alamitos, CA: IEEE Computer Society Press, pp.302-309, 1994. 8. A. P . Dawid and A. M. Skene, "Maximum likelihood estimation of observer error-rates using the EM algorithm," Applied Statistics, vol.28, no.1, pp.2G-28, 1979. 9. P. Smyth, M. C . Burl, U. M. Fayyad, P . Perona, 'Knowledge discovery in large image databases: dealing with uncertainties in ground truth,' in Knowledge Discovery in Databases 2, U. M. Fayyad, G. Piatetsky-Shapiro, P . Smyth, R. Uthurasamy (eds.), AAAI/MIT Press, to appear, 1995. 10. M. S. Chesters, "Human visual perception and ROC methodology in medical imaging," Phys. Med. BioI., vol.37, no.7, pp.1433-1476, 1992. 11. M. C. Burl, U. M . Fayyad, P. Perona, P . Smyth, "Automated analysis of radar imagery of Venus: handling lack of ground truth," in Proceedings of the IEEE Conference on Image Processing, Austin, November 1994. 12. W . Buntine, "Operations for learning with graphical models," Journal of Artificial Intelligence Research, 2, pp.159-225, 1994.
949 |@word km:5 accounting:3 carry:1 initial:1 contains:1 score:1 subjective:16 current:1 comparing:3 planet:1 realistic:1 visible:5 plot:1 intelligence:1 geologic:1 item:1 smith:1 provides:1 quantized:1 detecting:1 location:1 oak:2 chester:1 consists:1 baldi:5 manner:1 spacecraft:1 expected:1 rapid:1 examine:2 provided:4 estimating:1 matched:1 underlying:1 rivest:1 developed:1 sky:1 quantitative:1 collecting:1 voting:1 bernardo:1 classifier:1 assoc:1 medical:5 grant:1 appear:2 positive:1 before:2 tid:2 engineering:2 scientist:19 local:3 treat:1 severely:1 lugosi:2 therein:1 pit:3 limited:1 hunt:1 practical:1 procedure:8 empirical:1 projection:1 suggest:1 context:1 map:1 straightforward:2 independently:2 survey:1 resolution:1 simplicity:1 factored:1 estimator:1 rule:1 insight:1 notion:1 sar:3 usama:5 smyth:10 origin:1 agreement:1 dawid:2 element:1 recognition:8 particularly:1 located:1 summit:3 database:4 observed:5 electrical:1 calculate:1 region:4 remote:3 intuition:1 radar:4 division:1 basis:1 easily:1 various:2 train:2 investigative:1 describe:1 vlt:4 artificial:1 quite:1 apparent:1 cvpr:1 otherwise:2 ability:2 statistic:2 topographic:1 think:1 gp:2 noisy:3 indication:1 mission:1 combining:2 poorly:1 tax:1 forth:1 los:1 convergence:1 requirement:1 produce:3 object:2 depending:1 geology:1 ex:3 expectationmaximization:1 indicate:1 differ:1 appropriateness:1 human:8 opinion:3 bin:2 require:1 strictly:1 extension:2 ground:11 roi:6 visually:2 algorithmic:2 earth:2 purpose:1 estimation:2 label:35 sensitive:1 weighted:2 mit:2 clearly:1 gaussian:4 modified:1 rather:3 geologist:4 probabilistically:1 derived:3 focus:2 modelling:1 likelihood:5 detect:1 elsevier:1 vl:1 typically:1 inaccurate:1 pasadena:3 perona:8 hidden:3 pixel:2 issue:4 overall:2 classification:1 spatial:2 special:1 field:1 aware:1 having:2 ng:1 manually:1 biology:1 represents:1 lit:2 others:1 primarily:1 few:1 modern:1 resulted:2 national:1 individual:2 consisting:1 detection:13 interest:4 circular:1 evaluation:2 mixture:1 grove:2 necessary:2 experience:1 literally:1 theoretical:1 arpa:1 modeling:1 maximization:1 rdi:4 characterize:1 buntine:1 teacher:1 synthetic:1 combined:2 fundamental:1 automating:1 probabilistic:5 off:1 contract:1 michael:5 together:1 mouse:1 venusian:1 aaai:1 central:1 ambiguity:1 iip:1 choose:1 padhraic:5 opposed:1 containing:1 imagery:1 expert:9 li:1 sible:1 aggressive:1 account:1 automation:1 north:1 later:1 observer:1 optimistic:2 analyze:1 bayes:1 capability:2 elicited:1 slope:1 lindley:1 contribution:1 square:3 bright:1 il:1 moon:1 accuracy:1 who:1 characteristic:3 gathered:1 correspond:5 vlo:1 bayesian:3 accurately:1 worth:1 drive:2 piatetsky:1 detector:2 llt:6 phys:1 manual:1 ed:3 pp:7 involved:1 obvious:1 treatment:1 knowledge:2 back:1 magellan:6 originally:1 supervised:1 follow:1 methodology:1 april:1 amer:1 evaluated:2 just:1 correlation:2 until:1 working:1 ei:2 lack:1 french:2 mode:1 indicated:1 scientific:1 calibrating:2 contain:1 true:5 normalized:1 burl:9 assigned:1 laboratory:3 noted:1 m:1 criterion:1 trying:1 outline:1 complete:1 image:28 discussed:1 interpretation:1 significant:7 cambridge:1 outlined:1 reliability:2 supervision:1 surface:3 operating:1 gt:1 aeronautics:1 posterior:7 recent:2 perspective:1 belongs:1 certain:1 binary:1 came:1 caltech:3 scoring:1 transmitted:1 greater:1 maximize:1 ii:4 full:2 multiple:3 infer:4 jet:3 determination:1 cross:1 equally:1 vision:2 expectation:2 normalization:1 background:2 pierro:1 publisher:1 degroot:1 med:1 counting:1 presence:1 automated:2 independence:1 identified:1 venus:13 knowing:1 administration:1 labeller:5 locating:1 ignored:1 generally:1 amount:1 clutter:1 dark:1 statist:1 category:1 telescope:1 diameter:1 generate:1 shapiro:1 estimated:4 per:1 diagnosis:3 vol:4 group:1 yit:10 imaging:1 pietro:4 year:1 volcano:37 uncertainty:8 place:1 throughout:1 patch:2 aspect:1 argument:1 fayyad:9 relatively:1 skene:2 department:1 according:3 miscalibration:1 em:10 lp:1 labellings:2 modification:1 swamped:1 equation:3 remains:2 discus:3 turn:1 available:1 operation:1 apply:3 appropriate:1 pierre:5 jd:1 original:1 clustering:1 graphical:1 rdl:1 society:1 objective:1 alamitos:1 degrades:1 concentration:1 diagonal:1 majority:1 propulsion:3 topic:1 collected:2 consensus:2 trivial:1 pet:2 assuming:3 november:1 situation:1 variability:1 locate:1 lltt:2 inferred:1 rating:2 introduced:1 namely:1 pair:1 hanson:1 california:2 planetary:6 hour:2 address:1 able:1 pattern:10 perception:1 appeared:1 challenge:1 reliable:1 event:1 critical:1 natural:1 advanced:1 scheme:1 noool4:1 technology:3 axis:3 carried:4 unconditionally:1 literature:1 acknowledgement:1 discovery:2 relative:3 filtering:1 versus:1 validation:1 tld:3 viewpoint:2 austin:1 repeat:1 placed:1 supported:1 bias:1 institute:2 taking:1 subimage:1 absolute:1 curve:2 ignores:1 made:1 collection:1 dome:1 counted:1 far:1 aperture:1 unreliable:1 dealing:1 investigating:1 receiver:1 latent:2 table:6 nature:1 reasonably:1 ca:4 ignoring:1 flank:1 depicting:1 noise:2 alarm:7 scored:2 body:1 overconfidence:1 roc:7 scattered:1 screen:1 fashion:1 guise:1 inferring:7 thematic:1 wish:1 sub:1 minute:1 specific:1 sensing:3 list:4 evidence:1 exists:1 false:7 labelling:14 photograph:1 simply:1 distinguishable:1 likely:1 visual:1 expressed:2 holland:1 truth:13 ma:1 conditional:2 bioi:1 quantifying:1 twofold:2 labelled:5 absence:1 man:2 considerable:2 typical:2 averaging:1 kearns:1 conservative:2 nil:2 total:3 svd:6 experimental:1 vote:1 evaluate:1 handling:1
7,188
95
305 ALVINN: AN AUTONOMOUS LAND VEHICLE IN A NEURAL NETWORK Dean A. Pomerleau Computer Science Department Carnegie Mellon University Pittsburgh, PA 15213 ABSTRACT ALVINN (Autonomous Land Vehicle In a Neural Network) is a 3-layer back-propagation network designed for the task of road following. Currently ALVINN takes images from a camera and a laser range finder as input and produces as output the direction the vehicle should travel in order to follow the road. Training has been conducted using simulated road images. Successful tests on the Carnegie Mellon autonomous navigation test vehicle indicate that the network can effectively follow real roads under certain field conditions. The representation developed to perfOIm the task differs dramatically when the networlc is trained under various conditions, suggesting the possibility of a novel adaptive autonomous navigation system capable of tailoring its processing to the conditions at hand. INTRODUCTION Autonomous navigation has been a difficult problem for traditional vision and robotic techniques, primarily because of the noise and variability associated with real world scenes. Autonomous navigation systems based on traditional image processing and pattern recognition techniques often perform well under certain conditions but have problems with others. Part of the difficulty stems from the fact that the processing performed by these systems remains fixed across various driving situations. Artificial neural networks have displayed promising performance and flexibility in other domains characterized by high degrees of noise and variability, such as handwritten character recognition [Jackel et al., 1988] [Pawlicki et al., 1988] and speech recognition [Waibel et al., 1988]. ALVINN (Autonomous Land Vehicle In a Neural Network) is a connectionist approach to the navigational task of road following. Specifically, ALVINN is an artificial neural network designed to control the NAVLAB, the Carnegie Mellon autonomous navigation test vehicle. NETWORK ARCmTECTURE ALVINN's current architecture consists of a single hidden layer back-propagation network 306 Pomerleau Road Intensity Feedback Unit 45 Direction Output Units 8x32 Range Finder Input Retina 30x32 Video Input Retina Figure 1: ALVINN Architecture (See Figure 1). The input layer is divided into three sets of units: two "retinas" and a single intensity feedback unit. The two retinas correspond to the two forms of sensory input available on the NAVLAB vehicle; video and range information. The first retina, consisting of 3002 units, receives video camera input from a road scene. The activation level of each unit in this retina is proportional to the intensity in the blue color band of the corresponding patch of the image. The blue band of the color image is used because it provides the highest contrast between the road and the non-road. The second retina, consisting of 8x32 units, receives input from a laser range finder. The activation level of each unit in this retina is proportional to the proximity of the corresponding area in the image. The road intensity feedback unit indicates whether the road is lighter or darker than the non-road in the previous image. Each of these 1217 input units is fully connected to the hidden layer of 29 units, which is in tum fully connected to the output layer. The output layer consists of 46 units, divided into two groups. The first set of 45 units is a linear representation of the tum curvature along which the vehicle should travel in order to head towards the road center. The middle unit represents the "travel straight ahead" condition while units to the left and right of the center represent successively sharper left and right turns. The network is trained with a desired output vector of all zeros except for a "hill" of activation centered on the unit representing the correct tum curvature, which is the curvature which would bring the vehicle to the road center 7 meters ahead of its current position. More specifically, the desired activation levels for ALVlNN: An Autonomous Land Vehicle in a Neural Network Real Road Image Simulated Road Image Figure 2: Real and simulated road images the nine units centered around the correct tum curvature unit are 0.10, 0.32, 0.61, 0.89, 1.00,0.89,0.61,0.32 and 0.10. During testing, the tum curvature dictated by the network is taken to be the curvature represented by the output unit with the highest activation level. The final output unit is a road intensity feedback unit which indicates whether the road is lighter or darker than the non-road in the current image. During testing, the activation of the output road intensity feedback unit is recirculated to the input layer in the style of Jordan [Jordan, 1988] to aid the network's processing by providing rudimentary infonnation concerning the relative intensities of the road and the non-road in the previous image. TRAINING AND PERFORMANCE Training on actual road images is logistically difficult, because in order to develop a general representation, the network must be presented with a large number of training exemplaIS depicting roads under a wide variety of conditions. Collection of such a data set would be difficult, and changes in parameters such as camera orientation would require collecting an entirely new set of road images. To avoid these difficulties we have developed a simulated road generator which creates road images to be used as training exemplars for the network. Figure 2 depicts the video images of one real and one artificial road. Although not shown in Figure 2, the road generator also creates corresponding simulated range finder images. At the relatively low resolution being used it is difficult to distinguish between real and simulated roads. NetwoIk: training is performed using these artificial road "snapshots" and the Warp back- 307 ? 308 Pomerleau Figure 3: NAVLAB, the eMU autonomous navigation test vehicle. propagation simulator described in [Pomerleau et al., 1988]. Training involves first creating a set of 1200 road snapshots depicting roads with a wide variety of retinal orientations and positions, under a variety of lighting conditions and with realistic noise levels. Back-propagation is then conducted using this set of exemplars until only asymptotic performance improvements appear likely. During the early stages of training, the input road intensity unit is given a random activation level. This is done to prevent the network from merely learning to copy the activation level of the input road intensity unit to the output road intensity unit, since their activation levels should almost always be identical because the relative intensity of the road and the non-road does not often change between two successive images. Once the network has developed a representation that uses image characteristics to determine the activation level for the output road intensity unit, the network is given as input whether the road would have been darker or lighter than the non-road in the previous image. Using this extra information concerning the relative brightness of the road and the non-road, the network is better able to determine the correct direction for the vehicle to trave1. After 40 epochs of training on the 1200 simulated road snapshots, the network correctly dictates a tum curvature within two units of the correct answer approximately 90% of the time on novel simulated road images. The primary testing of the ALVINN's performance has been conducted on the NAVLAB (See Figure 3). The NAVLAB is a modified Chevy van equipped with 3 Sun computers, a Warp, a video camera, and a laser range finder, which serves as a testbed for the CMU autonomous land vehicle project [Thorpe et al., 1987]. Performance of the network to date is comparable to that achieved by the best traditional vision-based autonomous navigation algorithm at CMU under the limited conditions tested. Specifically, the network can accurately drive the NAVLAB at a speed of 1/2 meter per second along a 400 meter path through a wooded ALVINN: An Autonomous Land Vehicle in a Neural Network Weights to Direction Output Units t~li'C jjllil Weight to Output Feedback Unit D Weights from Video Camera Retina Road 1 Edges Weight from Input Feedback Unit n Weight from Bias Unit ? Weights from Range Finder Retina Excitatory Periphery Connections Inhibitory Central Connections Figure 4: Diagram of weights projecting to and from a typical hidden unit in a network trained on roads with a fixed width. The schematics on the right are aids for interpretation. area of the CMU campus under sunny fall conditions. Under similar conditions on the same course, the ALV group at CMU has recently achieved similar driving accuracy at a speed of one meter per second by implementing their image processing autonomous navigation algorithm on the Watp computer. In contrast, the ALVINN network is currently simulated using only an on-boani Sun computer, and dramatic speedups are expected when tests are perfonned using the Watp. NETWORK REPRESENTATION The representation developed by the network to perfonn the road following task depends dramatically on the characteristics of the training set. When trained on examples of roads with a fixed width, the network develops a representations consisting of overlapping road filters. Figure 4 is a diagram of the weights projecting to and from a single hidden unit . in such a network. As indicated by the weights to and from the feedback units, this hidden unit expects the road to be lighter than the non-road in the previous image and supports the road being lighter than the non-road in the current image. More specifically, the weights from the 309 310 Pomerleau Weights to Direction Output Units 1::tllII:UIll t k? Weight to Output Feedback Unit - mDIRoad ~ N on-road ~ Weights from Video Camera Retina . ' : ' .' : .. Weight from Input Feedback Unit . : ' : .' 11111111 1111111 I[ Weight from Bias Unit 111111111111 ? .: . Weights from Range Finder Retina Figure 5: Diagram of weights projecting to and from a typical hidden unit in a network trained on roads with different widths. video camera retina support the interpretation that this hidden unit is a filter for two light roads, one slightly left and the other slightly right or center (See schematic to the right of the weights from the video retina in Figure 4). This interpretation is also supported by the weights from the range finder retina, which has a much wider field of view than the video camera. This hidden unit is excited if there is high range activity (Le. obstacles) in the periphery and inhibited if there is high range activity in the central region of the scene where this hidden unit expects the road to be (See schematic to the right of the weights from the range finder retina in Figure 4). Finally, the two road filter interpretation is reflected in the weights from this hidden unit to the direction output units. Specifically, this hidden unit has two groups of excitatory connections to the output units, one group dictating a slight left turn and the other group dictating a slight right turn. Hidden units which act as filters for 1 to 3 roads are the representation structures most commonly developed when the network is trained on roads with a fixed width. The network develops a very different representation when trained on snapshots with widely varying road widths. A typical hidden unit from this type of representation is depicted in figure 5. One important feature to notice from the feedback weights is that this unit is filtering for a road which is darlcer than the non-road. More importantly, it is evident from the video camera retina weights that this hidden unit is a filter solely for the left edge of the road (See schematic to the right of the weights from the range finder ALVINN: An Autonomous Land Vehicle in a Neural Network retina in Figure 5). This hidden unit supports a rather wide range of travel directions. This is to be expected, since the correct travel direction for a road with an edge at a particular location varies substantially depending on the road's width. This hidden unit would cooperate with hidden units that detect the right road edge to determine the correct travel direction in any particular situation. DISCUSSION AND EXTENSIONS The distinct representations developed for different circumstances illustrate a key advantage provided by neural networks for autonomous navigation. Namely, in this paradigm the data, not the programmer, determines the salient image features crucial to accurate road navigation. From a practical standpoint, this data responsiveness has dramatically sped ALVINN's development. Once a realistic artificial road generator was developed, back-propagation producted in half an hour a relatively successful road following system. It took many months of algorithm development and parameter tuning by the vision and autonomous navigation groups at CMU to reach a similar level of performance using traditional image processing and pattern recognition techniques. More speculatively, the flexibility of neural network representations provides the possibility of a very different type of autonomous navigation system in which the salient sensory features are determined for specific driving conditions. By interactively training the network on real road images taken as a human drives the NAVLAB, we hope to develop a system that adapts its processing to accommodate current circumstances. This is in contrast with other autonomous navigation systems at CMU [Thorpe et al., 1987] and elsewhere [Dunlay & Seida, 1988] [Dickmanns & Zapp, 1986] [Kuan et al., 1988]. Each of these implementations has relied on a fixed, highly structured and therefore relatively inflexible algorithm for finding and following the road, regardless of the conditions at hand. There are difficulties involved with training "on-the-fly" with real images. If the network is not presented with sufficient variability in its training exemplars to cover the conditions it is likely to encounter when it takes over driving from the human operator, it will not develop a sufficiently robust representation and will perform poorly. In addition, the network must not solely be shown examples of accurate driving, but also how to recover (i.e. return to the road center) once a mistake has been made. Partial initial training on a variety of simulated road images should help eliminate these difficulties and facilitate better performance. Another important advantage gained through the use of neural networks for autonomous navigation is the ease with which they assimilate data from independent sensors. The current ALVINN implementation processes data from two sources, the video camera and the laser range finder. During training, the network discovers how information from each source relates to the task, and weights each accordingly. As an example, range data is in some sense less important for the task of road following than is the video data. The range data contains information concerning the position of obstacles in the scene, but nothing explicit about the location of the road. As a result, the range data is given less significance in the representation, as is illustrated by the relatively small 311 312 Pomerleau magnitude weights from the range finder retina in the weight diagrams. Figures 4 and 5 illustrate that the range finder connections do correlate with the connections from the video camera, and do contribute to choosing the correct travel direction. Specifically, in both figures, obstacles located outside the area in which the hidden unit expects the road to be located increase the hidden unit's activation level while obstacles located within the expected road boundaries inhibit the hidden unit. However the contributions from the range finger connections aren't necessary for reasonable performance. When ALVINN was tested with normal video input but an obstacle-free range finder image as constant input, there was no noticeable degradation in driving performance. Obviously under off-road driving conditions obstacle avoidance would become much more important and hence one would expect the range finder retina to playa much more significant role in the network's representation. We are currently working on an off-road version of ALVINN to test this hypothesis. Other current directions for this project include conducting more extensive tests of the network's performance under a variety of weather and lighting conditions. These will be crucial for making legitimate performance comparisons between ALVINN and other autonomous navigation techniques. We are also working to increase driving speed by implementing the network simulation on the on-board Warp computer. Additional extensions involve exploring different network architectures for the road following task. These include 1) giving the network additional feedback information by using Elman's [Elman, 1988] technique of recirculating hidden activation levels, 2) adding a second hidden layer to facilitate better internal representations, and 3) adding local connectivity to give the network a priori knowledge of the two dimensional nature of the input In the area of planning, interesting extensions include stopping for, or planning a path around, obstacles. One area of planning that clearly needs work is dealing sensibly with road forks and intersections. Currently upon reaching a fork, the network may output two widely discrepant travel directions, one for each choice. The result is often an oscillation in the dictated travel direction and hence inaccurate road following. Beyond dealing with individual intersections, we would eventually like to integrate a map into the system to enable global point-to-point path planning. CONCLUSION More extensive testing must be performed before definitive conclusions can be drawn concerning the peiformance of ALVINN versus other road followers. We are optimistic concerning the eventual contributions neural networks will make to the area of autonomous navigation. But perhaps just as interesting are the possibilities of contributions in the other direction. We hope that exploring autonomous navigation, and in particular some of the extensions outlined in this paper, will have a significant impact on the field of neural networks. We certainly believe it is important to begin researching and evaluating neural networks in real world situations, and we think autonomous navigation is an interesting application for such an approach. ALVINN: An Autonomous Land Vehicle in a Neural Network Acknowledgements This work would not have been possible without the input and support provided by Dave Touretzky, Joseph Tebelskis, George Gusciora and the CMU Warp group, and particularly Charles Thorpe, Till Crisman, Martial Hebert, David Simon, and rest of the CMU ALV group. This research was supported by the Office of Naval Research under Contracts NOOOI4-87-K-0385, NOOOI4-87-K-0533 and NOOOI4-86-K-0678, by National Science Foundation Grant EET-8716324, by the Defense Advanced Research Projects Agency (DOD) monitored by the Space and Naval Warfare Systems Command under Contract NOOO39-87-C-0251, and by the Strategic Computing Initiative of DARPA, through ARPA Order 5351, and monitored by the U.S. Army Engineer Topographic Laboratories under contract DACA76-85-C-0003 titled "Road Following". References [Dickmanns & Zapp, 1986] Dickmanns, E.D., Zapp, A. (1986) A curvature-based scheme for improving road vehicle guidance by computer vision. "Mobile Robots", SPIE-Proc. Vol. 727, Cambridge, MA. [Elman, 1988] Elman, J.L, (1988) Finding structure in time. Technical report 8801. Center for Research in Language, University of California, San Diego. [Dunlay & Seida, 1988] Dunlay, R.T., Seida, S. (1988) Parallel off-road perception processing on the ALV. Proc. SPIE Mobile Robot Conference, Cambridge MA. [Jackel et al., 1988] Jackel, L.D., Graf, H.P., Hubbard, W., Denker, J.S., Henderson, D., Guyon, 1. (1988) An application of neural net chips: Handwritten digit recognition. Proceedings of IEEE International Conference on Neural Networks, San Diego, CA. [Jordan, 1988] Jordan, M.l. (1988) Supervised learning and systems with excess degrees of freedom. COINS Tech. Report 88-27, Computer and Infolll1ation Science, University of Massachusetts, Amherst MA. [Kuan et al., 1988] Kuan, D., Phipps, G. and Hsueh, A.-C. Autonomous Robotic Vehicle Road Following. IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol. 10, Sept. 1988. [pawlicki et al., 1988] Pawlicki, T.E, Lee, D.S., Hull, J.J., Srihari, S.N. (1988) Neural network models and their application to handwritten digit recognition. Proceedings of IEEE International Conference on Neural Networks, San Diego, CA. [pomerleau et al., 1988] Pomerleau, D.A., Gusciora, G.L., Touretzky, D.S., and Kung, H.T. (1988) Neural network simulation at Waq> speed: How we got 17 million connections per second. Proceedings of IEEE International Conference on Neural Networks, San Diego, CA. [Th0IJle et al., 1987] Thorpe, c., Herbert, M., Kanade, T., Shafer S. and the members of the Strategic Computing Vision Lab (1987) Vision and navigation for the Carnegie Mellon NAVLAB. Annual Revi~ of Computer Science Vol. 11, Ed. Joseph Traub, Annual Reviews Inc., Palo Alto, CA. [Waibel et al., 1988] Waibel, A, Hanazawa, T., Hinton, G., Shikano, K., Lang, K. (1988) Phoneme recognition: Neural Networks vs. Hidden Markov Models. Proceedings from Int. Conf. on Acoustics, Speech and Signal Processing, New York, New York. 313
95 |@word version:1 middle:1 simulation:2 excited:1 brightness:1 dramatic:1 accommodate:1 initial:1 contains:1 current:7 activation:12 lang:1 follower:1 must:3 realistic:2 tailoring:1 designed:2 v:1 half:1 intelligence:1 accordingly:1 provides:2 contribute:1 location:2 successive:1 along:2 become:1 initiative:1 consists:2 expected:3 elman:4 planning:4 simulator:1 actual:1 equipped:1 project:3 provided:2 campus:1 begin:1 alto:1 substantially:1 developed:7 finding:2 perfonn:1 collecting:1 act:1 sensibly:1 control:1 unit:57 grant:1 appear:1 before:1 phipps:1 local:1 mistake:1 path:3 solely:2 approximately:1 ease:1 limited:1 range:23 practical:1 camera:11 testing:4 differs:1 digit:2 area:6 got:1 dictate:1 weather:1 road:90 operator:1 dean:1 map:1 center:6 peiformance:1 regardless:1 resolution:1 x32:3 legitimate:1 avoidance:1 importantly:1 autonomous:26 diego:4 lighter:5 us:1 hypothesis:1 pa:1 recognition:7 particularly:1 located:3 role:1 fork:2 fly:1 region:1 connected:2 sun:2 highest:2 inhibit:1 agency:1 trained:7 creates:2 upon:1 darpa:1 chip:1 various:2 represented:1 finger:1 laser:4 distinct:1 artificial:5 choosing:1 outside:1 widely:2 topographic:1 think:1 hanazawa:1 final:1 kuan:3 obviously:1 advantage:2 net:1 took:1 wooded:1 date:1 till:1 flexibility:2 poorly:1 adapts:1 produce:1 wider:1 depending:1 develop:3 illustrate:2 help:1 exemplar:3 noticeable:1 involves:1 indicate:1 direction:14 correct:7 filter:5 hull:1 centered:2 human:2 enable:1 programmer:1 implementing:2 require:1 daca76:1 extension:4 exploring:2 proximity:1 around:2 sufficiently:1 normal:1 driving:8 early:1 proc:2 travel:9 currently:4 infonnation:1 jackel:3 palo:1 hubbard:1 hope:2 clearly:1 sensor:1 always:1 modified:1 rather:1 reaching:1 avoid:1 varying:1 command:1 mobile:2 office:1 naval:2 improvement:1 indicates:2 tech:1 contrast:3 detect:1 sense:1 warfare:1 stopping:1 inaccurate:1 eliminate:1 hidden:23 orientation:2 priori:1 development:2 field:3 once:3 identical:1 represents:1 report:2 others:1 connectionist:1 develops:2 inhibited:1 primarily:1 retina:20 thorpe:4 national:1 individual:1 consisting:3 freedom:1 assimilate:1 possibility:3 highly:1 certainly:1 henderson:1 navigation:19 traub:1 light:1 uill:1 accurate:2 edge:4 capable:1 partial:1 necessary:1 desired:2 guidance:1 arpa:1 obstacle:7 cover:1 strategic:2 expects:3 dod:1 successful:2 conducted:3 answer:1 varies:1 international:3 amherst:1 contract:3 off:3 lee:1 connectivity:1 speculatively:1 central:2 successively:1 interactively:1 conf:1 creating:1 style:1 return:1 li:1 suggesting:1 retinal:1 int:1 inc:1 depends:1 vehicle:18 performed:3 view:1 lab:1 optimistic:1 relied:1 recover:1 parallel:1 simon:1 alv:3 contribution:3 accuracy:1 phoneme:1 characteristic:2 conducting:1 correspond:1 handwritten:3 accurately:1 lighting:2 drive:2 straight:1 dave:1 reach:1 touretzky:2 ed:1 involved:1 associated:1 spie:2 monitored:2 noooi4:3 massachusetts:1 color:2 knowledge:1 back:5 tum:6 supervised:1 follow:2 reflected:1 done:1 just:1 stage:1 until:1 hand:2 receives:2 working:2 overlapping:1 propagation:5 perhaps:1 indicated:1 believe:1 facilitate:2 hence:2 laboratory:1 illustrated:1 during:4 width:6 hill:1 evident:1 bring:1 rudimentary:1 image:30 netwoik:1 cooperate:1 novel:2 recently:1 charles:1 discovers:1 sped:1 million:1 interpretation:4 slight:2 mellon:4 significant:2 gusciora:2 cambridge:2 tuning:1 outlined:1 language:1 robot:2 playa:1 curvature:8 dictated:2 periphery:2 certain:2 dickmanns:3 responsiveness:1 herbert:1 additional:2 george:1 determine:3 paradigm:1 signal:1 relates:1 stem:1 technical:1 characterized:1 divided:2 concerning:5 finder:15 schematic:4 impact:1 vision:6 cmu:8 circumstance:2 represent:1 achieved:2 navlab:8 addition:1 discrepant:1 diagram:4 source:2 crucial:2 standpoint:1 extra:1 rest:1 member:1 jordan:4 variety:5 architecture:3 whether:3 defense:1 titled:1 speech:2 york:2 nine:1 dictating:2 dramatically:3 involve:1 band:2 inhibitory:1 notice:1 correctly:1 per:3 blue:2 carnegie:4 vol:3 group:8 key:1 salient:2 drawn:1 prevent:1 merely:1 almost:1 reasonable:1 guyon:1 patch:1 oscillation:1 comparable:1 entirely:1 layer:8 networlc:1 distinguish:1 annual:2 activity:2 ahead:2 scene:4 speed:4 tebelskis:1 relatively:4 speedup:1 department:1 structured:1 waibel:3 sunny:1 across:1 slightly:2 inflexible:1 character:1 joseph:2 making:1 projecting:3 taken:2 remains:1 turn:3 eventually:1 serf:1 available:1 denker:1 encounter:1 coin:1 include:3 giving:1 primary:1 traditional:4 simulated:10 providing:1 difficult:4 sharper:1 pomerleau:8 implementation:2 perform:2 snapshot:4 markov:1 displayed:1 situation:3 hinton:1 variability:3 head:1 intensity:12 david:1 namely:1 extensive:2 connection:7 california:1 acoustic:1 testbed:1 emu:1 hour:1 trans:1 able:1 beyond:1 pattern:3 perception:1 navigational:1 video:15 perfonned:1 difficulty:4 advanced:1 representing:1 scheme:1 martial:1 sept:1 epoch:1 review:1 acknowledgement:1 meter:4 relative:3 asymptotic:1 graf:1 fully:2 expect:1 interesting:3 proportional:2 filtering:1 versus:1 generator:3 foundation:1 integrate:1 degree:2 sufficient:1 land:8 excitatory:2 course:1 elsewhere:1 supported:2 copy:1 free:1 hebert:1 bias:2 warp:4 wide:3 fall:1 van:1 feedback:12 boundary:1 world:2 evaluating:1 sensory:2 collection:1 adaptive:1 commonly:1 made:1 san:4 arcmtecture:1 correlate:1 excess:1 eet:1 dealing:2 global:1 robotic:2 pittsburgh:1 shikano:1 promising:1 nature:1 kanade:1 robust:1 ca:4 depicting:2 improving:1 alvinn:18 domain:1 significance:1 noise:3 definitive:1 shafer:1 nothing:1 depicts:1 board:1 darker:3 aid:2 position:3 explicit:1 specific:1 adding:2 effectively:1 gained:1 magnitude:1 chevy:1 aren:1 depicted:1 intersection:2 likely:2 army:1 srihari:1 determines:1 ma:3 month:1 towards:1 eventual:1 change:2 specifically:6 except:1 typical:3 determined:1 degradation:1 engineer:1 internal:1 support:4 kung:1 tested:2
7,189
950
Using Voice Transformations to Create Additional Training Talkers for Word Spotting Eric I. Chang and Richard P. Lippmann MIT Lincoln Laboratory Lexington, MA 02173-0073, USA eichang@sst.ll.mit.edu and rpl@sst.ll.mit.edu Abstract Speech recognizers provide good performance for most users but the error rate often increases dramatically for a small percentage of talkers who are "different" from those talkers used for training. One expensive solution to this problem is to gather more training data in an attempt to sample these outlier users. A second solution, explored in this paper, is to artificially enlarge the number of training talkers by transforming the speech of existing training talkers. This approach is similar to enlarging the training set for OCR digit recognition by warping the training digit images, but is more difficult because continuous speech has a much larger number of dimensions (e.g. linguistic, phonetic, style, temporal, spectral) that differ across talkers. We explored the use of simple linear spectral warping to enlarge a 48-talker training data base used for word spotting. The average detection rate overall was increased by 2.9 percentage points (from 68.3% to 71.2%) for male speakers and 2.5 percentage points (from 64.8% to 67.3%) for female speakers. This increase is small but similar to that obtained by doubling the amount of training data. 1 INTRODUCTION Speech recognizers, optical character recognizers, and other types of pattern classifiers used for human interface applications often provide good performance for most users. Performance is often, however, low and unacceptable for a small percentage of "outlier" users who are presumably not represented in the training data. One expensive solution to this problem is to obtain more training data in the hope of including users from these outlier 876 Eric I. Chang, Richard P. Lippmann classes. Other approaches already used for speech recognition are to use input features and distance metrics that are relatively invariant to linguistically unimportant differences between talkers and to adapt a recognizer for individual talkers. Talker adaptation is difficult for word spotting and with poor outlier users because the recognition error rate is high and talkers often can not be prompted to recite standard phrases that can be used for adaptation. An alternative approach, that has not been fully explored for speech recognition, is to artificially expand the number of training talkers using voice transformations. Transforming the speech of one talker to make it sound like that of another is difficult because speech varies across many difficult-to-measure dimensions including linguistic, phonetic, duration, spectra, style, and accent. The transformation task is thus more difficult than in optical character recognition where a small set of warping functions can be successfully applied to character images to enlarge the number of training images (Drucker, 1993). This paper demonstrates how a transformation accomplished by warping the spectra of training talkers to create more training data can improve the performance of a whole-word word spotter on a large spontaneous-speech data base. 2 BASELINE WORD SPOTTER A hybrid radial basis function (RBF) - hidden Markov model (HMM) keyword spotter has been developed over the past few years that provides state-of-the-art performance for a whole-word word spotter on the large spontaneous-speech credit-card speech corpus. This system spots 20 target keywords, includes one general filler class, and uses a Viterbi decoding backtrace as described in (Chang, 1994) to backpropagate errors over a sequence of input speech frames. This neural network word spotter is trained on target and background classes, normalizes target outputs using the background output, and thresholds the resulting score to generate putative hits, as shown in Figure 1. Putative hits in this figure are input patterns which generate normalized scores above a threshold. The performance of this, and other spotting systems, is analyzed by plotting a detection versus false alarm rate curve. This curve is generated by adjusting the classifier output threshold to allow few or many putative hits. The figure of merit (FOM) is defined as the average keyword detection rate when the false alarm rate ranges from 1 to 10 false alarms per keyword per hour. The previous best FOM for this word spotter is 67.8% when trained using 24 male talkers and tested on 11 male talkers, and 65.9% when trained using 24 female talkers and tested on 11 female talkers. The overall FOM for all talkers is 66.3%. ATIVE HITS CONTINUOUS SPEECH INPUT Figure 1: NEURAL NETWORK WORDSPOTTER Block diagram of neural network word spotter. Using Voice Transfonnations to Create Additional Training Talkers for Word Spotting 877 3 TALKER VARIABILITY FOM scores of test talkers vary over a wide range. When training on 48 talkers and then performing testing on 22 talkers from the 70 conversations in the NIST Switchboard credit card database, the FOM ofthe test talkers varies from 16.7% to 100%. Most talkers perform well above 50%, but there are two female talkers with FOM's of 16.7% and 21.4%. The low FOM for individual speakers indicates a lack of training data with voice qualities that are similar to these test speakers. 4 CREATING MORE TRAINING DATA USING VOICE TRANSFORMATIONS Talker adaptation is difficult for word spotting because error rates are high and talkers often can not be prompted to verify adaptation phrases. Our approach to increasing performance across talkers uses voice transformation techniques to generate more varied training examples of keywords as shown in Figure 2. Other researchers have used talker transformation techniques to produce more natural synthesized speech (lwahashi, 1994, Mizuno, 1994), but using talker transformation techniques to generate more training data is novel. We have implemented a new voice transformation technique which utilizes the Sinusoidal Transform Analysis/Synthesis System (STS) described in (Quatieri, 1992). This technique attempts to transform one talker's speech pattern to that of a different talker. The STS generates a 512 point spectral envelope of the input speech 100 times a second and also separates pitch and voicing information. Separation of vocal tract characteristic and pitch information has allowed the implementation of pitch and time transformations in previous work (Quatieri, 1992). The system has been modified to generate and accept a spectral en- VOICE TRANSFORMATION SYSTEM ORIGINAL SPEECH TRANSFORMED SPEECH Figure 2: Generating more training data by artificially transforming original speech training data. 878 Eric I. Chang, Richard P. Lippmann velo.pe file fro.m an input speech sample. We info.nnally explo.red different techniques to' transfo.nn the spectral envelo.pe to' generate mo.re varied training examples by listening to' transfo.nned speech. This resulted in the fo.llo.wing algo.rithm that transfo.nns a talker's vo.ice by scaling the spectral envelo.pe o.f training talkers. Training co.nversatio.ns are upsampled fro.m 8000 Hz to' 10,000 Hz to' be co.mpatible with existing STS co.ding so.ftware. 1. 2. The STS system pro.cesses the upsampled files and generates a 512 point spectral envelo.pe o.f the input speech wavefo.rm at a frame rate o.f 100 frames a seco.nd and with a windo.w length o.f appro.ximately 2.5 times the length o.f each pitch perio.d. 3. A new spectral envelo.pe is generated by linearly expanding o.r co.mpressing the spectral axis. Each spectral po.int is identified by its index, ranging fro.m 0 to' 511. To. transfo.rm a spectral pro.file by 2, the new spectral value at frequency f is generated by averaging the spectral values aro.und the o.riginal spectral pro.file at frequency o.f 0.5 f The transfo.nnatio.n process is illustrated in Figure 3. In this figure, an o.riginal spectral envelo.pe is being expanded by two.. The spectral value at index 150 is thus transfo.rmed to. spectral index 300 in the new envelo.pe and the o.riginal spectral info.rmatio.n at high frequencies is lo.st. 4. The transfo.rmed spectral value is used to. resynthesize a speech wavefo.rm using the vocal tract excitatio.n info.nnatio.n extracted fro.m the o.riginal file. Vo.ice transfo.rmatio.n with the STS co.der allo.ws listening to. transfo.rmed speech but requires lo.ng co.mputatio.n. We simplified o.ur appro.ach to' o.ne o.f mo.difying the spectral scale in the spectral do.main directly within a mel-scale filterbank analysis pro.gram. The inco.ming speech sample is pro.cessed with an FFT to' calculate spectral magnitudes. Then spectral magnitudes are linearly transfo.rmed. Lastly mel-scale filtering is perfo.rmed with 10 linearly spaced filters up to' 1000 Hz and lo.garithmically spaced filters fro.m 1000 Hz up. A co.sine transfo.rm is then used to' generate mel-scaled cepstral values that are used by the wo.rdspo.tter. Much faster pro.cessing can be achieved by applying the spectral transfo.rmatio.n as part o.f the filterbank analysis. Fo.r example, while perfo.rming spectral transfo.nnatio.n using the STS algo.rithm takes up to' approximately 10 times real time, spectral transfo.rmatio.n within the mel-scale filterbank pro.gram can be acco.mplished within 1110 real time o.n a Sparc 10 wo.rkstatio.n. The rapid pro.cessing rate allo.ws o.n-line spectral transfo.rmatio.n. 5 WORD SPOTTING EXPERIMENTS Linear warping in the spectral do.main, which is used in the above algo.rithm, is co.rrect when the vo.cal tract is mo.delled as a series o.f lo.ssless aco.ustic tubes and the excitatio.n so.urce is at o.ne end o.f the vo.cal tract (Wakita, 1977). Wakita sho.wed that if the vo.cal tract is mo.delled as a series o.f equal length, lo.ssless, and co.ncatenated aco.ustic tubes, then the ratio. o.f the areas between the tubes determines the relative reso.nant frequencies o.f the vo.cal tract, while the o.verall length o.fthe vo.cal tract linearly scales fo.rmant frequencies. Preliminary research was co.nducted using linear scaling with spectral ratio.s ranging fro.m 0.6 to. 1.8 to. alter test utterances. After listening to. the STS transfo.rmed speech and also. o.bserving Using Voice Transformations to Create Additional Training Talkers for Word Spotting Original Spectral Envelope 879 III I I o 150 511 " Transformed Spectral Envelope o 300 511 Spectral Value Index Figure 3: An example of the spectral transformation algorithm where the original spectral envelope frequency scale is expanded by 2. spectrograms of the transformed speech, it was found that speech transformed using ratios between 0.9 and 1.1 are reasonably natural and can represent speech without introducing artifacts. Using discriminative training techniques such as FOM training carries the risk of overtraining the wordspotter on the training set and obtaining results that are poor on the testing set. To delay the onset of overtraining, we artificially transformed each training set conversation during each epoch using a different random linear transformation ratio. The transformation ratio used for each conversation is calculated using the following formula: ratio == a + N (0,0.06), where a is the transformation ratio that matches each training speaker to the average of the training set speakers, and N is a normally distributed random variable with a mean of 0.0 and standard deviation of 0.06. For each training conversation, the long term averages of formant frequencies for formant 1, 2, and 3 are calculated. A least square estimation is then performed to match the formant frequencies of each training set conversation to the group average formant frequencies. The transformation equation is described below: 880 Eric I. Chang, Richard P. Lippmann 100 ? 90 ? ? ? t ? :--! ~ 80 o - a- ?? .,. ? ? NORMAL 50 Figure 4: o 234 5 678 9 10 NUMBER OF EPOCHS OF FOM TRAINING Average detection accuracy (FOM) for the male training and testing set versus the number of epochs of FOM training. The transform ratio for each individual conversation is calculated to improve the naturalness of the transformed speech. In preliminary experiments, each conversation was transformed with fixed ratios of 0.9, 0.95, 1.05, and 1.1. However, for a speaker with already high formant frequencies, pushing the formant frequencies higher may make the transformed speech sound unnatural. By incorporating the individual formant matching ratio into the transformation ratio, speakers with high formant frequencies are not transformed to very high frequencies and speakers with low formant frequencies are not transformed to even lower formant frequency ranges. Male and female conversations from the NIST credit card database were used separately to train separate word spotters. Both the male and the female partition of data used 24 conversations for training and 11 conversations for testing. Keyword occurrences were extracted from each training conversation and used as the data for initialization of the neural network word spotter. Also, each training conversation was broken up into sentence length segments to be used for embedded reestimation in which the keyword models are joined with the filler models and the parameters of all the models are jointly estimated. After embedded reestimation, Figure of Merit training as described in (Lippmann. 1994) was performed for up to 10 epochs. During each epoch. each training conversation is transformed using a transform ratio randomly generated as described above. The performance of the word spotter after each iteration of training is evaluated on both the training set and the testing set. 6 WORD SPOTTING RESULTS Training and testing set FOM scores for the male speakers and the female speakers are shown in Figure 4 and Figure 5 respectively. The x axis plots the number of epochs of FOM training where each epoch represents presenting all 24 training conversations once. The FOM for word spotters trained with the normal training conversations and word spotters Using Voice Transformations to Create Additional Training Talkers for Word Spotting 881 100 90 -?j ~ - =I - -. " - ? .- au ? .. - ra- ". TRAIN ? 80 60 ~= ? \ 50 o NORMAL 23456789 10 NUMBER OF EPOCHS OF FOM TRAINING Figure 5: Average detection accuracy (FOM) for the female training and testing set versus the number of iterations of FOM training. trained with artificially expanded training conversations are shown in each plot. After the first epoch, the FOM improves significantly. With only the original training conversations (normal), the testing set FOM rapidly levels off while the training set FOM keeps on improving. When the training conversations are artificially expanded, the training set FOM is below the training set FOM from the normal training set due to more difficult training data. However, the testing set FOM continues to improve as more epochs of FOM training are performed. When comparing the FOM of wordspotters trained on the two sets of data after ten epochs of training, the FOM for the expanded set was 2.9 percentage points above the normal FOM for male speakers and 2.5 percentage points above the normal FOM for female speakers. For comparison, Carlson has reported that for a high performance word spotter on this database, doubling the amount of training data typically increases the FOM by 2 to 4 percentage points (Carlson, 1994). 7 SUMMARY Lack of training data has always been a constraint in training speech recognizers. This research presents a voice transformation technique which increases the variety among training talkers. The resulting more varied training set provided up to 2.9 percentage points of improvement in the figure of merit (average detection rate) of a high performance word spotter. This improvement is similar to the increase in performance provided by doubling the amount of training data (Carlson, 1994). This technique can also be applied to other speech recognition systems such as continuous speech recognition, talker identification, and isolated speech recognition. 882 Eric I. Chang. Richard P. Lippmcmn ACKNOWLEDGEMENT This work was sponsored by the Advanced Research Projects Agency. The views expressed are those of the authors and do not reflect the official policy or position of the U.S. Government. We wish to thank Tom Quatieri for providing his sinusoidal transform analysis/synthesis system. BIBLIOGRAPHY B. Carlson and D. Seward. (1994) Diagnostic Evaluation and Analysis of Insufficient and Task-Independent Training Data on Speech Recognition. In Proceedings Speech Research Symposium XlV, Johns Hopkins University. E. Chang and R. Lippmann. (1994) Figure of Merit Training for Detection and Spotting. In Neural Information Processing Systems 6, G. Tesauro, J. Cohen, and J. Alspector, (Eds.), Morgan Kaufmann: San Mateo, CA. H. Drucker, R. Schapire, and P. Simard. (1993) Improving Performance in Neural Networks Using a Boosting Algorithm. In Neural Information Processing Systems 5, S. Hanson, J. Cowan, and C. L. Giles, (Eds.), Morgan Kaufmann: San Mateo, California. N. Iwahashi and Y. Sagisaka. (1994) Speech Spectrum Transformation by Speaker Interpolation. In Proceedings International Conference on Acoustics Speech and Signal Processing, Vol. 1,461-464. R. Lippmann, E. Chang & C. Jankowski. (1994) Wordspotter Training Using Figure-ofMerit Back Propagation. In Proceedings of International Conference on Acoustics Speech and Signal Processing, Vol. 1,389-392. H. Mizuno and M. Abe. (1994) Voice Conversion Based on Piecewise Linear Conversion Rules of Formant Frequency and Spectrum Tilt. In Proceedings International Conference Oil Acoustics Speech and Signal Processing, Vol. 1,469-472. T. Quatieri and R. McAulay. (1992) Shape Invariant Time-Scale and Pitch Modification of Speech. In IEEE Trans. Signal Processing, vol 40, no 3. pp. 497-510. Hisashi Wakita. (1977) Normalization of Vowels by Vocal-Tract Length and Its Application to Vowel Identification. In IEEE Trans. Acoustics, Speech, and Signal Processing, vol. ASSP-25, No.2., pp. 183-192.
950 |@word nd:1 llo:1 perfo:2 carry:1 series:2 score:4 past:1 existing:2 comparing:1 wakita:3 john:1 partition:1 shape:1 plot:2 sponsored:1 provides:1 boosting:1 unacceptable:1 symposium:1 ra:1 rapid:1 alspector:1 ming:1 increasing:1 provided:2 project:1 developed:1 lexington:1 transformation:21 temporal:1 transfonnations:1 classifier:2 demonstrates:1 hit:4 rm:4 filterbank:3 scaled:1 normally:1 ice:2 interpolation:1 approximately:1 initialization:1 au:1 mateo:2 co:10 range:3 testing:9 block:1 spot:1 digit:2 area:1 significantly:1 matching:1 word:24 radial:1 vocal:3 upsampled:2 cal:5 risk:1 applying:1 duration:1 rule:1 his:1 spontaneous:2 target:3 quatieri:4 user:6 us:2 expensive:2 recognition:9 continues:1 database:3 ding:1 calculate:1 keyword:5 transforming:3 und:1 broken:1 agency:1 trained:6 segment:1 algo:3 eric:5 inco:1 basis:1 po:1 represented:1 train:2 larger:1 seco:1 formant:11 transform:5 jointly:1 sequence:1 aro:1 adaptation:4 rapidly:1 lincoln:1 ach:1 produce:1 generating:1 tract:8 keywords:2 implemented:1 differ:1 filter:2 human:1 government:1 preliminary:2 credit:3 normal:7 presumably:1 viterbi:1 sho:1 talker:40 mo:4 vary:1 recognizer:1 estimation:1 linguistically:1 nant:1 create:5 successfully:1 hope:1 mit:3 always:1 modified:1 appro:2 linguistic:2 improvement:2 indicates:1 baseline:1 nn:1 typically:1 accept:1 hidden:1 w:2 expand:1 transformed:11 overall:2 among:1 art:1 equal:1 once:1 enlarge:3 ng:1 represents:1 alter:1 piecewise:1 richard:5 few:2 randomly:1 resulted:1 individual:4 vowel:2 attempt:2 detection:7 evaluation:1 male:8 analyzed:1 re:1 isolated:1 increased:1 giles:1 phrase:2 introducing:1 deviation:1 delay:1 reported:1 varies:2 nns:1 st:1 international:3 off:1 decoding:1 synthesis:2 excitatio:2 hopkins:1 reflect:1 tube:3 creating:1 wing:1 style:2 simard:1 sinusoidal:2 hisashi:1 includes:1 int:1 sts:7 onset:1 sine:1 performed:3 view:1 red:1 ative:1 square:1 accuracy:2 kaufmann:2 who:2 characteristic:1 spaced:2 ofthe:1 identification:2 researcher:1 overtraining:2 wed:1 fo:3 ed:2 frequency:16 pp:2 adjusting:1 conversation:18 improves:1 back:1 higher:1 tom:1 evaluated:1 lastly:1 recite:1 lack:2 propagation:1 accent:1 artifact:1 quality:1 oil:1 usa:1 normalized:1 verify:1 laboratory:1 illustrated:1 ll:2 during:2 speaker:14 mel:4 presenting:1 vo:7 interface:1 pro:8 image:3 ranging:2 novel:1 aco:2 cohen:1 tilt:1 synthesized:1 cessing:2 recognizers:4 base:2 xlv:1 female:9 sparc:1 tesauro:1 phonetic:2 accomplished:1 der:1 morgan:2 additional:4 spectrogram:1 signal:5 sound:2 faster:1 adapt:1 match:2 long:1 pitch:5 metric:1 iteration:2 represent:1 normalization:1 achieved:1 background:2 separately:1 diagram:1 envelope:4 rming:1 file:5 hz:4 cowan:1 iii:1 fft:1 variety:1 identified:1 drucker:2 listening:3 unnatural:1 wo:2 speech:43 dramatically:1 sst:2 unimportant:1 amount:3 ten:1 generate:7 schapire:1 percentage:8 jankowski:1 estimated:1 diagnostic:1 per:2 vol:5 group:1 threshold:3 ce:1 spotter:14 year:1 utilizes:1 putative:3 separation:1 scaling:2 constraint:1 bibliography:1 generates:2 performing:1 expanded:5 optical:2 relatively:1 poor:2 across:3 character:3 ur:1 modification:1 outlier:4 invariant:2 equation:1 fom:29 merit:4 rmed:6 end:1 naturalness:1 ocr:1 spectral:34 voicing:1 occurrence:1 alternative:1 voice:12 original:5 carlson:4 pushing:1 warping:5 rmatio:5 already:2 fthe:1 distance:1 separate:2 card:3 thank:1 hmm:1 length:6 tter:1 index:4 prompted:2 ratio:12 providing:1 insufficient:1 difficult:7 reso:1 info:3 implementation:1 policy:1 perform:1 conversion:2 markov:1 nist:2 variability:1 assp:1 frame:3 varied:3 abe:1 backtrace:1 sentence:1 hanson:1 california:1 acoustic:4 hour:1 iwahashi:1 trans:2 spotting:11 below:2 pattern:3 including:2 natural:2 hybrid:1 advanced:1 improve:3 ne:2 axis:2 rpl:1 fro:6 utterance:1 epoch:11 acknowledgement:1 relative:1 embedded:2 fully:1 allo:2 filtering:1 versus:3 gather:1 switchboard:1 plotting:1 normalizes:1 lo:5 summary:1 allow:1 wide:1 cepstral:1 riginal:4 distributed:1 curve:2 dimension:2 calculated:3 gram:2 wordspotter:3 author:1 san:2 simplified:1 lippmann:7 keep:1 reestimation:2 corpus:1 resynthesize:1 discriminative:1 spectrum:4 continuous:3 reasonably:1 expanding:1 ca:1 obtaining:1 improving:2 artificially:6 official:1 main:2 linearly:4 whole:2 alarm:3 allowed:1 en:1 rithm:3 n:1 position:1 wish:1 pe:7 formula:1 enlarging:1 explored:3 incorporating:1 false:3 magnitude:2 backpropagate:1 expressed:1 doubling:3 joined:1 chang:8 determines:1 extracted:2 ma:1 rbf:1 averaging:1 explo:1 filler:2 tested:2
7,190
951
Reinforcement Learning Algorithm for Partially Observable Markov Decision Problems Tommi Jaakkola tommi@psyche.mit.edu Satinder P. Singh singh@psyche.mit.edu Michael I. Jordan jordan@psyche.mit.edu Department of Brain and Cognitive Sciences, BId. E10 Massachusetts Institute of Technology Cambridge, MA 02139 Abstract Increasing attention has been paid to reinforcement learning algorithms in recent years, partly due to successes in the theoretical analysis of their behavior in Markov environments. If the Markov assumption is removed, however, neither generally the algorithms nor the analyses continue to be usable. We propose and analyze a new learning algorithm to solve a certain class of non-Markov decision problems. Our algorithm applies to problems in which the environment is Markov, but the learner has restricted access to state information. The algorithm involves a Monte-Carlo policy evaluation combined with a policy improvement method that is similar to that of Markov decision problems and is guaranteed to converge to a local maximum. The algorithm operates in the space of stochastic policies, a space which can yield a policy that performs considerably better than any deterministic policy. Although the space of stochastic policies is continuous-even for a discrete action space-our algorithm is computationally tractable. 346 1 Tommi Jaakkola, Satinder P. Singh, Michaell. Jordan INTRODUCTION Reinforcement learning provides a sound framework for credit assignment in unknown stochastic dynamic environments. For Markov environments a variety of different reinforcement learning algorithms have been devised to predict and control the environment (e.g., the TD(A) algorithm of Sutton, 1988, and the Q-Iearning algorithm of Watkins, 1989). Ties to the theory of dynamic programming (DP) and the theory of stochastic approximation have been exploited, providing tools that have allowed these algorithms to be analyzed theoretically (Dayan, 1992; Tsitsiklis, 1994; Jaakkola, Jordan, & Singh, 1994; Watkins & Dayan, 1992). Although current reinforcement learning algorithms are based on the assumption that the learning problem can be cast as Markov decision problem (MDP), many practical problems resist being treated as an MDP. Unfortunately, if the Markov assumption is removed examples can be found where current algorithms cease to perform well (Singh, Jaakkola, & Jordan, 1994). Moreover, the theoretical analyses rely heavily on the Markov assumption. The non-Markov nature of the environment can arise in many ways. The most direct extension of MDP's is to deprive the learner of perfect information about the state of the environment. Much as in the case of Hidden Markov Models (HMM's), the underlying environment is assumed to be Markov, but the data do not appear to be Markovian to the learner. This extension not only allows for a tractable theoretical analysis, but is also appealing for practical purposes. The decision problems we consider here are of this type. The analog of the HMM for control problems is the partially observable Markov decision process (POMDP; see e.g., Monahan, 1982). Unlike HMM's, however, there is no known computationally tractable procedure for POMDP's. The problem is that once the state estimates have been obtained, DP must be performed in the continuous space of probabilities of state occupancies, and this DP process is computationally infeasible except for small state spaces. In this paper we describe an alternative approach for POMDP's that avoids the state estimation problem and works directly in the space of (stochastic) control policies. (See Singh, et al., 1994, for additional material on stochastic policies.) 2 PARTIAL OBSERVABILITY A Markov decision problem can be generalized to a POMDP by restricting the state information available to the learner. Accordingly, we define the learning problem as follows. There is an underlying MDP with states S {SI, S2, ... , SN} and transition probability pOII! " the probability of jumping from state S to state s' when action a is taken in state s. For every state and every action a (random) reward is provided to the learner. In the POMDP setting, the learner is not allowed to observe the state directly but only via messages containing information about the state. At each time step t an observable message mt is drawn from a finite set of messages according to an unknown probability distribution P(mlst) 1. We assume that the learner does = 1 For simplicity we assume that this distribution depends only on the current state. The analyses go through also with distributions dependent on the past states and actions Reinforcement Learning Algorithm for Markov Decision Problems 347 not possess any prior information about the underlying MDP beyond the number of messages and actions. The goal for the learner is to come up with a policy-a mapping from messages to actions-that gives the highest expected reward. As discussed in Singh et al. (1994), stochastic policies can yield considerably higher expected rewards than deterministic policies in the case of POMDP's. To make this statement precise requires an appropriate technical definition of "expected reward," because in general it is impossible to find a policy, stochastic or not, that maximizes the expected reward for each observable message separately. We take the timeaverage reward as a measure of performance, that is, the total accrued reward per number of steps taken (Bertsekas, 1987; Schwartz, 1993). This approach requires the assumption that every state of the underlying controllable Markov chain is reachable. In this paper we focus on a direct approach to solving the learning problem. Direct approaches are to be compared to indirect approaches, in which the learner first identifies the parameters of the underlying MDP, and then utilizes DP to obtain the policy. As we noted earlier, indirect approaches lead to computationally intractable algorithms. Our approach can be viewed as providing a generalization of the direct approach to MDP's to the case of POMDP's. 3 A MONTE-CARLO POLICY EVALUATION Advantages of Monte-Carlo methods for policy evaluation in MDP's have been reviewed recently (Barto and Duff, 1994). Here we present a method for calculating the value of a stochastic policy that has the flavor of a Monte-Carlo algorithm. To motivate such an approach let us first consider a simple case where the average reward is known and generalize the well-defined MDP value function to the POMDP setting. In the Markov case the value function can be written as (cf. Bertsekas, 1987): N V(s) = N-oo~ lim ~ E{R(st, Ut) - Risl = s} (1) t=l where St and at refer to the state and the action taken at the tth step respectively. This form generalizes easily to the level of messages by taking an additional expectation: (2) V(m) = E {V(s)ls -+ m} where s -+ m refers to all the instances where m is observed in sand E{?ls -+ m} is a Monte-Carlo expectation. This generalization yields a POMDP value function given by (3) V(m) = P(slm)V(s) I: ,Em in which P(slm) define the limit occupancy probabilities over the underlying states for each message m. As is seen in the next section value functions of this type can be used to refine the currently followed control policy to yield a higher average reward. Let us now consider how the generalized value functions can be computed based on the observations. We propose a recursive Monte-Carlo algorithm to effectively compute the averages involved in the definition of the value function. In the simple 348 Tommi Jaakkola, Satinder P. Singh, Michael I. Jordan case when the average payoff is known this algorithm is given by f3t(m) Xt(m) Xt(m) (1 - Kt(m) htf3t-l(m) + Kt(m) lIt(m) Xt(m) (1- Kt(m))lIt-l(m) + f3t(m)[R(st, at) - (4) R] (5) where Xt(m) is the indicator function for message m, Kt(m) is the number of times m has occurred, and 'Yt is a discount factor converging to one in the limit. This algorithm can be viewed as recursive averaging of (discounted) sample sequences of different lengths each of which has been started at a different occurrence of message m . This can be seen by unfolding the recursion, yielding an explicit expression for lit (m). To this end, let tk denote the time step corresponding to the ph occurrence of message m and for clarity let Rt = R(st, Ut) - R for every t. Using these the recursion yields: 1 lIt(m) = Kt(m) [ Rtl + rl,l R t1 +1 + ... + r1,t-tl Rt (6) where we have for simplicity used rk ,T to indicate the discounting at the Tth step in the kth sequence. Comparing the above expression to equation 1 indicates that the discount factor has to converge to one in the limit since the averages in V(s) or V(m) involve no discounting . To address the question of convergence of this algorithm let us first assume a constant discounting (that is, 'Yt 'Y < 1). In this case, the algorithm produces at best an approximation to the value function. For large K(m) the convergence rate by which this approximate solution is found can be characterized in terms of the bias and variance. This gives Bias{V(m)} Q( (1 - r)-l / K(m) and Var{V(m)} Q( (1 r)-2 / K(m) where r = Ehtk-tk-l} is the expected effective discounting between observations. Now, in order to find the correct value function we need an appropriate way of letting 'Yt - 1 in the limit. However, not all such schedules lead to convergent algorithms; setting 'Yt = 1 for all t , for example, would not. By making use of the above bounds a feasible schedule guaranteeing a vanishing bias and variance can be found . For instance, since 'Y > 7 we can choose 'Yk(m) = 1 - K(m)I/4. Much faster schedules are possible to obtain by estimating r. = Let us now revise the algorithm to take into account the fact that the learner in fact has no prior knowledge of the average reward. In this case the true average reward appearing in the above algorithm needs to be replaced with an incrementally updated estimate R t - l . To improve the effect this changing estimate has on the values we transform the value function whenever the estimate is updated. This transformation is given by Xt(m) (1 - Kt(m) )Ct-1(m) + f3t(m) (7) (8) lIt(m) - Ct(m)(Rt - Rt-l) and, as a result, the new values are as if they had been computed using the current estimate of the average reward. Reinforcement Learning Algorithm for Markov Decision Problems 349 To carry these results to the control setting and assign a figure of merit to stochastic policies we need a quantity related to the actions for each observed message. As in the case of MDP's, this is readily achieved by replacing m in the algorithm just described by (m, a). In terms of equation 6, for example, this means that the sequences started from m are classified according to the actions taken when m is observed. The above analysis goes through when m is replaced by (m, a), yielding "Q-values" on the level of messages: (9) In the next section we show how these values can be used to search efficiently for a better policy. 4 POLICY IMPROVEMENT THEOREM Here we present a policy improvement theorem that enables the learner to search efficiently for a better policy in the continuous policy space using the "Q-values" Q(m, a) described in the previous section. The theorem allows the policy refinement to be done in a way that is similar to policy improvement in a MDP setting. Theorem 1 Let the current stochastic policy 1I"(alm) lead to Q-values Q1r(m, a) on the level of messages. For any policy 11"1 (aim) define J1r 1 (m) = L: 11"1 (alm)[Q1r(m, a) - V 1r (m)] a The change in the average reward resulting from changing the current policy according to 1I"(alm) -+ (1- {)1I"(alm) + {1I"1(alm) is given by ~R1r = {L: p1r(m)J1r 1 (m) + O({2) m where p1r (m) are the occupancy probabilities for messages associated with the current policy. The proof is given in Appendix. In terms of policy improvement the theorem can be interpreted as follows . Choose the policy 1I"1(alm) such that 1 J1r (m) = max[Q1r(m, a) a V 1r (m)] (10) If now J1r 1 (m) > 0 for some m then we can change the current policy towards 11"1 and expect an increase in the average reward as shown by the theorem . The { factor suggests local changes in the policy space and the policy can be refined until m~l J1r\ m) 0 for all m which constitutes a local maximum for this policy improvement method. Note that the new direction 1I"1(alm) in the policy space can be chosen separately for each m. = 5 THE ALGORITHM Based on the theoretical analysis presented above we can construct an algorithm that performs well in a POMDP setting. The algorithm is composed of two parts: First, 350 Tommi Jaakkola, Satinder P. Singh, Michael I. Jordan Q(m, a) values-analogous to the Q-values in MDP-are calculated via a MonteCarlo approach. This is followed by a policy improvement step which is achieved by increasing the probability of taking the best action as defined by Q(m,a). The new policy is guaranteed to yield a higher average reward (see Theorem 1) as long as for somem max[Q(m, a) - V(m)] > 0 (11) a This condition being false constitutes a local maximum for the algorithm. Examples illustrating that this indeed is a local maximum can be found fairly easily. In practice, it is not feasible to wait for the Monte-Carlo policy evaluation to converge but to try to improve the policy before the convergence. The policy can be refined concurrently with the Monte-Carlo method according to 1I"(almn) -+ 1I"(almn) + ?[Qn(m n, a) - Vn(m n)] (12) with normalization. Other asynchronous or synchronous on-online updating schemes can also be used. Note that if Qn(m, a) = Q(m, a) then this change would be statistically equivalent to that of the batch version with the concomitant guarantees of giving a higher average reward. 6 CONCLUSIONS In this paper we have proposed and theoretically analyzed an algorithm that solves a reinforcement learning problem in a POMDP setting, where the learner has restricted access to the state of the environment. As the underlying MDP is not known the problem appears to the learner to have a non-Markov nature. The average reward was chosen as the figure of merit for the learning problem and stochastic policies were used to provide higher average rewards than can be achieved with deterministic policies. This extension from MDP's to POMDP's greatly increases the domain of potential applications of reinforcement learning methods. The simplicity of the algorithm stems partly from a Monte-Carlo approach to obtaining action-dependent values for each message. These new "Q-values" were shown to give rise to a simple policy improvement result that enables the learner to gradually improve the policy in the continuous space of probabilistic policies. The batch version of the algorithm was shown to converge to a local maximum. We also proposed an on-line version of the algorithm in which the policy is changed concurrently with the calculation of the "Q-values." The policy improvement of the on-line version resembles that of learning automata. APPENDIX Let us denote the policy after the change by 11"!. Assume first that we have access to Q1I"(s, a), the Q-values for the underlying MDP, and to p1l"< (slm), the occupancy probabilities after the policy refinement. Define J(m, 11"!, 11"!, 11") =L 1I"!(alm) L p1l"< (slm)[Q1I"(s, a) - V1I"(s)] (13) where we have used the notation that the policies on the left hand side correspond to the policies on the right respectively. To show how the average reward depends 351 Reinforcement Learning Algorithm for Markov Decision Problems on this quantity we need to make use of the following facts. The Q-values for the underlying MDP satisfy (Bellman's equation) Q1r(s, a) = R(s, a) - R1r + L:P~3' V 1r (s') (14) " In addition, 2:0 7r(alm)Q1r(s, a) = V 1r (s), implying that J(m, 7r!, 7r!, 7r f ) = 0 (see eq. 13). These facts allow us to write J(m, 7r f , 7r f , 7r) - J(m, 7r f , 7r f , 7r f ) L: 7r f (alm) L: p 1r ?slm)[Q1r(s, a) - V 1r (s) - Q1r< (s, a) + V 1r <(s)] (15) 3 By weighting this result for each class by p 1r < (m) and summing over the messages the probability weightings for the last two terms become equal and the terms cancel. This procedure gives us R1r < - R1r = L:P1r?m)J(m, 7r f ,7r f , 7r) (16) m This result does not allow the learner to assess the effect of the policy refinement on the average reward since the JO term contains information not available to the learner. However, making use of the fact that the policy has been changed only slightly this problem can be avoided. As 7r! is a policy satisfying maxmo l7r f (alm) -7r(alm)1 ::; (, it can then be shown that there exists a constant C such that the maximum change in P(slm), pes), P(m) is bounded by Cf. Using these bounds and indicating the difference between 7r! and 7r dependent quantities by ~ we get L:[7r(alm) + ~7r(alm)] L)P1r(slm) + ~p1r(slm)][Q1r(s, a) - V 1r (s)] o o 3 o 3 where the second equality follows from third from the bounds stated earlier. 2:0 7r(alm)[Q1r(s, a) - V1r(s)] = 0 and the The equation characterizing the change in the average reward due to the policy change (eq. 16) can be now rewritten as follows: R1r< - R1r = L: p 1r <(m)J(m, 7r f , 7r, 7r) + 0?(2) m 352 Tommi Jaakkola, Satinder P. Singh, Michael I. Jordan m a where the bounds (see above) have been used for Vir' (m) - p1r(m). This completes 0 the proof. Acknowledgments The authors thank Rich Sutton for pointing out errors at early stages of this work. This project was supported in part by a grant from the McDonnell-Pew Foundation, by a grant from ATR Human Information Processing Research Laboratories, by a grant from Siemens Corporation and by grant NOOOI4-94-1-0777 from the Office of Naval Research. Michael I. Jordan is a NSF Presidential Young Investigator. References Barto, A., and Duff, M. (1994). Monte-Carlo matrix inversion and reinforcement learning. In Advances of Neural Information Processing Systems 6, San Mateo, CA, 1994. Morgan Kaufmann. Bertsekas, D. P. (1987). Dynamic Programming: Deterministic and Stochastic Models. Englewood Cliffs, NJ: Prentice-Hall. Dayan, P. (1992). The convergence of TD(A) for general A. Machine Learning, 8, 341-362. Jaakkola, T., Jordan M.I., and Singh, S. P. (1994). On the convergence of stochastic iterative Dynamic Programming algorithms. Neural Computation 6, 1185-1201. Monahan, G. (1982). A survey of partially observable Markov decision processes. Management Science, 28, 1-16. Singh, S. P., Jaakkola, T., Jordan, M.1. (1994). Learning without state estimation in partially observable environments. In Proceedings of the Eleventh Machine Learning Conference. Sutton, R. S. (1988). Learning to predict by the methods of temporal differences. Machine Learning, 3, 9-44. Schwartz, A. (1993). A reinforcement learning method for maximizing un discounted rewards. In Proceedings of the Tenth Machine Learning Conference. Tsitsiklis J. N. (1994). Asynchronous stochastic approximation and Q-Iearning. Machine Learning 16, 185-202. Watkins, C.J.C.H. (1989). Learning from delayed rewards. PhD Thesis, University of Cambridge, England. Watkins, C.J .C.H, & Dayan, P. (1992). Q-Iearning. Machine Learning, 8, 279-292.
951 |@word illustrating:1 version:4 inversion:1 paid:1 carry:1 contains:1 past:1 current:8 comparing:1 si:1 must:1 readily:1 written:1 enables:2 implying:1 v1r:1 accordingly:1 vanishing:1 provides:1 direct:4 become:1 eleventh:1 alm:15 theoretically:2 indeed:1 expected:5 behavior:1 nor:1 brain:1 bellman:1 discounted:2 td:2 increasing:2 provided:1 estimating:1 moreover:1 underlying:9 maximizes:1 notation:1 bounded:1 project:1 interpreted:1 transformation:1 corporation:1 nj:1 guarantee:1 temporal:1 every:4 iearning:3 tie:1 slm:8 schwartz:2 control:5 vir:1 grant:4 appear:1 bertsekas:3 t1:1 before:1 local:6 limit:4 sutton:3 cliff:1 resembles:1 mateo:1 suggests:1 statistically:1 practical:2 acknowledgment:1 recursive:2 practice:1 procedure:2 refers:1 wait:1 get:1 prentice:1 impossible:1 equivalent:1 deterministic:4 yt:4 maximizing:1 go:2 attention:1 l:2 automaton:1 pomdp:12 survey:1 simplicity:3 analogous:1 updated:2 heavily:1 programming:3 satisfying:1 updating:1 observed:3 removed:2 highest:1 yk:1 environment:10 reward:23 dynamic:4 motivate:1 singh:12 solving:1 learner:16 easily:2 indirect:2 describe:1 effective:1 monte:10 refined:2 solve:1 r1r:6 presidential:1 transform:1 online:1 advantage:1 sequence:3 propose:2 convergence:5 r1:1 produce:1 perfect:1 guaranteeing:1 tk:2 oo:1 eq:2 solves:1 involves:1 come:1 indicate:1 tommi:6 direction:1 correct:1 stochastic:15 human:1 material:1 sand:1 assign:1 timeaverage:1 generalization:2 extension:3 credit:1 hall:1 mapping:1 predict:2 pointing:1 early:1 purpose:1 estimation:2 currently:1 v1i:1 tool:1 unfolding:1 mit:3 concurrently:2 aim:1 barto:2 jaakkola:9 office:1 focus:1 naval:1 improvement:9 indicates:1 greatly:1 dayan:4 dependent:3 hidden:1 l7r:1 fairly:1 equal:1 once:1 construct:1 lit:5 cancel:1 constitutes:2 composed:1 delayed:1 replaced:2 message:17 englewood:1 evaluation:4 analyzed:2 yielding:2 chain:1 kt:6 partial:1 jumping:1 theoretical:4 instance:2 earlier:2 markovian:1 assignment:1 f3t:3 considerably:2 combined:1 st:4 accrued:1 probabilistic:1 michael:5 jo:1 thesis:1 management:1 containing:1 choose:2 cognitive:1 usable:1 account:1 potential:1 satisfy:1 depends:2 performed:1 try:1 analyze:1 ass:1 variance:2 kaufmann:1 efficiently:2 yield:6 correspond:1 generalize:1 carlo:10 classified:1 whenever:1 definition:2 involved:1 associated:1 proof:2 massachusetts:1 revise:1 noooi4:1 lim:1 ut:2 knowledge:1 schedule:3 appears:1 higher:5 done:1 just:1 stage:1 until:1 hand:1 replacing:1 incrementally:1 mdp:16 effect:2 true:1 discounting:4 equality:1 laboratory:1 noted:1 generalized:2 performs:2 recently:1 mt:1 rl:1 analog:1 discussed:1 occurred:1 refer:1 cambridge:2 pew:1 had:1 reachable:1 access:3 recent:1 certain:1 continue:1 success:1 exploited:1 seen:2 morgan:1 additional:2 q1r:9 converge:4 sound:1 stem:1 technical:1 faster:1 characterized:1 calculation:1 england:1 long:1 devised:1 converging:1 expectation:2 normalization:1 achieved:3 addition:1 separately:2 completes:1 unlike:1 posse:1 jordan:11 bid:1 variety:1 observability:1 synchronous:1 expression:2 action:11 generally:1 involve:1 discount:2 ph:1 tth:2 nsf:1 per:1 discrete:1 write:1 drawn:1 clarity:1 changing:2 neither:1 tenth:1 year:1 j1r:5 vn:1 utilizes:1 decision:11 appendix:2 bound:4 ct:2 guaranteed:2 followed:2 convergent:1 refine:1 department:1 according:4 mcdonnell:1 slightly:1 em:1 psyche:3 appealing:1 making:2 restricted:2 gradually:1 taken:4 computationally:4 equation:4 montecarlo:1 letting:1 merit:2 tractable:3 end:1 available:2 generalizes:1 rewritten:1 observe:1 appropriate:2 occurrence:2 appearing:1 alternative:1 batch:2 cf:2 calculating:1 giving:1 question:1 quantity:3 rt:4 dp:4 kth:1 thank:1 atr:1 hmm:3 length:1 providing:2 concomitant:1 unfortunately:1 statement:1 stated:1 rise:1 policy:56 unknown:2 perform:1 observation:2 p1l:2 markov:22 finite:1 payoff:1 precise:1 duff:2 michaell:1 cast:1 resist:1 address:1 beyond:1 max:2 treated:1 rely:1 indicator:1 recursion:2 occupancy:4 improve:3 scheme:1 technology:1 identifies:1 started:2 sn:1 prior:2 expect:1 var:1 foundation:1 changed:2 supported:1 last:1 asynchronous:2 infeasible:1 tsitsiklis:2 bias:3 side:1 allow:2 institute:1 taking:2 characterizing:1 calculated:1 transition:1 avoids:1 rich:1 qn:2 author:1 monahan:2 reinforcement:12 refinement:3 avoided:1 san:1 approximate:1 observable:6 rtl:1 satinder:5 summing:1 assumed:1 continuous:4 search:2 iterative:1 un:1 reviewed:1 nature:2 ca:1 controllable:1 obtaining:1 domain:1 s2:1 arise:1 allowed:2 tl:1 q1i:2 explicit:1 pe:1 watkins:4 weighting:2 third:1 young:1 rk:1 theorem:7 xt:5 cease:1 intractable:1 exists:1 restricting:1 false:1 effectively:1 phd:1 flavor:1 partially:4 applies:1 ma:1 goal:1 viewed:2 towards:1 feasible:2 change:8 except:1 operates:1 averaging:1 total:1 partly:2 e10:1 siemens:1 indicating:1 investigator:1
7,191
952
Hyperparameters, Evidence and Generalisation for an Unrealisable Rule Glenn Marion and David Saad glennyGed.ac.uk, D.SaadGed.ac.uk Department of Physics, University of Edinburgh, Edinburgh, EH9 3JZ, U.K. Abstract Using a statistical mechanical formalism we calculate the evidence, generalisation error and consistency measure for a linear perceptron trained and tested on a set of examples generated by a non linear teacher. The teacher is said to be unrealisable because the student can never model it without error. Our model allows us to interpolate between the known case of a linear teacher, and an unrealisable, nonlinear teacher. A comparison of the hyperparameters which maximise the evidence with those that optimise the performance measures reveals that, in the non-linear case, the evidence procedure is a misleading guide to optimising performance. Finally, we explore the extent to which the evidence procedure is unreliable and find that, despite being sub-optimal, in some circumstances it might be a useful method for fixing the hyperparameters. 1 INTRODUCTION The analysis of supervised learning or learning from examples is a major field of research within neural networks. In general, we have a probabilistic 1 teacher, which maps an N dimensional input vector x to output Yt(x) according to some distribution P(Yt I x). We are supplied with a data set v= ({Yt(x lJ ), xlJ} : J.' = l..p) generated from P(Yt I x) by independently sampling the input distribution, P(x), p times. One attempts to optimise a model mapping (a student), parameterised by lThis accommodates teachers with deterministic output corrupted by noise. 256 Glenn Marion, David Saad some vector w, with respect to the underlying teacher. The training error Ew (V) is some measure of the difference between the student and the teacher outputs over the set V. Simply minimising the training error leads to the problem of over-fitting. In order to make successful predictions out-with the set V it is essential to have some prior preference for particular rules. Occams razor is an expression of our preference for the simplest rules which account for the data. Clearly Ew(V) is an unsatisfactory performance measure since it is limited to the training examples. Very often we are interested in the students ability to model a random example drawn from P(Yt I x)P(x), but not necessarily in the training set, one measure of this performance is the generalisation error. It is also desirable to predict, or estimate, the level of this error. The teacher is said to be an unrealisable rule, for the student in question, if the minimum generalisation error is non-zero. One can consider the Supervised Learning Paradigm within the context of Bayesian Inference. In particular MacKay [MacKay 92(a)] advocates the evidence procedure as a 'principled' method which, in some situations, does seem to improve performance [Thodberg 93]. However, in others, as MacKay points out the evidence procedure can be misleading [MacKay 92(b )]. In this paper we do not seek to comment on the validity of of the evidence procedure as an approximation to Hierarchical Bayes (see for example [Wolpert and Strauss 94]). Rather, we ask which performance measures do we seek to optimise and under what conditions will the evidence procedure optimise them? Theoretical results have been obtained for a linear percept ron trained on data produced by a linear perceptron [Bruce and Saad 94]. They suggest that the evidence procedure is a useful guide to optimising the learning algorithm's performance. In what follows we examine the evidence procedure for the case of a linear perceptron learning a non linear teacher. In the next section we review the Bayesian scheme and define the evidence and the relevant performance measures. In section 3 we introduce our student and teacher and discuss the calculation. Finally, in section 4 we examine the extent to which the evidence procedure optimises performance. 2 BAYESIAN FORMALISM 2.1 THE EVIDENCE If we take Ew(V) to be the usual sum squared error and assume that our data is corrupted by Gaussian noise with variance 1/2/3 then the probability, or likelihood, ofthe data(V) being produced given the model wand /3 is P(D 1/3, w) ex: e-~Ew(1)). In order to incorporate Occams Razor we also assume a prior distribution on the teacher rules, that is, we believe a priori in some rules more strongly than others. Specifically we believe that pew I ,) ex: e-"'(C(w). MUltiplying the likelihood by the prior we obtain the post training or student distribution 2 P( w I V", /3) ex: e-~Ew(1))-''YC(w). It is clear that the most probable model w? is given by minimising the composite cost function /3Ew(V)+,C(w) with respect to the weights (w). This formalises the trade off between fitting the data and minimising student complexity. In this sense the Bayesian viewpoint coincides with the usual backprop standpoint. 2Integrating this over f3 and 'Y gives us the posterior P(w I 1?. Hyperparameters, Evidence and Generalisation for an Unrealisable Rule 257 In fact, it should be noted that stochastic minimisation can also give rise to the same post training distribution [Seung et aI92). The parameters (3 and, are known as the hyperparameters. Here we consider C(w) = wtw in which case, is termed the weight decay. The evidence is the normalisation constant in the above expression for the post training distribution. P('D I 'Y,(3) = Jn dWjP('D I (3, w)P(w I,) J That is, the probability of the data set ('D) given the hyperparameters. The evidence procedure fixes the hyperparameters to the values that maximise this probability. 2.2 THE PERFORMANCE MEASURES Many performance measures have been introduced in the literature (See e.g., [Krogh and Hertz 92) and [Seung et aI92)) . Here, we consider the squared difference between the average (over the post training distribution) of the student output (y.(x)}w and that of the teacher, Yt(x) , averaged over all possible test questions and teacher outputs, P(Yt, x) and finally over all possible sets of data, 'D. fg = ((Yt(x) - (Y. (x?)w ?}P(X,Yf).'l) This is equivalent to the generalisation error given by Krogh and Hertz. Another factor we can consider is the variance of the output over the student distribution ({y.(x) - (y.(x)}wP}w,P(x)' This gives us a measure of the confidence we should have in our post training distribution and could possibly be calculated if we could estimate the input distribution P(x). Here we extend Bruce and Saad's definition [Bruce and Saad 94] of the consistency measure Dc to include unrealisable = IiIIlp_oo fg, rules by adding the asymptotic error fr: Dc = ({y.(x) - (y.(x)}w}2}w,p(x),'P - fg + fr;' We regard Dc = 0 as optimal since then the variance over our student distribution is an accurate prediction of the decaying part of the generalisation error. We can consider both these performance measures as objective functions measuring the students ability to mimic the underlying teacher. Clearly, they can only be calculated in theory and perhaps, estimated in practice. In contrast, the evidence is only a function of our assumptions and the data and the evidence procedure is, therefore, a practical method of setting the hyperparameters. 3 THE MODEL In our model the student is simply a linear perceptron. The output for an input vector xl' is given by = w .xl' / v'N. The examples, against which the student is trained and tested, are produced by sampling the input distribution, P(x) and then generating outputs from the distribution, Y: P(Yt I x) = t 0=1 P(y~ I x, O)P(x I O)PA 2:0=1 P(O)P(x I 0) 258 Glenn Marion, David Saad -1.0 I I I -0.6 0.0 0.6 1.0 ? Figure 1: A 2-teacher in 1D : The average output (Yt}P(yl%') (i) for Dw = 0 , (ii) for Dw > 0 (0'%'1 = 0'%',) and (iii) with Dw > 0 (0'%'1 ?- 0'%,,) . where P(Yt I x, n) <X exp([Yt - wn.xF /20'2), P(x I n) is N(an,O',;o) 3 and PA is chosen such that I:~=l PA=1. Thus, each component in the sum is a linear perceptron, whose output is corrupted by Gaussian noise of variance 0'2, and we refer to this teacher as an n-teacher. In what follows , for simplicity, we consider a two teacher (n=2) with an = O. The parameter Dw = Jv Iw 1-w 2 12 and the input distribution determine the form of the teacher. This is shown in Figure 1. which displays the average output of a 2-teacher with one dimensional input vector. For 0'Xl =0'X2' Dw controls the variance about a linear mean output, and for fixed O'XI ?- 0'%'2' Dw controls the nonlinearity of the teacher. In the latter case, in the large N limit the variance of P(Yt I x) is zero. We can now explicitly write the evidence and perform the integration over the student parameters (over weights) . Taking the logarithm of the resulting expression leads to In P(1) I >" ,13) = - N 1(1) where the 1 is analogous to a free energy in statistical physics. 1 >.. - 1(1) = -In 2 11' a 13 1 1 1 + -In - + -lndetg + -ln211' + -P 'g'kPk - e 2 11' 2N 2 N J J and, n gjk1 = L Afk + >"Ojk p a=- n=1 N Here we are using the convention that summations are implied where repeated indices occur. 3Where N(x, 0'2) denotes a normal distribution with mean x and variance 0'2 . Hyperparameters, Evidence and Generalisation for an Un realisable Rule 259 The performance measures for this model are 2 {g = (~'x PA{w?w? - Oc = where, Nt (trg}'V - u2 (Wj}w = Pl:gl:j ' {g and 2w?(Wj}w + (Wj}!}}'V + {r; u;eff = PAu;o In order to pursue the calculation we consider the average of I(V) over all possible data sets just as, earlier, we defined our performance measures as averages over all data sets. This is some what artificial as we would normally be able to calculate I(V) and be interested in the generalisation error for our learning algorithm given a particular instance of the data. However, here we consider the thermodynamic limit (i.e., N,p - 00 s.t. 0 = piN = const.) in which, due to our sampling assumptions, the behaviours for typical examples of V coincide with that of the average. Details of the calculation will be published else where [Marion and Saad 95]. 4 RESULTS AND DISCUSSION We can now examine the evidence and the performance measures for our unlearnable problem. We note that in two limits we recover the learnable, linear teacher, case. Specifically if the probability of picking one of the component teachers is zero or if both component teacher vectors are aligned. In what follows we set Pi = P~ and normalise the components of the teacher such that Iwol = l. Firstly let us consider the performance measures. The asymptotic value of both {g and loci for large 0 is PiP~u;lu;:lDwlu;eff' This is the minimum generalisation error attainable and reflects the effective noise level due to the mismatch between student and teacher. We note here that the generalisation error is a function of ~ rather than f3 and 'Y independently. Figure 2a shows the generalisation error plotted against o. The addition of unlearn ability (Dw > 0) has a similar effect to the addition of noise on the examples. The appearance of the hump can be easily understood; If there is no noise or ~ is large enough then there is a steady reduction in {g. However, if this is not so then for small 0 the student learns this effective noise and the generalisation error increases with o . As the student gets more examples the effects of the noise begin to average out and the student starts to learn the rule. The point at which the generalisation error starts to decrease is influenced by the effective noise level and the prior constraint. Figure 2b shows the absolute value of the consistency measure v's 0 for non-optimal f3. Again we see that unlearn ability acts as an effective noise. For a few examples with ~ small or with large effective noise the student distribution is narrowed until the Oc is zero. However, the generalisation error is still increasing (as described above) and loci increases to a local maximum, it then asymptotically tends to { ,q . If there is no noise or ~ is large enough then loci steadily reduces as the number of examples increases. We now examine the evidence procedure. Firstly we define f3ev ( 'Y) and 'Yev (f3) to be the hyperparameters which maximise the evidence. The evidence procedure 260 Glenn Marion, David Saad 4,-------------------, 3 l ,I I l I 4 ,r,, (Ui) 3 ".. : ""'... " 1 ,~.. ............ . ... -... ""_ . . . . --..... _- J 18 2 (ii) I .....................................?... (i) O~---r~-,----~--~ o 1 2 (iii) --- 0 0 3 1 2 3 tl " (b) Consistency Measure (a) Generalisation error Figure 2: The performance measures: Graph a shows (g for finite lambda. a(i) and a(ii) are the learnable case with noise in the latter case. a(iii) shows that the effect of adding unlearn ability is qualitatively the same as adding noise. Graph b. shows the modulus of the consistency error v's a. Curves b(i) and b(ii) are the learnable case without and with noise respectively. Curve b(iii) is an unlearnable case with the same noise level. picks the point in hyperparameter space where these curves coincide. We denote the asymptotic values of 13ev(-y) and 'Yev(13) in the limit of large a by 1300 and 'Yoo respectively. In the linear case (Dw = 0) the evidence procedure assignments of the hyperparameters (for finite a) coincide with 1300 and 'Yoo and also optimise (g and 6c in agreement with [Bruce and Saad 94] . This is shown in Figure 3a where we plot the 13 which optimises the evidence (13ev) , the consistency measure (136 c) and the generalisation error (13!g) versus 'Y. The point at which the three curves coincide is the point in the 13-'Y plane identified by the evidence procedure. However, we note here that, if one of the hyperparameters is poorly determined then maximising the evidence with respect to the other is a misleading guide to optimising performance even in the linear case. The results for an unrealisable rule in the linear regime (Dw > 0, lrXI = lrX:l) are similar to the learnable case but with an increased noise due to the unlearn ability. The evidence procedure still optimises performance. In the non-linear regime (Dw > 0 , lrXI ?- lrX:l) the evidence procedure fails to minimise either performance measure. This is shown in Figure 3b where the evidence procedure point does not lie on 13!g ('Y) or 136 c (-y). Indeed, its hyperparameter assignments do not coincide with 1300 and 'Yoo but are a dependent. How badly does the evidence procedure fail? We define the percentage degradation in generalisation performance as I'\, = 100 * 9 (Aev) - (;Pt) / (;pt. Where Aev is the evidence procedure assignment and (;pt is the optimal generalisation error with respect to A. This is plotted in Figure 4a. We also define 1'\,6 = 100* 16c(Aev)1 /(g(A ev ). This measures the error in using the variance of the ?( Hyperparameters, Evidence and Generalisation for an Unrealisable Rule 1.0.....---------..,. ../ .....?.. 0.8- /Jopt. o.e ./... - ... _ ............ ", (i) ,. ". ~-"":: " '--1' 0.4. "", \" 0.2 (Ii) ". ".::r- \..(~) 0.0 I I 0.2 0.4. I .., o.e (a) Linear Case 0.5.....-----------, /Jopt. 0.3 (1) \~~"-/~ . (11),','/~; ,.' I 0.8 ,_ _ _.... 0.2 0.1 ,',',,, 0.0 261 1.0 0.0-f---'T"---,---r---1 0.0 0.6 1.0 2.0 .., (b) Non-Linear Case Figure 3: The evidence procedure:Optimal f3 v's /. In both graphs for (i) the evidence(f3ev), (ii) the generalisation error (f3f g ) and (iii) the consistency measure (f36J . The point which the evidence procedure picks in the linear case is that where all three curves coincide, whereas in the non linear case it coincides only with f3ev . post training distribution to estimate the generalisation error as a percentage of the generalisation error itself. Examples of this quantity are plotted in Figure 4b. There are three important points to note concerning I'\, and 1'\,6 . Firstly, the larger the deviation from a linear rule the greater is the error. Secondly, that it is the magnitude of the effective noise due to unlearnability relative to the real noise which determines this error. In other words , if the real noise is large enough to swamp the non-linearity of the rule then the evidence procedure will not be very misleading . Finally, the magnitude of the error for relatively large deviations from linearity is only a few percent and thus the evidence procedure might well be a reasonable, if not optimal, method for setting the hyperparameters. However, clearly it would be preferable to improve our student space to enable it to model the teacher . 5 CONCLUSION We have examined the generalisation error, the consistency measure and the evidence procedure within a model which allows us to interpolate between a learnable and an unlearnable scenario. We have seen that the unlearnability acts like an effective noise on the examples. Furthermore, we have seen that for a linear student the evidence procedure breaks down, in that it fails to optimise performance, when the teacher output is non-linear. However, even for relatively large deviations of the teacher from linearity the evidence procedure is close to optimal. Bayesian methods, such as the evidence procedure, are based on the assumption that the student or hypothesis space contains the teacher generating the data. In our case, in the non-linear regime, this is clearly not true and so it is perhaps not surprising that the evidence procedure is sub-optimal. Whether or not such a breakdown of the evidence procedure is a generic feature of a mismatch between the hypothesis space and the teacher is a matter for further study. 262 Glenn Marion, David Saad f\ 0.4-/J \\ J 0.3- J IC .' ! 0.1-1 0.0 (i) 3 (iii) \ \ " ",. , .,.., \ IC, / 2 " " ... ~.,... . . .................. --- _ ..-. 0.0 4 (Ii) \ \ 0.2- _. ,, ,, ,, , 5 o~,-----------------~ 0.5 t.o ----2.0 (I) (Ii) (iii) \ ,.0\0., "\ , ' .. !, ,I ! \\ V 0 0 """ "1 . .... - 3 2 Cl (a) ., .., 4 Cl (b) Figure 4: The relative degradation in performance compared to the optimal when using the evidence procedure to set the hyperparameters. Graph (a) shows the percentage degradation in generalisation performance K, ? a(i) has Dw = 1 with the real noise level u 1. a(ii) has this noise level reduced to u 0.1 and a(iii) has increased non-linearity, Dw = 3, and u = 1. Graph (b) shows the error made in predicting the generalisation error from the variance of the post training distribution as a percentage of the generalisation error itself, "'6 . b(i) and b(ii) have the same parameter values as a(i) and a(ii), whilst b(iii) has Dw = 3 and u = 0.1 = = Acknowledgments We are very grateful to Alastair Bruce and Peter Sollich for useful discussions. GM is supported by an E.P.S.R.C. studentship. References Bruce, A.D. and Saad, D. (1994) Statistical mechanics of hypothesis evaluation. J. of Phys. A: Math. Gen. 27:3355-3363 Krogh, A. and Hertz, J. (1992) Generalisation in a linear perceptron in the presence of noise. J. of Phys. A: Math. Gen. 25:1135-1147 MacKay, D.J.C. (1992a) Bayesian interpolation. Neural Compo 4:415-447 MacKay, D.J.C. (1992b) A practical Bayesian framework for backprop networks. Neural Compo 4:448-472 Marion, G. and Saad, D. (1995) A statistical mechanical analysis of a Bayesian inference scheme for an unrealisable rule. To appear in J. of Phys. A: Math. Gen. Seung, H. S, Sompolinsky, H., Tishby, N. (1992) Statistical mechanics of learning from examples. Phys. Rev. A, 45:6056-6091 Thodberg, H.H. (1994) Bayesian backprop in action:pruning, ensembles, error bars and application to spectroscopy. Advances in Neural Information Processing Systems 6:208-215. Cowan et al.(Eds.), Morgan Kauffmann, San Mateo, CA Wolpert, D. H and Strauss, C. E. M. (1994) What Bayes has to say about the evidence procedure. To appear in Maximum entropy and Bayesian methods. G . Heidbreder (Ed.), Kluwer.
952 |@word seek:2 attainable:1 pick:2 reduction:1 contains:1 nt:1 surprising:1 plot:1 plane:1 compo:2 math:3 ron:1 preference:2 firstly:3 fitting:2 advocate:1 introduce:1 indeed:1 examine:4 mechanic:2 trg:1 increasing:1 begin:1 underlying:2 linearity:4 heidbreder:1 what:6 pursue:1 whilst:1 act:2 f3f:1 preferable:1 uk:2 control:2 normally:1 unlearnable:3 appear:2 maximise:3 understood:1 local:1 tends:1 limit:4 despite:1 interpolation:1 might:2 mateo:1 examined:1 limited:1 averaged:1 practical:2 acknowledgment:1 practice:1 procedure:32 composite:1 confidence:1 integrating:1 word:1 suggest:1 alastair:1 get:1 close:1 context:1 equivalent:1 map:1 deterministic:1 yt:13 unlearnability:2 independently:2 simplicity:1 rule:15 dw:13 swamp:1 analogous:1 kauffmann:1 pt:3 gm:1 hypothesis:3 agreement:1 pa:4 breakdown:1 calculate:2 wj:3 sompolinsky:1 trade:1 decrease:1 principled:1 complexity:1 ui:1 seung:3 trained:3 grateful:1 easily:1 effective:7 artificial:1 whose:1 larger:1 say:1 ability:6 itself:2 fr:2 relevant:1 aligned:1 gen:3 poorly:1 generating:2 ac:2 fixing:1 krogh:3 convention:1 stochastic:1 enable:1 eff:2 backprop:3 behaviour:1 fix:1 probable:1 summation:1 secondly:1 pl:1 ic:2 normal:1 exp:1 mapping:1 predict:1 major:1 iw:1 reflects:1 clearly:4 gaussian:2 rather:2 pip:1 minimisation:1 unsatisfactory:1 likelihood:2 contrast:1 afk:1 sense:1 inference:2 dependent:1 lj:1 interested:2 priori:1 integration:1 mackay:6 field:1 optimises:3 never:1 f3:5 sampling:3 optimising:3 mimic:1 others:2 few:2 interpolate:2 attempt:1 normalisation:1 hump:1 evaluation:1 accurate:1 logarithm:1 plotted:3 theoretical:1 instance:1 formalism:2 earlier:1 increased:2 measuring:1 assignment:3 cost:1 deviation:3 marion:7 successful:1 tishby:1 teacher:32 corrupted:3 probabilistic:1 physic:2 off:1 yl:1 picking:1 squared:2 again:1 possibly:1 lambda:1 account:1 realisable:1 student:23 matter:1 explicitly:1 break:1 start:2 bayes:2 decaying:1 recover:1 narrowed:1 bruce:6 variance:9 percept:1 ensemble:1 ofthe:1 bayesian:10 produced:3 lu:1 multiplying:1 published:1 kpk:1 influenced:1 phys:4 ed:2 definition:1 against:2 energy:1 steadily:1 ask:1 supervised:2 strongly:1 furthermore:1 parameterised:1 just:1 until:1 nonlinear:1 yf:1 perhaps:2 believe:2 modulus:1 effect:3 validity:1 true:1 wp:1 razor:2 noted:1 steady:1 coincides:2 oc:2 formalises:1 percent:1 unlearn:4 extend:1 kluwer:1 refer:1 pew:1 consistency:8 nonlinearity:1 posterior:1 termed:1 scenario:1 seen:2 minimum:2 greater:1 morgan:1 determine:1 paradigm:1 ii:11 thermodynamic:1 desirable:1 reduces:1 xf:1 calculation:3 minimising:3 concerning:1 post:7 prediction:2 circumstance:1 addition:2 whereas:1 aev:3 else:1 standpoint:1 saad:12 comment:1 cowan:1 seem:1 presence:1 iii:9 enough:3 wn:1 identified:1 minimise:1 whether:1 expression:3 peter:1 action:1 useful:3 clear:1 simplest:1 reduced:1 supplied:1 percentage:4 estimated:1 write:1 hyperparameter:2 drawn:1 jv:1 wtw:1 asymptotically:1 graph:5 sum:2 wand:1 lrx:2 reasonable:1 eh9:1 display:1 badly:1 occur:1 constraint:1 x2:1 relatively:2 department:1 according:1 hertz:3 sollich:1 xlj:1 jopt:2 rev:1 discus:1 pin:1 fail:1 locus:3 hierarchical:1 generic:1 jn:1 denotes:1 include:1 const:1 implied:1 objective:1 question:2 quantity:1 usual:2 said:2 accommodates:1 normalise:1 extent:2 maximising:1 index:1 rise:1 perform:1 finite:2 yev:2 situation:1 dc:3 david:5 introduced:1 mechanical:2 able:1 bar:1 mismatch:2 yc:1 ev:3 regime:3 optimise:6 predicting:1 scheme:2 improve:2 misleading:4 prior:4 review:1 literature:1 asymptotic:3 relative:2 versus:1 thodberg:2 viewpoint:1 occam:2 pi:1 gl:1 supported:1 free:1 guide:3 perceptron:6 taking:1 absolute:1 fg:3 edinburgh:2 regard:1 curve:5 calculated:2 studentship:1 qualitatively:1 made:1 coincide:6 san:1 pruning:1 unreliable:1 reveals:1 xi:1 un:1 lthis:1 glenn:5 jz:1 learn:1 ca:1 spectroscopy:1 necessarily:1 cl:2 noise:24 hyperparameters:15 repeated:1 tl:1 sub:2 fails:2 xl:3 lie:1 learns:1 down:1 learnable:5 decay:1 evidence:48 essential:1 strauss:2 adding:3 magnitude:2 pau:1 wolpert:2 entropy:1 simply:2 explore:1 appearance:1 u2:1 determines:1 generalisation:28 specifically:2 typical:1 determined:1 degradation:3 ew:6 ojk:1 latter:2 incorporate:1 yoo:3 tested:2 ex:3
7,192
953
A Non-linear Information Maximisation Algorithm that Performs Blind Separation. Anthony J. Bell tonylOsalk.edu Terrence J. Sejnowski terrylOsalk.edu Computational Neurobiology Laboratory The Salk Institute 10010 N. Torrey Pines Road La Jolla, California 92037-1099 and Department of Biology University of California at San Diego La Jolla CA 92093 Abstract A new learning algorithm is derived which performs online stochastic gradient ascent in the mutual information between outputs and inputs of a network. In the absence of a priori knowledge about the 'signal' and 'noise' components of the input, propagation of information depends on calibrating network non-linearities to the detailed higher-order moments of the input density functions. By incidentally minimising mutual information between outputs, as well as maximising their individual entropies, the network 'factorises' the input into independent components. As an example application, we have achieved near-perfect separation of ten digitally mixed speech signals. Our simulations lead us to believe that our network performs better at blind separation than the HeraultJ utten network, reflecting the fact that it is derived rigorously from the mutual information objective. 468 1 Anthony J. Bell, Terrence J. Sejnowski Introduction Unsupervised learning algorithms based on information theoretic principles have tended to focus on linear decorrelation (Barlow & Foldiak 1989) or maximisation of signal-to-noise ratios assuming Gaussian sources (Linsker 1992). With the exception of (Becker 1992), there has been little attempt to use non-linearity in networks to achieve something a linear network could not. Non-linear networks, however, are capable of computing more general statistics than those second-order ones involved in decorrelation, and as a consequence they are capable of dealing with signals (and noises) which have detailed higher-order structure. The success of the 'H-J' networks at blind separation (Jutten & Herault 1991) suggests that it should be possible to separate statistically independent components, by using learning rules which make use of moments of all orders. This paper takes a principled approach to this problem, by starting with the question of how to maximise the information passed on in non-linear feed-forward network. Starting with an analysis of a single unit, the approach is extended to a network mapping N inputs to N outputs. In the process, it will be shown that, under certain fairly weak conditions, the N ---. N network forms a minimally redundant encoding ofthe inputs, and that it therefore performs Independent Component Analysis (ICA) . 2 Information maximisation The information that output Y contains about input X is defined as: = H(Y) - I(Y, X) H(YIX) (1) where H(Y) is the entropy (information) in the output, while H(YIX) is whatever information the output has which didn't come from the input. In the case that we have no noise (or rather, we don't know what is noise and what is signal in the input), the mapping between X and Y is deterministic and H(YIX) has its lowest possible value of -00 . Despite this, we may still differentiate eq.l as follows (see [5]): 8 8 8w I(Y, X) = 8w H(Y) (2) Thus in the noiseless case, the mutual information can be maximised by maximising the entropy alone. 2.1 One input, one output. Consider an input variable, x, passed through a transforming function, g( x), to produce an output variable, y, as in Fig.2.1(a). In the case that g(x) is monotonically increasing or decreasing (ie: has a unique inverse), the probability density function (pdf) of the output fy(y) can be written as a function of the pdfofthe input fx(x), (Papoulis, eq. 5-5): f (y) = fx(x) y 8y/8x (3) A Non-Linear Information Maximization Algorithm That Performs Blind Separation -1: 469 The entropy of the output, H(y), is given by: H(y) = -E[lnfy(y)] = fy(y)lnfy(y)dy (4) where E[.] denotes expected value. Substituting eq.3 into eq.4 gives H(y) = E [In ~:] (5) - E[lnfx(x)] The second term on the right may be considered to be unaffected by alterations in a parameter, w, determining g(x). Therefore in order to maximise the entropy of y by changing w, we need only concentrate on maximising the first term, which is the average log of how the input affects the output. This can be done by considering the 'training set' of x's to approximate the density fx(x), and deriving an 'online', stochastic gradient descent learning rule: ~W ex oH OW = ~ (In OY) = (oy) -1 ~ (oy) OW OX OX OW ox (6) In the case of the logistic transfer function y = (1 + e- U )-l , u = wx + Wo in which the input is multiplied by a weight wand added to a bias-weight wo, the terms above evaluate as: wY(l - y) (7) y(l - y)(l + wX(l - 2y)) (8) Dividing eq.8 by eq.7 gives the learning rule for the logistic function, as calculated from the general rule of eq.6: 1 ex - ~w W + x(l - 2y) (9) Similar reasoning leads to the rule for the bias-weight: ~wo ex 1- 2y (10) The effect of these two rules can be seen in Fig. 1a. For example, if the input pdf fx(x) was gaussian, then the ~wo-rule would centre the steepest part of the sigmoid curve on the peak of fx(x), matching input density to output slope, in a manner suggested intuitively by eq.3. The ~w-rule would then scale the slope of the sigmoid curve to match the variance of fx (x). For example, narrow pdfs would lead to sharply-sloping sigmoids. The ~w-rule is basically anti-Hebbian 1 , with an anti-decay term. The anti-Hebbian term keeps y away from one uninformative situation: that of y being saturated to 0 or 1. But anti-Hebbian rules alone make weights go to zero, so the anti-decay term (l/w) keeps y away from the other uninformative situation: when w is so small that y stays around 0.5. The effect of these two balanced forces is to produce an output pdf fy(y) which is close to the flat unit distribution, which is the maximum entropy distribution for a variable 11? y = tanh(wx + wo) then ~w ex !;; - 2xy 470 Anthony 1. Bell, Terrence 1. Sejnowski 4 ~---' 0 --~---4"""",""::~---- X y (a) ----~~~--~----X ox (b) Figure 1: (a) Optimal information flow in sigmoidal neurons (Schraudolph et al 1992). Input x having density function fx(x), in this case a gaussian, is passed through a non-linear function g(x). The information in the resulting density, fy(Y) depends on matching the mean and variance of x to the threshold and slope of g( x). In (b) fy(y) is plotted for different values of the weight w. The optimal weight, Wopt transmits most information. bounded between 0 and 1. Fig. 1b illustrates a family of these distributions, with the highest entropy one occuring at Wopt . A rule which maximises information for one input and one output may be suggestive for structures such as synapses and photoreceptors which must position the gain of their non-linearity at a level appropriate to the average value and size of the input fluctuations. However, to see the advantages of this approach in artificial neural networks, we now analyse the case of multi-dimensional inputs and outputs. 2.2 N inputs, N outputs. Consider a network with an input vector x, a weight matrix Wand a monotonically transformed output vector y = g(Wx + wo). Analogously to eq.3, the multivariate probability density function of y can be written (Papoulis, eq. 6-63): f ( ) = fx(x) y y (11) IJI where IJI is the absolute value ofthe Jacobian of the transformation. The Jacobian is the determinant of the matrix of partial derivatives: [ 8Xl ~ 8x n ~] ~ 8Xl ~ 8xn J = det: : (12) The derivation proceeds as in the previous section except instead of maximising In(8y/8x), now we maximise In IJI. For sigmoidal units, y = (1 + e-U)-l, U = A Non-Linear Information Maximization Algorithm That Performs Blind Separation 471 Wx + Wo, the resulting learning rules are familiar in form: D. W ex: [WTrl + x(1 - D.wo ex: 1- 2y 2y? (13) (14) except that now x, y, Wo and 1 are vectors (1 is a vector of ones), W is a matrix, and the anti-Hebbian term has become an outer product. The anti-decay term has generalised to the inverse of the transpose of the weight matrix. For an individual weight, Wij, this rule amounts to: cof Wij D.wij ex: det W + xj(l - 2Yi) (15) where cof Wij, the cofactor of Wij, is (-1 )i+j times the determinant of the matrix obtained by removing the ith row and the jth column from W. This rule is the same as the one for the single unit mapping, except that instead of W = 0 being an unstable point of the dynamics, now any degenerate weight matrix is unstable, since det W = 0 if W is degenerate. This fact enables different output units, Yi, to learn to represent different components in the input. When the weight vectors entering two output units become too similar, det W becomes small and the natural dynamic of learning causes these weight vectors to diverge. This effect is mediated by the numerator, cof Wij. When this cofactor becomes small, it indicates that there is a degeneracy in the weight matrix of the rest of the layer (ie: those weights not associated with input Xj or output Yi). In this case, any degeneracy in W has less to do with the specific weight Wij that we are adjusting. 3 Finding independent components - blind separation Maximising the information contained in a layer of units involves maximising the entropy of the individual units while minimising the mutual information (the redundancy) between them. Considering two units: (16) For I(Yl, Y2) to be zero, Yl and Y2 must be statistically independent of each other, so that f Y1 Y2(Yl, Y2) = fYl (ydfY2(Y2). Achieving such a representation is variously called factorial code learning, redundancy reduction (Barlow 1961, Atick 1992), or independent component analysis (ICA), and in the general case of continuously valued variables of arbitrary distributions, no learning algorithm has been shown to converge to such a representation. Our method will converge to a minimum redundancy, factorial representation as long as the individual entropy terms in eq.16 do not override the redundancy term, making an I(Yl, Y2) = 0 solution sub-optimal. One way to ensure this cannot occur is if we have a priori knowledge of the general form of the pdfs of the independent components. Then we can tailor our choice of node-function to be optimal for transmitting information about these components. For example, unit distributions require piecewise linear node-functions for highest H(y), while the more common gaussian forms require roughly sigmoidal curves. Once we have chosen our nodefunctions appropriately, we can be sure that an output node Y cannot have higher 472 Anthony 1. Bell, Terrence 1. Sejnowski- slt--_---<x\._ _---4!'.~ o o o )<----~ unknown mixing process 1'-----..:?0 BUND SEPARATION (learnt weights) WA '" after learning: 1-4.091 0.13 0.09 0.07 1-2.921 0.00 0.02 -0.02 -0.06 0.02 0.03 0.00 -0.07 0.14 1-3.501 -0.07 -0.01 0.02 -0.06 -0.08 1-2.2ij 11.971 0.02 -0.01 0.04 Figure 2: (a) In blind separation, sources, s, have been linearly scrambled by a matrix, A, to form the inputs to the network, x. We must recover the sources at our output y, by somehow inverting the mapping A with our weight matrix W. The problem: we know nothing about A or the sources. (b) A successful 'unscrambling' occurs when WA is a 'permutation' matrix. This one resulted from separating five speech signals with our algorithm. entropy by representing some combination of independent components than by representing just one. When this condition is satisfied for all output units, the residual goal, of minimising the mutual information between the outputs, will dominate the learning. See [5] for further discussion of this. With this caveat in mind, we turn to the problem of blind separation, (Jutten & Herault 1991), illustrated in Fig.2. A set of sources, Sl, ... , SN, (different people speaking, music, noise etc) are presumed to be mixed approximately linearly so that all we receive is N superpositions of them, Xl, ... , X N, which are input to our single-layer information maximisation network. Providing the mixing matrix, A, is non-singular then the original sources can be recovered if our weight matrix, W, is such that W A is a 'permutation' matrix containing only one high value in each row and column. Unfortunately we know nothing about the sources or the mixing matrix. However, if the sources are statistically independent and non-gaussian, then the information in the output nodes will be maximised when each output transmits one independent component only. This problem cannot be solved in general by linear decorrelation techniques such as those of (Barlow & Foldicik 1989) since second-order statistics can only produce symmetrical decorrelation matrices. We have tested the algorithm in eq.13 and eq.14 on digitally mixed speech signals, and it reliably converges to separate the individual sources. In one example, five separately-recorded speech signals from three individuals were sampled at 8kHz. Three-second segments from each were linearly mixed using a matrix of random values between 0.2 and 4. Each resulting mixture formed an incomprehensible babble. Time points were generated at random, and for each, the corresponding five mixed values were entered into the network, and weights were altered according to eq.13 and eq.14. After on the order of 500,000 points were presented, the network A Non-Linear Information Maximization Algorithm That Performs Blind Separation 473 had converged so that WA was the matrix shown in Fig.2b. As can be seen from the permutation structure of this matrix, on average, 95% of each output unit is dedicated to one source only, with each unit carrying a different source. Any residual interference from the four other sources was inaudible. We have not yet performed any systematic studies on rates of convergence or existence of local minima. However the algorithm has converged to separate N independent components in all our tests (for 2 ~ N ~ 10). In contrast, we have not been able to obtain convergence of the H-J network on our data set for N > 2. Finally, the kind of linear static mixing we have been using is not equivalent to what would be picked up by N microphones positioned around a room. However, (Platt & Faggin 1992) in their work on the H-J net, discuss extensions for dealing with time-delays and non-static filtering, which may also be applicable to our methods. 4 Discussion The problem of Independent Component Analysis (ICA) has become popular recently in the signal processing community, partly as a result ofthe success ofthe H-J network. The H-J network is identical to the linear decorrelation network of (Barlow & Foldicik 1989) except for non-linearities in the anti-Hebb rule which normally performs only decorrelation. These non-linearities are chosen somewhat arbitrarily in the hope that their polynomial expansions will yield higher-order cross-cumulants useful in converging to independence (Comon et aI, 1991). The H-J algorithm lacks an objective function, but these insights have led (Comon 1994) to propose minimising mutual information between outputs (see also Becker 1992). Since mutual information cannot be expressed as a simple function of the variables concerned, Comon expands it as a function of cumulants of increasing order. In this paper, we have shown that mutual information, and through it, ICA, can be tackled directly (in the sense of eq.16) through a stochastic gradient approach. Sigmoidal units, being bounded, are limited in their 'channel capacity' . Weights transmitting information try, by following eq.13, to project their inputs to where they can make a lot of difference to the output, as measured by the log of the Jacobian of the transformation. In the process, each set of statistically 'dependent' information is channelled to the same output unit. The non-linearity is crucial. If the network were just linear, the weight matrix would grow without bound since the learning rule would be: (17) This reflects the fact that the information in the outputs grows with their variance. The non-linearity also supplies the higher-order cross-moments necessary to maximise the infinite-order expansion of the information. For example, when y = tanh( u), the learning rule has the form D.. W ex: [wT] -1 - 2yxT , from which we can write that the weights stabilise (or (D..W) = 0) when 1= 2(tanh(u)uT ). Since tanh is an odd function, its series expansion is of the form tanh( u) = L: j bj u 2P+1 , the bj being coefficients, and thus this convergence criterion amounts to the condition L:i,j bijp(U;P+1Uj) = 0 for all output unit pairs i 1= j, for p = 0,1,2,3 ..., 474 Anthony J. Bell, Terrence J. Sejnowski and for the coefficients bijp coming from the Taylor series expansion of the tanh function. These and other issues are covered more completely in a forthcoming paper (Bell & Sejnowski 1995). Applications to blind deconvolution (removing the effects of unknown causal filtering) are also described, and the limitations of the approach are discussed. Acknowledgements We are much indebted to Nici Schraudolph, who not only supplied the original idea in Fig.l and shared his unpublished calculations [13], but also provided detailed criticism at every stage of the work. Much constructive advice also came from Paul Viola and Alexandre Pouget. References [1] Atick J.J. 1992. Could information theory provide an ecological theory of sensory processing, Network 3, 213-251 [2] Barlow H.B. 1961. Possible principles underlying the transformation of sensory messages, in Sensory Communication, Rosenblith W.A. (ed), MIT press [3] Barlow H.B. & Foldicik P. 1989. Adaptation and decorrelation in the cortex, in Durbin R. et al (eds) The Computing Neuron, Addison-Wesley [4] Becker S. 1992. An information-theoretic unsupervised learning algorithm for neural networks, Ph.D. thesis, Dept. of Compo Sci., Univ. of Toronto [5] Bell A.J. & Sejnowski T.J. 1995. An information-maximisation approach to blind separation and blind deconvolution, Neural Computation, in press [6] Comon P., Jutten C. & Herault J. 1991. Blind separation of sources, part II: problems statement, Signal processing, 24, 11-21 [7] Comon P. 1994. Independent component analysis, a new concept? Signal processing, 36, 287-314 [8] Hopfield J.J. 1991. Olfactory computation and object perception, Proc. N atl. Acad. Sci. USA, vol. 88, pp.6462-6466 [9] Jutten C. & Herault J. 1991. Blind separation of sources, part I: an adaptive algorithm based on neuromimetic architecture, Signal processing, 24, 1-10 [10] Linsker R. 1992. Local synaptic learning rules suffice to maximise mutual information in a linear network, Neural Computation, 4, 691-702 [11] Papoulis A. 1984. Probability, random variables and stochastic processes, 2nd edition, McGraw-Hill, New York [12] Platt J .C. & Faggin F. 1992. Networks for the separation of sources that are superimposed and delayed, in Moody J.E et al (eds) Adv. Neur. Inf. Proc. Sys. 4, Morgan-Kaufmann [13] Schraudolph N.N., Hart W .E. & Belew R.K. 1992. Optimal information flow in sigmoidal neurons, unpublished manuscript
953 |@word determinant:2 polynomial:1 nd:1 simulation:1 papoulis:3 reduction:1 moment:3 contains:1 series:2 recovered:1 yet:1 written:2 must:3 wx:5 enables:1 alone:2 maximised:2 sys:1 steepest:1 ith:1 compo:1 caveat:1 node:4 toronto:1 sigmoidal:5 five:3 become:3 supply:1 olfactory:1 manner:1 expected:1 presumed:1 ica:4 roughly:1 multi:1 decreasing:1 little:1 considering:2 increasing:2 becomes:2 project:1 provided:1 linearity:7 bounded:2 underlying:1 didn:1 sloping:1 lowest:1 what:3 suffice:1 kind:1 finding:1 transformation:3 every:1 expands:1 platt:2 whatever:1 unit:16 normally:1 generalised:1 maximise:5 local:2 consequence:1 acad:1 despite:1 encoding:1 fluctuation:1 approximately:1 minimally:1 suggests:1 limited:1 statistically:4 unique:1 maximisation:5 bell:7 matching:2 road:1 cannot:4 close:1 equivalent:1 deterministic:1 go:1 starting:2 pouget:1 rule:18 insight:1 deriving:1 dominate:1 oh:1 his:1 atl:1 fx:8 diego:1 solved:1 adv:1 highest:2 digitally:2 principled:1 transforming:1 balanced:1 rigorously:1 dynamic:2 carrying:1 segment:1 completely:1 hopfield:1 derivation:1 univ:1 sejnowski:7 artificial:1 valued:1 statistic:2 torrey:1 analyse:1 online:2 differentiate:1 advantage:1 net:1 propose:1 product:1 coming:1 adaptation:1 entered:1 mixing:4 degenerate:2 achieve:1 convergence:3 produce:3 incidentally:1 perfect:1 converges:1 object:1 measured:1 ij:1 odd:1 eq:17 dividing:1 involves:1 come:1 concentrate:1 stochastic:4 incomprehensible:1 babble:1 require:2 extension:1 around:2 considered:1 mapping:4 bj:2 pine:1 substituting:1 proc:2 applicable:1 tanh:6 superposition:1 reflects:1 hope:1 mit:1 gaussian:5 yix:3 rather:1 derived:2 focus:1 pdfs:2 indicates:1 superimposed:1 contrast:1 criticism:1 sense:1 stabilise:1 dependent:1 wij:7 transformed:1 issue:1 priori:2 herault:4 fairly:1 mutual:10 once:1 having:1 biology:1 identical:1 unsupervised:2 linsker:2 piecewise:1 channelled:1 resulted:1 individual:6 delayed:1 variously:1 familiar:1 attempt:1 message:1 saturated:1 mixture:1 capable:2 partial:1 necessary:1 xy:1 taylor:1 plotted:1 causal:1 column:2 cumulants:2 maximization:3 delay:1 successful:1 too:1 faggin:2 learnt:1 yxt:1 density:7 peak:1 ie:2 stay:1 systematic:1 terrence:5 yl:4 diverge:1 analogously:1 continuously:1 moody:1 transmitting:2 thesis:1 satisfied:1 recorded:1 containing:1 derivative:1 alteration:1 utten:1 coefficient:2 blind:14 depends:2 performed:1 try:1 picked:1 lot:1 recover:1 slope:3 formed:1 variance:3 who:1 kaufmann:1 yield:1 ofthe:4 weak:1 basically:1 fyl:1 unaffected:1 indebted:1 converged:2 synapsis:1 tended:1 rosenblith:1 ed:3 synaptic:1 pp:1 involved:1 transmits:2 associated:1 static:2 degeneracy:2 gain:1 sampled:1 adjusting:1 iji:3 popular:1 knowledge:2 ut:1 positioned:1 reflecting:1 manuscript:1 feed:1 alexandre:1 higher:5 wesley:1 done:1 ox:4 just:2 atick:2 stage:1 propagation:1 lack:1 somehow:1 jutten:4 logistic:2 grows:1 believe:1 usa:1 effect:4 calibrating:1 concept:1 y2:6 barlow:6 entering:1 laboratory:1 illustrated:1 numerator:1 criterion:1 pdf:3 override:1 hill:1 theoretic:2 occuring:1 performs:8 dedicated:1 reasoning:1 recently:1 sigmoid:2 common:1 khz:1 discussed:1 cofactor:2 ai:1 centre:1 had:1 cortex:1 etc:1 something:1 multivariate:1 foldiak:1 jolla:2 inf:1 certain:1 ecological:1 success:2 arbitrarily:1 came:1 yi:3 seen:2 minimum:2 morgan:1 somewhat:1 nici:1 converge:2 redundant:1 monotonically:2 signal:12 ii:1 hebbian:4 match:1 calculation:1 minimising:4 schraudolph:3 long:1 cross:2 hart:1 converging:1 noiseless:1 represent:1 achieved:1 receive:1 uninformative:2 separately:1 singular:1 source:15 grow:1 crucial:1 appropriately:1 rest:1 ascent:1 sure:1 flow:2 near:1 concerned:1 affect:1 xj:2 independence:1 forthcoming:1 architecture:1 idea:1 det:4 passed:3 becker:3 wo:9 speech:4 speaking:1 cause:1 york:1 useful:1 detailed:3 covered:1 factorial:2 amount:2 ten:1 ph:1 sl:1 supplied:1 write:1 vol:1 redundancy:4 four:1 threshold:1 achieving:1 changing:1 wand:2 inverse:2 tailor:1 family:1 separation:15 dy:1 wopt:2 layer:3 bound:1 tackled:1 durbin:1 occur:1 sharply:1 flat:1 department:1 according:1 neur:1 combination:1 making:1 foldicik:3 comon:5 intuitively:1 interference:1 inaudible:1 turn:1 discus:1 know:3 mind:1 addison:1 neuromimetic:1 multiplied:1 away:2 appropriate:1 existence:1 original:2 denotes:1 ensure:1 belew:1 music:1 uj:1 objective:2 question:1 added:1 occurs:1 gradient:3 ow:3 separate:3 separating:1 capacity:1 sci:2 outer:1 evaluate:1 fy:5 unstable:2 maximising:6 assuming:1 code:1 ratio:1 providing:1 unfortunately:1 statement:1 reliably:1 unknown:2 maximises:1 neuron:3 descent:1 anti:8 situation:2 neurobiology:1 extended:1 viola:1 communication:1 y1:1 arbitrary:1 community:1 inverting:1 pair:1 unpublished:2 california:2 narrow:1 able:1 suggested:1 proceeds:1 wy:1 perception:1 decorrelation:7 natural:1 force:1 residual:2 representing:2 altered:1 factorises:1 mediated:1 sn:1 acknowledgement:1 determining:1 permutation:3 mixed:5 oy:3 limitation:1 filtering:2 principle:2 row:2 transpose:1 jth:1 bias:2 institute:1 slt:1 absolute:1 curve:3 calculated:1 xn:1 sensory:3 forward:1 adaptive:1 san:1 approximate:1 mcgraw:1 keep:2 dealing:2 suggestive:1 photoreceptors:1 symmetrical:1 don:1 scrambled:1 learn:1 transfer:1 channel:1 ca:1 expansion:4 anthony:5 linearly:3 noise:6 paul:1 edition:1 nothing:2 fig:6 advice:1 hebb:1 salk:1 sub:1 position:1 xl:3 jacobian:3 removing:2 specific:1 decay:3 deconvolution:2 sigmoids:1 illustrates:1 entropy:10 led:1 expressed:1 contained:1 goal:1 room:1 shared:1 absence:1 infinite:1 except:4 wt:1 microphone:1 called:1 partly:1 la:2 exception:1 people:1 cof:3 constructive:1 dept:1 tested:1 ex:8
7,193
954
Anatomical origin and computational role of diversity in the response properties of cortical neurons Kalanit Grill Spectort Shimon Edelmant Rafael Malacht Depts of tApplied Mathematics and Computer Science and tN eurobiology The Weizmann Institute of Science Rehovot 76100, Israel {kalanit.edelman. malach }~wisdom . weizmann .ac.il Abstract The maximization of diversity of neuronal response properties has been recently suggested as an organizing principle for the formation of such prominent features of the functional architecture of the brain as the cortical columns and the associated patchy projection patterns (Malach, 1994). We show that (1) maximal diversity is attained when the ratio of dendritic and axonal arbor sizes is equal to one, as found in many cortical areas and across species (Lund et al., 1993; Malach, 1994), and (2) that maximization of diversity leads to better performance in systems of receptive fields implementing steerable/shiftable filters, and in matching spatially distributed signals, a problem that arises in many high-level visual tasks. 1 Anatomical substrate for sampling diversity A fundamental feature of cortical architecture is its columnar organization, manifested in the tendency of neurons with similar properties to be organized in columns that run perpendicular to the cortical surface. This organization of the cortex was initially discovered by physiological experiments (Mouncastle, 1957; Hubel and Wiesel, 1962), and subsequently confirmed with the demonstration of histologically defined columns. Tracing experiments have shown that axonal projections throughout the cerebral cortex tend to be organized in vertically aligned clusters or patches. In particular, intrinsic horizontal connections linking neighboring cortical sites, which may extend up to 2 - 3 mm, have a striking tendency to arborize selectively in preferred sites, forming distinct axonal patches 200 - 300 J.lm in diameter. Recently, it has been observed that the size of these patches matches closely the average diameter of individual dendritic arbors of upper-layer pyramidal cells 118 Kalanit Grill Spector, Shimon Edelman, Rafael Malach .... 2&00 10(0 : : &00 , / '" ! .._._.r::~ ._. ~._ . _._;~ \ ~ ~ ~ ~ ~ ,*,*".",p.d ~ hm pa~ ) N BO 80 100 '0 0:2 0" 0S 0.8 t 1 :2 1 .. 1S t IS r?? o betiIJMn I'IWI'ot'I .. d ptIId! Figure 1: Left: histograms of the percentage of patch-originated input to the neurons, plotted for three values of the ratio r between the dendritic arbor and the patch diameter (0,5, 1.0, 2.0). The flattest histogram is obtained for r = 1.0 Right: the di versity of neuronal properties (as defined in section 1) vs. r. The maximum is attained for r = 1.0, a value compatible with the anatomical data. (see Malach, 1994, for a review). Determining the functional significance of this correlation, which is a fundamental property that holds throughout various cortical areas and across species (Lund et al., 1993), may shed light on the general principles of operation of the cortical architecture. One such driving principle may be the maximization of diversity of response properties in the neuronal population (Malach, 1994). According to this hypothesis, matching the sizes of the axonal patches and the dendritic arbors causes neighboring neurons to develop slightly different functional selectivity profiles, resulting in an even spread of response preferences across the cortical population, and in an improvement of the brain's ability to process the variety of stimuli likely to be encountered in the environment. 1 To test the effect of the ratio between axonal patch and dendritic arbor size on the diversity of the neuronal population, we conducted computer simulations based on anatomical data concerning patchy projections (Rockland and Lund, 1982; Lund et al., 1993; Malach, 1992; Malach et al., 1993). The patches were modeled by disks, placed at regular intervals of twice the patch diameter, as revealed by anatomical labeling. Dendritic arbors were also modeled by disks, whose radii were manipulated in different simulations. The arbors were placed randomly over the axonal patches, at a density of 10,000 neurons per patch. We then calculated the amount of patchrelated information sampled by each neuron, defined to be proportional to the area of overlap of the dendritic tree and the patch. The results of the calculations for three 1 Necessary conditions for obtaining dendritic sampling diversity are that dendritic arbors cross freely through column borders, and that dendrites which cross column borders sample with equal probability from patch and inter-patch compartments. These assumptions were shown to be valid in (Malach, 1992; Malach, 1994). Diversity in the Response Properties of Cortical Neurons 119 values of the ratio of patch and arbor diameters appear in Figure 1. The presence of two peaks in the histogram obtained with the arbor/patch ratio r = 0.5 indicates that two dominant groups are formed in the population, the first receiving most of its input from the patch, and the second - from the inter-patch sources. A value of r = 2.0, for which the dendritic arbors are larger than the axonal patch size, yields near uniformity of sampling properties, with most of the neurons receiving mostly patch-originated input , as apparent from the single large peak in the histogram. To quantify the notion of diversity, we defined it as diversity "'< I~; I > -1, where n(p) is the number of neurons that receive p percent of their inputs from the patch, and < . > denotes average over p. Figure 1, right, shows that diversity is maximized when the size of the dendritic arbors matches that of the axonal patches, in accordance with the anatomical data. This result confirms the diversity maximization hypothesis stated in (Malach, 1994). 2 Orientation tuning: a functional manifestation of sam pIing diversity The orientation columns in VI are perhaps the best-known example of functional architecture found in the cortex (Bubel and Wiesel, 1962). Cortical maps obtained by optical imaging (Grinvald et al., 1986) reveal that orientation columns are patchy rather then slab-like: domains corresponding to a single orientation appear as a mosaic of round patches, which tend to form pinwheel-like structures. Incremental changes in the orientation of the stimulus lead to smooth shifts in the position of these domains. We hypothesized that this smooth variation in orientation selectivity found in VI originates in patchy projections, combined with diversity in the response properties of neurons sampling from these projections. The simulations described in the rest of this section substantiate this hypothesis. Computer simulations. The goal of the simulations was to demonstrate that a limited number of discretely tuned elements can give rise to a continuum of responses. We did not try to explain how the original set of discrete orientations can be formed by projections from the LGN to the striate cortex; several models for this step can be found in the literature (Bubel and Wiesel, 1962; Vidyasagar, 1985).2 In setting the size of the original discrete orientation columns we followed the notion of a point image (MacIlwain, 1986), defined as the minimal cortical separation of cells with non-overlapping RFs. Each column was tuned to a specific angle, and located at an approximately constant distance from another column with the same orientation tuning (we allowed some scatter in the location of the RFs). The RFs of adjacent units with the same orientation preference were overlapping, and the amount of overlap 2ln particular, it has been argued (Vidyasagar, 1985) that the receptive fields at the output of the LGN are already broadly tuned for a small number of discrete orientations (possibly just horizontal and vertical), and that at the cortical level the entire spectrum of orientations is generated from the discrete set present in the geniculate projection. 120 Kalanit Grill Spector, Shimon Edelman, Rafael Malach 3< 007 32 OOS 30 OOs II 0 28 ! 00' 1 ! . 26 g 003 .. 24 ? 00' 22 001 0 0 20 .0 20 30 '" runber ~ shlttlng fIlters So SO 70 '?0 .0 20 30 '" nunber cI shttng tlMlI'1 so . Figure 2: The effects of (independent) noise in the basis RFs and in the steering/shifting coefficients. Left: the approximation error vs. the number of basis RFs used in the linear combination. Right: the signal to noise ratio vs. the number of basis RFs. The SNR values were calculated as 10 loglO (signal energy/noise energy). Adding RFs to the basis increases the accuracy of the resultant interpolated RF. was determined by the number of RFs incorporated into the network. The preferred orientations were equally spaced at angles between 0 and 1r. The RFs used in the simulations were modeled by a product of a 2D Gaussian G 1 , centered at rj, with orientation selectivity G 2, and optimal angle Oi: G(r, rj, 0, Oi) = G1(r, rj)G 2(O, Oi). According to the recent results on shiftable/steerable filters (Simon celli et al., 1992), a RF located at and tuned to the orientation ,po can be obtained by a linear combination of basis RFs, as follows: ro G(r, ro, 0, ,po) M-IN-I L L bj(ro)ki(,po)G(r,rj,O,Oi) j=O i=O M-I N-I 2: bj(ro)G1(r, rj) 2: ki(,po)G2(O,Oi) j=O (1 ) i=O From equation 1 it is clear that the linear combination is equivalent to an outer product of the shifted and the steered RFs, with {ki(,pO)}~~1 and {bj(ro)}~~l denoting the steering and shifting coefficients, respectively. Because orientation and localization are independent parameters, the steering coefficients can be calculated separately from the shifting coefficients. The number of steering coefficients depends on the polar Fourier bandwidth of the basis RF, while the number of steering filters is inversely proportional to the basis RF size (Grill-Spector et al., 1995). In the presence of noise this minimal basis has to be extended (see Figure 2). The results ofthe simulation for several RF sizes are shown in Figure 3, left. As expected, the number of basis RFs required to approximate a desired RF is inversely proportional to the 121 Diversity in the Response Properties of Cortical Neurons The dependency oI1he nlnber d RF9 a1 the venene. '. ,5 2 van.,08 25 35 Figure 3: Left: error of the steering/shifting approximation for several basis RF sizes. Right: the number of basis RFs required to achieve a given error for different sizes of the basis RFs. The dashed line is the hyperbola num RFs x size = const. size of the basis RFs (Figure 3, right). Steerability and biological considerations. The anatomical finding that the columnar "borders" are freely crossed by dendritic and axonal arbors (Malach, 1992), and the mathematical properties of shiftable/steerable filters outlined above suggest that the columnar architecture in VI provides a basis for creating a continuum of RF properties, rather that being a form of organizing RFs in discrete bins. Computationally, this may be possible if the input to neurons is a linear combination of outputs of several RFs, as in equation 1. The anatomical basis for this computation may be provided by intrinsic cortical connections. It is known that long-range (I"V 1mm) connections tend to link cells with like orientation preference, while the short-range (I"V 400 J.lm) connections are made to cells of diverse orientation preferences (Malach et al., 1993). We suggest that the former provide the inputs necessary to shift the position of the desired RF, while the latter participate in steering the RF to an arbitrary angle (see Grill-Spector et al., 1995, for details). 3 Matching with patchy connections Many visual tasks require matching between images taken at different points in space (as in binocular stereopsis) or time (as in motion processing). The first and foremost problem faced by a biological system in solving these tasks is that the images to be compared are not represented as such anywhere in the system: instead of images, there are patterns of activities of neurons, with RFs that are overlapping, are not located on a precise grid, and are subject to mixing by patchy projections in each successive stage of processing. In this section, we show that a system composed of scattered RFs with smooth and overlapping tuning functions can, as a matter of 122 Kalanit Grill Spector, Shimon Edelman, Rafael Malach fact, perform matching precisely by allowing patchy connections between domains. Moreover, the weights that must be given to the various inputs that feed a RF carrying out the match are identical to the coefficients that would be generated by a learning algorithm required to capture a certain well-defined input-output relationship from pairs of examples. DOMAIN A Figure 4: Unit C receives patchy input from areas A and B which contain receptors with overlapping RFs. Consider a unit C, sampling two domains A and B through a Gaussian-profile dendritic patch equal in size to that of the axonal arbor of cells feeding A and B (Figure 4). The task faced by unit C is to determine the degree to which the activity patterns in domains A and B match. Let <Pjp be the response of the j'th unit in A to an input x-;': A. . = exp{ - ( .. Xp - ")2 Xj 20'2 'l'JP } (2) where xj be the optimal pattern to which the j'th unit is tuned (the response Bjp of a unit in B is of similar form). If, for example, domains A and B contain orientation selective cells, then xj would be the optimal combination of orientation and location of a bar stimulus. For simplicity we assume that all the RFs are of the same size 0', that unit C samples the same number of neurons N from both domains, and that the input from each domain to unit C is a linear combination of the responses of the units in each area. The input to C from domain A, with x-;, presented to the system is then: N Ain =L j=1 aj<pjp (3) Diversity in the Response Properties of Cortical Neurons 123 The problem is to find coefficients {aj} and {bj } such that on a given set of inputs {x-;} the outputs of domains A and B will match. We define the matching error as follows: Em = t (~a,~,p -~b'8'p)' (4) Proposition 1 The desired coefficients, minimizing Em, can be generated by an algorithm trained to learn an input/output mapping from a set of examples. This proposition can be proved by taking the derivative of Em with respect to the coefficients (Grill-Spector et al., 1995). Learning here can be carried out by radial basis function (RBF) approximation (Poggio and Girosi, 1990), which is particularly suitable for our purpose, because its basis functions can be regarded as multidimensional Gaussian RFs. 4 Summary Our results show that maximal diversity of neuronal response properties is attained when the ratio of dendritic and axonal arbor sizes is equal to 1, a value found in many cortical areas and across species (Lund et al., 1993; Malach, 1994). Maximization of diversity also leads to better performance in systems of receptive fields implementing steerablejshiftable filters, which may be necessary for generating the seemingly continuous range of orientation selectivity found in VI, and in ma.tching spatially distributed signals. This cortical organization principle may, therefore, have the double advantage of accounting for the formation of the cortical columns and the associated patchy projection patterns, and of explaining how systems of receptive fields can support functions such as the generation of precise response tuning from imprecise distributed inputs, and the matching of distributed signals, a problem that arises in visual tasks such as stereopsis, motion processing, and recognition. References Grill-Spector, K. , Edelman, S., and Malach, R. (1995). Anatomical origin and computational role of diversity in the response properties of cortical neurons. In Aertsen, A., editor, Brain Theory: biological basis and computational theory of vision. Elsevier. in press. Grinvald, A., Lieke, T., Frostigand, R., Gilbert, C. , and Wiesel, T. (1986). Functional architecture of the cortex as revealed by optical imaging of intrinsic signals. Nature, 324:361-364. Rubel, D. and Wiesel, T . (1962) . Receptive fields, binocular interactions and functional architecture in the cat's visual cortex. Journal of Physiology, 160:106-154. 124 Kalanit Grill Spector, Shimon Edelman, Rafael Malach Lund, J., Yoshita, S., and Levitt , J. (1993). Comparison of intrinsic connections in different areas of macaque cerebral cortex. Cerebral Cortex, 3:148- 162. MacIlwain, J. (1986). Point images in the visual system: new interest in an old idea. Trends in Neurosciences, 9:354- 358 . Malach, R. (1992) . Dendritic sampling across processing streams in monkey striate cortex. Journal of Comparative Neurobiology, 315:305-312. Malach, R. (1994). Cortical columns as devices for maximizing neuronal diversity. Trends in Neurosciences, 17:101- 104. Malach, R., Amir, Y., Harel, M., and Grinvald, A. (1993) . Relationship between intrinsic connections and functional architecture,revealed by optical imaging and in vivo targeted biocytine injections in primate striate cortex. Proceedings of the National Academy of Science, USA, 90:10469- 10473. Mouncastle, V. (1957). Modality and topographic properties of single neurons of cat's somatic sensory cortex. Journal of Neurophysiology, 20:408-434 . Poggio, T. and Girosi, F. (1990). Regularization algorithms for learning that are equivalent to multilayer networks . Science, 247:978- 982. Rockland, K. and Lund, J. (1982). Widespread periodic intrinsic connections in the tree shrew visual cortex. Science, 215: 1532-1534. Simoncelli, E., Freeman, W., Adelson, E., and Heeger, D. (1992). Shiftable multiscale transformations. IEEE Transactions on Information Theory, 38:587-607. Vidyasagar, T. (1985). Geniculate orientation biases as cartesian coordinates for cortical orientation detectors. In Models for the visual cortex, pages 390- 395 . Wiley, New York.
954 |@word neurophysiology:1 wiesel:5 disk:2 confirms:1 simulation:7 accounting:1 tuned:5 denoting:1 scatter:1 must:1 girosi:2 v:3 device:1 amir:1 short:1 num:1 provides:1 location:2 preference:4 successive:1 mathematical:1 edelman:6 inter:2 expected:1 brain:3 freeman:1 versity:1 provided:1 moreover:1 israel:1 monkey:1 finding:1 transformation:1 multidimensional:1 shed:1 ro:5 originates:1 unit:10 appear:2 vertically:1 accordance:1 receptor:1 approximately:1 twice:1 limited:1 perpendicular:1 range:3 weizmann:2 steerable:3 area:7 physiology:1 projection:9 matching:7 imprecise:1 radial:1 regular:1 suggest:2 gilbert:1 equivalent:2 map:1 maximizing:1 simplicity:1 regarded:1 population:4 notion:2 variation:1 coordinate:1 substrate:1 mosaic:1 hypothesis:3 origin:2 pa:1 element:1 trend:2 recognition:1 particularly:1 located:3 malach:21 observed:1 role:2 capture:1 environment:1 trained:1 uniformity:1 solving:1 carrying:1 localization:1 basis:18 po:5 various:2 represented:1 cat:2 distinct:1 labeling:1 formation:2 whose:1 apparent:1 spector:8 larger:1 ability:1 g1:2 topographic:1 seemingly:1 advantage:1 shrew:1 interaction:1 maximal:2 product:2 neighboring:2 aligned:1 organizing:2 rockland:2 mixing:1 achieve:1 academy:1 cluster:1 double:1 generating:1 incremental:1 comparative:1 develop:1 ac:1 quantify:1 radius:1 closely:1 filter:6 subsequently:1 centered:1 implementing:2 bin:1 argued:1 require:1 feeding:1 proposition:2 dendritic:15 biological:3 celli:1 mm:2 hold:1 exp:1 mapping:1 bj:4 lm:2 slab:1 driving:1 continuum:2 purpose:1 polar:1 geniculate:2 ain:1 loglo:1 gaussian:3 rather:2 kalanit:6 rubel:1 improvement:1 indicates:1 elsevier:1 entire:1 initially:1 selective:1 lgn:2 orientation:23 eurobiology:1 equal:4 field:5 sampling:6 identical:1 adelson:1 stimulus:3 randomly:1 manipulated:1 composed:1 harel:1 national:1 individual:1 organization:3 interest:1 light:1 necessary:3 poggio:2 tree:2 old:1 desired:3 plotted:1 minimal:2 column:12 patchy:9 maximization:5 shiftable:4 snr:1 conducted:1 dependency:1 periodic:1 combined:1 density:1 fundamental:2 peak:2 receiving:2 possibly:1 creating:1 steered:1 derivative:1 diversity:21 coefficient:9 matter:1 vi:4 depends:1 crossed:1 stream:1 try:1 iwi:1 simon:1 vivo:1 il:1 compartment:1 formed:2 accuracy:1 oi:5 maximized:1 yield:1 wisdom:1 spaced:1 ofthe:1 hyperbola:1 confirmed:1 explain:1 detector:1 energy:2 resultant:1 associated:2 di:1 sampled:1 proved:1 organized:2 feed:1 attained:3 response:15 just:1 binocular:2 anywhere:1 stage:1 correlation:1 receives:1 horizontal:2 multiscale:1 overlapping:5 widespread:1 bjp:1 aj:2 perhaps:1 reveal:1 usa:1 effect:2 hypothesized:1 contain:2 former:1 regularization:1 spatially:2 round:1 adjacent:1 substantiate:1 manifestation:1 prominent:1 demonstrate:1 tn:1 motion:2 percent:1 image:5 consideration:1 recently:2 functional:8 jp:1 cerebral:3 linking:1 extend:1 tuning:4 outlined:1 mathematics:1 grid:1 cortex:13 surface:1 dominant:1 recent:1 selectivity:4 certain:1 manifested:1 steering:7 freely:2 determine:1 signal:6 ii:1 dashed:1 simoncelli:1 rj:5 smooth:3 match:5 calculation:1 cross:2 long:1 concerning:1 equally:1 a1:1 multilayer:1 vision:1 foremost:1 histogram:4 cell:6 receive:1 separately:1 interval:1 pyramidal:1 source:1 modality:1 ot:1 rest:1 subject:1 tend:3 axonal:11 near:1 presence:2 revealed:3 variety:1 xj:3 architecture:8 bandwidth:1 idea:1 shift:2 grill:9 york:1 cause:1 clear:1 amount:2 diameter:5 percentage:1 shifted:1 neuroscience:2 per:1 anatomical:9 broadly:1 rehovot:1 discrete:5 diverse:1 group:1 imaging:3 run:1 angle:4 striking:1 throughout:2 patch:24 separation:1 layer:1 ki:3 followed:1 encountered:1 discretely:1 activity:2 precisely:1 interpolated:1 fourier:1 optical:3 injection:1 according:2 combination:6 across:5 slightly:1 em:3 sam:1 primate:1 taken:1 ln:1 equation:2 computationally:1 histologically:1 operation:1 arborize:1 original:2 denotes:1 const:1 already:1 receptive:5 striate:3 aertsen:1 distance:1 link:1 outer:1 participate:1 modeled:3 relationship:2 ratio:7 demonstration:1 minimizing:1 mostly:1 stated:1 rise:1 perform:1 allowing:1 upper:1 vertical:1 neuron:17 pinwheel:1 extended:1 incorporated:1 precise:2 neurobiology:1 discovered:1 somatic:1 arbitrary:1 pair:1 required:3 connection:9 macaque:1 suggested:1 bar:1 pattern:5 lund:7 rf:34 shifting:4 vidyasagar:3 overlap:2 suitable:1 inversely:2 carried:1 hm:1 faced:2 review:1 literature:1 determining:1 generation:1 proportional:3 oos:2 degree:1 xp:1 principle:4 editor:1 compatible:1 summary:1 placed:2 bias:1 institute:1 explaining:1 steerability:1 taking:1 tracing:1 distributed:4 van:1 calculated:3 cortical:22 valid:1 sensory:1 made:1 transaction:1 approximate:1 rafael:5 preferred:2 hubel:1 stereopsis:2 spectrum:1 continuous:1 learn:1 nature:1 obtaining:1 dendrite:1 pjp:2 domain:11 did:1 significance:1 spread:1 border:3 noise:4 profile:2 allowed:1 neuronal:6 site:2 levitt:1 scattered:1 wiley:1 position:2 originated:2 heeger:1 grinvald:3 shimon:5 specific:1 physiological:1 intrinsic:6 adding:1 ci:1 cartesian:1 depts:1 columnar:3 likely:1 forming:1 visual:7 g2:1 bo:1 ma:1 goal:1 targeted:1 rbf:1 change:1 determined:1 specie:3 arbor:15 tendency:2 selectively:1 support:1 latter:1 arises:2
7,194
955
Factorial Learning by Clustering Features Joshua B. Tenenbaum and Emanuel V. Todorov Department of Brain and Cognitive Sciences Massachusetts Institute of Technology Cambridge, MA 02139 {jbt.emo}~psyche . mit.edu Abstract We introduce a novel algorithm for factorial learning, motivated by segmentation problems in computational vision, in which the underlying factors correspond to clusters of highly correlated input features. The algorithm derives from a new kind of competitive clustering model, in which the cluster generators compete to explain each feature of the data set and cooperate to explain each input example, rather than competing for examples and cooperating on features, as in traditional clustering algorithms. A natural extension of the algorithm recovers hierarchical models of data generated from multiple unknown categories, each with a different, multiple causal structure. Several simulations demonstrate the power of this approach. 1 INTRODUCTION Unsupervised learning is the search for structure in data. Most unsupervised learning systems can be viewed as trying to invert a particular generative model of the data in order to recover the underlying causal structure of their world. Different learning algorithms are then primarily distinguished by the different generative models they embody, that is, the different kinds of structure they look for. Factorial learning, the subject of this paper, tries to find a set of independent causes that cooperate to produce the input examples. We focus on strong factorial learning, where the goal is to recover the actual degrees of freedom responsible for generating the observed data, as opposed to the more general weak approach, where the goal 562 Joshua B. Tenenbaum, Emmanuel V. Todorov Figure 1: A simple factorial learning problem. The learner observes an articulated hand in various configurations, with each example specified by the positions of 16 tracked features (shown as black dots). The learner might recover four underlying factors, corresponding to the positions of the fingers, each of which claims responsiblity for four features of the data set. is merely to recover some factorial model that explains the data efficiently. Strong factorial learning makes a claim about the nature of the world, while weak factorial learning only makes a claim about the nature of the learner's representations (although the two are clearly related). Standard subspace algorithms, such as principal component analysis, fit a linear, factorial model to the input data, but can only recover the true causal structure in very limited situations, such as when the data are generated by a linear combination of independent factors with significantly different variances (as in signal-from-noise separation). Recent work in factorial learning suggests that the general problem of recovering the true, multiple causal structure of an arbitrary, real-world data set is very difficult, and that specific approaches must be tailored to specific, but hopefully common, classes of problems (Foldiak, 1990; Saund, 1995; Dayan and Zemel, 1995). Our own interest in multiple cause learning was motivated by segmentation problems in computational vision, in which the underlying factors correspond ideally to disjoint clusters of highly correlated input features. Examples include the segmentation of articulated objects into functionally independent parts, or the segmentation of multiple-object motion sequences into tracks of individual objects. These problems, as well as many other problems of pattern recognition and analysis, share a common set of constraints which makes factorial learning both appropriate and tractable. Specifically, while each observed example depends on some combination of several factors, anyone input feature always depends on only one such factor (see Figure 1). Then the generative model decomposes into independent sets of functionally grouped input features, or functional parts (Tenenbaum, 1994). In this paper, we propose a learning algorithm that extracts these functional parts. The key simplifying assumption, which we call the membership constmint, states that each feature belongs to at most one functional part, and that this membership is constant over the set of training examples. The membership constraint allows us to treat the factorial learning problem as a novel kind of clustering problem. The cluster generators now compete to explain each feature of the data set and cooperate to explain each input example, rather than competing for examples and cooperating on features, as in traditional clustering systems such as K-means or mixture models. The following sections discuss the details of the feature clustering algorithm for extracting functional parts, a simple but illustrative example, and extensions. In particular, we demonstrate a natural way to relax the strict membership constraint and thus learn hierarchical models of data generated from multiple unknown categories, each with a different multiple causal structure. Factorial Learning by Clustering Features 2 563 THE FEATURE CLUSTERING ALGORITHM Our algorithm for extracting functional parts derives from a statistical mechanics formulation of the soft clustering problem (inspired by Rose, Gurewitz, and Fox, 1990; Hinton and Zemel, 1994). We take as input a data set {xt}, with I examples of J real-valued features. The best K-cluster representation of these J features is given by an optimal set of cluster parameters, {Ok}, and an optimal set of assignments, {Pi/.J. The assignment Pjk specifies the probability of assigning feature j to cluster k, and depends directly on Ejk = 2:i(Xj'l - fj~)2, the total squared difference (over the I training examples) between the observed feature values x}il and cluster k's predictions fj~. The parameters Ok define cluster k's generative model, and thus determine the predictions fj~(Ok). If we limit functional parts to clusters of linearly correlated features, then the appropriate generative model has fj~ = WjkYkil + Uj, with cluster parameters Ok = {Ykil , Wjk, Uj} to be estimated. That is, for each example i, part k predicts the value of input feature j as a linear function of some part-specific factor Ykil (such as finger position in Figure 1). For the purposes of this paper, we assume zero-mean features and ignore the Uj terms. Then Ejk = 2:i(Xj'l - WjkYkil )2. The optimal cluster parameters and assignments can now be found by maximizing the complete log likelihood of the data given the K-cluster representation, or equivalently, in the framework of, statistical mechanics, by minimizing the free energy F =E - 1 -H (3 1 = LLPjk(Ejk + -logpjk) (3 k j (1) subject to the membership constraints, 2:k pjk = 1, (\lj). Minimizing the energy, E = L j LPjkEjk, (2) k reduces the expected reconstruction error, leading to more accurate representations. Maximizing the entropy, (3) H = - LLPjklogPjk, j k distributes responsibility for each feature across many parts, thus decreasing the independence of the parts and leading to simpler representations (with fewer degrees of freedom). In line with Occam's Razor, minimizing the energy-entropy tradeoff finds the representation that, at a particular temperature 1/(3, best satisfies the conflicting requirements of low error and low complexity. We minimize the free energy with a generalized EM procedure (Neal and Hinton, 1994), setting derivatives to zero and iterating the resulting update equations: e-(3Ejk Pjk = Ykil 2:k e-(3Ejk LPjkWjkXj'l (4) (5) j Wjk = Lxj'lYtl. (6) 564 Joshua B. Tenenbaum. Emmanuel V. Todorov This update procedure assumes a normalization step Ykil = Ykil /(Eil (y(l)2)1/2 in each iteration, because without some additional constraint on the magnitudes of l (or Wjk), inverting the generative model fj~ = WjkYkil is an ill-posed problem. yt This algorithm maps naturally onto a simple network architecture. The hidden unit activities, representing the part-specific factors Ykil , are computed from the observations via bottom-up weights PjkWjk, normalized, and multiplied by top-down weights Wjk to generate the network's predictions fj~. The weights adapt according to a hybrid learning rule, with Wjk determined by a Hebb rule (as in subspace learning algorithms), and pjk determined by a competitive, softmax function of the reconstruction error Ejk (as in soft mixture models). xyl 3 LEARNING A HIERARCHY OF PARTS The following simulation illustrates the algorithm's behavior on a simple, part segmentation task. The training data consist of 60 examples with 16 features each, representing the horizontal positions of 16 points on an articulated hand in various configurations (as in Figure 1). The data for this example were generated by a hierarchical, random process that produced a low correlation between all 16 features, a moderate correlation between the four features on each finger, and a high correlation between the two features on each joint (two joints per finger). To fully explain this data set, the algorithm should be able to find a corresponding hierarchy of increasingly complex functional part representations. To evaluate the network's representation of this data set, we inspect the learned weights PjkWjk, which give the total contribution of feature j to part k in (5) . In Figure 2, these weights are plotted for several different values of /3, with gray boxes indicating zero weights, white indicating strong positive weights, and black indicating strong negative weights. The network was configured with K = 16 part units, to ensure that all potential parts could be found. When fewer than K distinct parts are found, some of the cluster units have identical parameters (appearing as identical columns in Figure 2). These results were generated by deterministic annealing, starting with /3 ? 1, and perturbing the weights slightly each time /3 was increased, in order to break symmetries. Figure 2 shows that the number of distinct parts found increases with /3, as more accurate (and more complex) representations become favored. In (4), we see that /3 controls the number of distinct parts via the strength of the competition for features. At /3 = 0, every part takes equal responsibility for every feature. Without competition, there can be no diversity, and thus only one distinct part is discovered at low /3, corresponding to the whole hand (Figure 2a). As /3 increases, the competition for features gets stiffer, and parts split into their component subparts. The network finds first four distinct parts (with four features each), corresponding to individual fingers (Figure 2c), and then eight distinct parts (with two features each), corresponding to individual joints (Figure 2d). Figure 2b shows an intermediate representation, with something between one and four parts. Four distinct columns are visible, but they do not cleanly segregate the features. Figure 3 plots the decrease in mean reconstruction error (expressed by the energy E) Factorial Learning by Clustering Features (a) 565 ~=1 ~= (b) 100 ~ ~~+-~~+-~~~~~~ ~ ~~+-r+~+-~-r~r+-r~ ~ r4~+-r+~4=P+-r~r+~~ Part k Part k (e) ~= 1000 ~=2oo00 (d) Part k Part k Figure 2: A hierarchy of functional part representations, parameterized by {3. F - - - e - - - - - -___ b d 2 3 4 log ~ Figure 3: A phase diagram distinguishes true parts (a, c, d) from spurious ones (b). 566 Joshua B. Tenenbaum. Emmanuel V. Todorov as (3 increases and more distinct parts emerge. Notice that within the three stable phases corresponding to good part decompositions (Figures 2a, 2c, 2d), E remains practically constant over wide variations in (3. In contrast, E varies rapidly at the boundaries between phases, where spurious part structure appears (Figure 2b). In general, good representations should lie at stable points of this phase diagram, where the error-complexity tradeoff is robust. Thus the actual number of parts in a particular data set, as well as their hierarchical structure, need not be known in advance, but can be inferred from the dynamics of learning. 4 LEARNING MULTIPLE CATEGORIES Until this point, we have assumed that each feature belongs to at most one part over the entire set of training examples, and tried to find the single K -part model that best explains the data as a whole. But the notion that a single model must explain the whole data set is quite restrictive. The data may contain several categories of examples, each characterized by a different pattern of feature correlations, and then we would like to learn a set of models, each capturing the distinctive part structure of one such category. Again we are motivated by human vision, which easily recognizes many categories of motion defined by high-level patterns of coordinated part movement, such as hand gestures and facial expressions. If we know which examples belong to which categories, learning multiple models is no harder than learning one, as in the previous section. A separate model can be fit to each category m of training examples, and the weights P'j'k wjk are frozen to produce a set of category templates. However, if the category identities are unknown, we face a novel kind of hierarchical learning task. We must simultaneously discover the optimal clustering of examples into categories, as well as the optimal clustering of features into parts within each category. We can formalize this hierarchical clustering problem as minimizing a familiar free energy, (7) in which gim. specifies the probability of assigning example i to category m, and Tim. is the associated cost. This cost is itself the free energy of the mth K -part model on the ith example, Tim = L LP'j'k(Ejk' j + _(31 logp'j'k) , (8) k in which P'j'k specifies the probability of assigning feature j to part k within category m, and Ejk' = (xj - Wjky~m.)2 is the usual reconstruction error from Section 2. This algorithm was tested on a data set of 256 hand configurations with 20 features each (similar to those in Figure 1), in which each example expresses one of four possible "gestures", i.e. patterns of feature correlation. As Table 1 indicates, the five features on each finger are highly correlated across the entire data set, while variable correlations between the four fingers distinguish the gesture categories. Note that a single model with four parts explains the full variance of the data just as well as the actual four-category generating process. However, most of the data Factorial Learning by Clustering Features 567 Table 1: The 20 features are grouped into either 2, 3, or 4 functional parts. Examples 1- 64 65 - 127 128 - 192 193 - 256 No. of parts 2 3 3 4 1 1 1-5 1-5 Part composition 10 11----20 10 11-15 16--20 6-10 11 20 6--10 11-15 16--20 can also be explained by one of several simpler models, making the learner's task a challenging balancing act between accuracy and simplicity. Figure 4 shows a typical representation learned for this data set. The algorithm was configured with M = 8 category models (each with K = 8 parts), but only four distinct categories of examples are found after annealing on a (holding {3 constant), and their weights prk wjk are depicted in Figure 4a. Each category faithfully captures one of the actual generating categories in Table 1, with the correct number and composition of functional parts. Figure 4b depicts the responsibility gi'm that each learned category m takes for each feature i. Notice the inevitable effect of a bias towards simpler representations. Many examples are misassigned relative to Table 1, when categories with fewer degrees of freedom than their true generating categories can explain them almost as accurately. 5 CONCLUSIONS AND FUTURE DIRECTIONS The notion that many data sets are best explained in terms of functionally independent clusters of correlated features resonates with similar proposals of Foldiak (1990), Saund (1995), Hinton and Zemel (1994), and Dayan and Zemel (1995). Our approach is unique in actually formulating the learning task as a clustering problem and explicitly extracting the functional parts of the data. Factorial learning by clustering features has three principal advantages. First, the free energy cost function for clustering yields a natural complexity scale-space of functional part representations, parameterized by {3. Second, the generalized EM learning algorithm is simple and quick, and maps easily onto a network architecture. Third, by nesting free energies, we can seamlessly compose objective functions for quite complex, hierarchical unsupservised learning problems, such as the multiple category, multiple part mixture problem of Section 4. The primary limitation of our approach is that when the generative model we assume does not in fact apply to the data, the algorithm may fail to recover any meaningful structure. In ongoing work, we are pursuing a more flexible generative model that allows the underlying causes to compete directly for arbitrary featureexample pairs ij, rather than limiting competition only to features j, as in Section 2, or only to examples i, as in conventional mixture models, or segregating competition for examples and features into hierarchical stages, as in Section 4. Because this introduces many more degrees of freedom, robust learning will require additional constraints, such as temporal continuity of examples or spatial continuity of features. 568 Joshua B. Tenenbaum, Emmanuel V. Todorov (a) Category 1 category 2 category 3 category 4 Part k Part k Part k Part k (b) 1 E 521 ..--------~..~w.~? Cl (*) 31------~????? . . . . .. 64 128 Example i 192 256 Figure 4: Learning multiple categories, each with a different part structure. Acknowledgements Both authors are Howard Hughes Medical Institute Predoctoral Fellows. We thank Whitman Richards, Yair Weiss, and Stephen Gilbert for helpful discussions. References Dayan, P. and Zemel, R. S. (1995). Computation, in press. Competition and multiple cause models. Neural Foldiak, P. (1990). Forming sparse representations by local anti-hebbian learning. Biological Cybernetics 64, 165-170. Hinton, G. E . and Zemel, R. S. (1994). Autoencoders, minimum description length and Helmholtz free energy. In J. D. Cowan, G. Tesauro, & J. Alspector (eds.), Advances in Neural Information Processing Systems 6. San Mateo, CA: Morgan Kaufmann, 3-10. Neal, R. M., and Hinton, G. E. (1994). A new view of the EM algorithm that justifies incremental and other variants. Rose, K., Gurewitz, F., and Fox, G. (1990). Statistical mechanics and phase transitions in clustering. Physical Review Letters 65, 945-948. Saund, E. (1995). A mUltiple cause mixture model for unsupervised learning. Neural Computation 7,51-71. Tenenbaum, J . (1994) . Functional parts. In A. Ram & K. Eiselt (eds.), Proceedings of the Sixteenth Annual Conference of the Cognitive Science Society. Hillsdale, N J: Lawrence Erlbaum, 864-869.
955 |@word cleanly:1 simulation:2 tried:1 simplifying:1 decomposition:1 harder:1 configuration:3 assigning:3 must:3 visible:1 plot:1 prk:1 update:2 generative:8 fewer:3 ith:1 simpler:3 five:1 become:1 compose:1 introduce:1 expected:1 behavior:1 alspector:1 embody:1 mechanic:3 brain:1 inspired:1 decreasing:1 eil:1 actual:4 discover:1 underlying:5 kind:4 temporal:1 fellow:1 every:2 act:1 control:1 unit:3 medical:1 positive:1 local:1 treat:1 limit:1 black:2 might:1 mateo:1 r4:1 suggests:1 challenging:1 limited:1 unique:1 responsible:1 hughes:1 procedure:2 significantly:1 get:1 onto:2 gilbert:1 conventional:1 map:2 deterministic:1 yt:1 maximizing:2 quick:1 starting:1 simplicity:1 rule:2 nesting:1 notion:2 variation:1 limiting:1 hierarchy:3 helmholtz:1 recognition:1 richards:1 predicts:1 observed:3 bottom:1 capture:1 decrease:1 movement:1 observes:1 lxj:1 rose:2 complexity:3 ideally:1 dynamic:1 distinctive:1 learner:4 whitman:1 easily:2 joint:3 various:2 finger:7 articulated:3 distinct:9 zemel:6 quite:2 posed:1 valued:1 relax:1 gi:1 itself:1 sequence:1 advantage:1 frozen:1 propose:1 reconstruction:4 rapidly:1 sixteenth:1 description:1 competition:6 wjk:7 cluster:15 requirement:1 produce:2 generating:4 incremental:1 object:3 tim:2 ij:1 strong:4 recovering:1 direction:1 correct:1 human:1 hillsdale:1 explains:3 require:1 pjk:4 biological:1 extension:2 practically:1 lawrence:1 claim:3 purpose:1 grouped:2 faithfully:1 mit:1 clearly:1 always:1 rather:3 focus:1 likelihood:1 indicates:1 seamlessly:1 contrast:1 helpful:1 dayan:3 membership:5 lj:1 entire:2 hidden:1 spurious:2 mth:1 ill:1 flexible:1 favored:1 spatial:1 softmax:1 equal:1 identical:2 look:1 unsupervised:3 inevitable:1 future:1 primarily:1 distinguishes:1 simultaneously:1 individual:3 familiar:1 phase:5 freedom:4 interest:1 highly:3 introduces:1 mixture:5 accurate:2 facial:1 fox:2 plotted:1 causal:5 increased:1 column:2 soft:2 logp:1 assignment:3 cost:3 erlbaum:1 varies:1 squared:1 again:1 opposed:1 gim:1 cognitive:2 derivative:1 leading:2 potential:1 diversity:1 configured:2 coordinated:1 explicitly:1 depends:3 try:1 saund:3 responsibility:3 break:1 view:1 competitive:2 recover:6 contribution:1 minimize:1 il:1 accuracy:1 variance:2 kaufmann:1 efficiently:1 correspond:2 yield:1 weak:2 accurately:1 produced:1 cybernetics:1 explain:7 ed:2 energy:10 naturally:1 associated:1 unsupservised:1 recovers:1 emanuel:1 massachusetts:1 segmentation:5 formalize:1 actually:1 appears:1 ok:4 wei:1 formulation:1 box:1 just:1 stage:1 correlation:6 until:1 hand:5 autoencoders:1 horizontal:1 hopefully:1 continuity:2 gray:1 effect:1 normalized:1 true:4 contain:1 jbt:1 neal:2 white:1 razor:1 illustrative:1 generalized:2 trying:1 complete:1 demonstrate:2 motion:2 temperature:1 fj:6 cooperate:3 novel:3 common:2 functional:13 physical:1 tracked:1 perturbing:1 belong:1 functionally:3 composition:2 cambridge:1 dot:1 stable:2 stiffer:1 something:1 own:1 recent:1 foldiak:3 belongs:2 moderate:1 tesauro:1 joshua:5 morgan:1 minimum:1 additional:2 determine:1 signal:1 stephen:1 multiple:14 full:1 reduces:1 hebbian:1 adapt:1 characterized:1 gesture:3 prediction:3 variant:1 vision:3 iteration:1 normalization:1 tailored:1 invert:1 proposal:1 annealing:2 diagram:2 strict:1 subject:2 cowan:1 call:1 extracting:3 intermediate:1 split:1 todorov:5 xj:3 fit:2 independence:1 architecture:2 competing:2 tradeoff:2 motivated:3 expression:1 cause:5 iterating:1 factorial:16 tenenbaum:7 category:29 generate:1 specifies:3 notice:2 estimated:1 disjoint:1 track:1 per:1 subpart:1 express:1 key:1 four:12 segregating:1 ram:1 cooperating:2 merely:1 compete:3 parameterized:2 letter:1 almost:1 pursuing:1 separation:1 capturing:1 distinguish:1 annual:1 activity:1 strength:1 constraint:6 anyone:1 formulating:1 department:1 according:1 combination:2 across:2 slightly:1 em:3 increasingly:1 psyche:1 lp:1 making:1 explained:2 equation:1 remains:1 emo:1 discus:1 fail:1 know:1 tractable:1 multiplied:1 eight:1 apply:1 hierarchical:8 appropriate:2 appearing:1 distinguished:1 yair:1 assumes:1 clustering:18 include:1 top:1 ensure:1 recognizes:1 restrictive:1 emmanuel:4 uj:3 society:1 objective:1 primary:1 usual:1 traditional:2 subspace:2 separate:1 thank:1 length:1 minimizing:4 equivalently:1 difficult:1 holding:1 negative:1 unknown:3 predoctoral:1 inspect:1 observation:1 howard:1 ejk:8 anti:1 situation:1 hinton:5 segregate:1 discovered:1 arbitrary:2 inferred:1 inverting:1 pair:1 specified:1 learned:3 conflicting:1 able:1 pattern:4 power:1 natural:3 hybrid:1 representing:2 technology:1 extract:1 gurewitz:2 review:1 acknowledgement:1 relative:1 fully:1 limitation:1 generator:2 degree:4 share:1 pi:1 occam:1 balancing:1 free:7 bias:1 institute:2 wide:1 template:1 face:1 emerge:1 sparse:1 boundary:1 world:3 transition:1 author:1 san:1 ignore:1 assumed:1 search:1 decomposes:1 table:4 nature:2 learn:2 robust:2 ca:1 symmetry:1 complex:3 cl:1 linearly:1 whole:3 noise:1 depicts:1 hebb:1 position:4 resonates:1 lie:1 third:1 down:1 specific:4 xt:1 derives:2 consist:1 magnitude:1 illustrates:1 justifies:1 entropy:2 depicted:1 forming:1 expressed:1 satisfies:1 ma:1 viewed:1 goal:2 identity:1 towards:1 specifically:1 determined:2 typical:1 distributes:1 principal:2 total:2 meaningful:1 indicating:3 logpjk:1 ongoing:1 evaluate:1 tested:1 correlated:5
7,195
956
Predicting the Risk of Complications in Coronary Artery Bypass Operations using Neural Networks Richard P. Lippmann, Linda Kukolich MIT Lincoln Laboratory 244 Wood Street Lexington, MA 02173-0073 Dr. David Shahian Lahey Clinic Burlington, MA 01805 Abstract Experiments demonstrated that sigmoid multilayer perceptron (MLP) networks provide slightly better risk prediction than conventional logistic regression when used to predict the risk of death, stroke, and renal failure on 1257 patients who underwent coronary artery bypass operations at the Lahey Clinic. MLP networks with no hidden layer and networks with one hidden layer were trained using stochastic gradient descent with early stopping. MLP networks and logistic regression used the same input features and were evaluated using bootstrap sampling with 50 replications. ROC areas for predicting mortality using preoperative input features were 70.5% for logistic regression and 76.0% for MLP networks. Regularization provided by early stopping was an important component of improved perfonnance. A simplified approach to generating confidence intervals for MLP risk predictions using an auxiliary "confidence MLP" was developed. The confidence MLP is trained to reproduce confidence intervals that were generated during training using the outputs of 50 MLP networks trained with different bootstrap samples. 1 INTRODUCTION In 1992 there were roughly 300,000 coronary artery bypass operations perfonned in the United States at a cost of roughly $44,000 per operation. The $13.2 billion total cost of these operations is a significant fraction of health care spending in the United States. This has led to recent interest in comparing the quality of cardiac surgery across hospitals using risk-adjusted procedures and large patient populations. It has also led to interest in better assessing risks for individual patients and in obtaining improved understanding of the patient and procedural characteristics that affect cardiac surgery outcomes. 1056 Richard P. Lippmann, Linda Kukolich, David Shahian INPUT FEATURES ~ -... -.... ... ~ -~ SELECT I-FEATURES I-AND REPLACE MISSING FEATURES r-r-- CLASSIFY -... CONFIDENCE NETWORK ... ~ RISK PROBABILITY CONFIDENCE INTERVAL Figure 1. Block diagram of a medical risk predictor. This paper describes a experiments that explore the use of neural networks to predict the risk of complications in coronary artery bypass graft (CABO) surgery. Previous approaches to risk prediction for bypass surgery used linear or logistic regression or a Bayesian approach which assumes input features used for risk prediction are independent (e.g. Edwards, 1994; Marshall, 1994; Higgins, 1992; O'Conner, 1992). Neural networks have the potential advantages of modeling complex interactions among input features, of allowing both categorical and continuous input features, and of allowing more flexibility in fitting the expected risk than a simple linear or logistic function. 2 RISK PREDICTION AND FEATURE SELECTION A block diagram of the medical risk prediction system used in these experiments is shown in Figure 1. Input features from a patient's medical record are provided as 105 raw inputs, a smaller subset of these features is selected, missing features in this subset are replaced with their most likely values from training data, and a reduced input feature vector is fed to a classifier and to a "confidence network". The classifier provides outputs that estimate the probability or risk of one type of complication. The confidence network provides upper and lower bounds on these risk estimates. Both logistic regression and multilayer sigmoid neural network (MLP) classifiers were evaluated in this study. Logistic regression is the most common approach to risk prediction. It is structurally equivalent to a feed-forward network with linear inputs and one output unit with a sigmoidal nonlinearity. Weights and offsets are estimated using a maximum likelihood criterion and iterative "batch" training. The reference logistic regression classifier used in these experiments was implemented with the SPlus glm function (Mathsoft, 1993) which uses iteratively reweighted least squares for training and no extra regularization such as weight decay. Multilayer feed-forward neural networks with no hidden nodes (denoted single-layer MLPs) and with one hidden layer and from 1 to 10 hidden nodes were also evaluated as implemented using LNKnet pattern classification software (Lippmann, 1993). An MLP committee classifier containing eight members trained using different initial random weights was also evaluated. All classifiers were evaluated using a data base of 1257 patients who underwent coronary artery bypass surgery from 1990 to 1994. Classifiers were used to predict mortality, postoperative strokes, and renal failure. Predictions were made after a patient's medical history was obtained (History), after pre-surgical tests had been performed (Post-test), immediately before the operation (preop), and immediately after the operation (Postop). Bootstrap sampling (Efron, 1993) was used to assess risk prediction accuracy because there were so few Predicting the Risk of Complications in Coronary Artery Bypass Operations 1057 patients with complications in this data base. The number of patients with complications was 33 or 2.6% for mortality, 25 or 2.0% for stroke, and 21 or 1.7% for renal failure. All experiments were performed using 50 bootstrap training sets where a risk prediction technique is trained with a bootstrap training set and evaluated using left-out patterns. NComplications NHigh mSTORY Age COPD (Chronic Obs. Pul. Disease) % True Hits 27/674 71126 5.6% 8nl 61105 6/21 11.3% 5.7% 26.6% 211447 111115 10/127 7/64 4.7% 6.6% 7.9% 10.9% 121113 91184 10.6% 4.9% 4.0% POST?TEST Pulmonary Ventricular Congestion X-ray Cardiomegaly X-ray Pulmonary Edema PREOP NTG (Nitroglycerin) IABP (Intraaortic Balloon Pump) Urgency Status MI When POSTOP Blood Used (Packed Cells) Perfusion Time Figure 2. Features selected to predict mortality. The initial set of 105 raw input features included binary (e.g. MalelFemale), categorical (e.g. MI When: none, old, recent, evolving), and continuous valued features (e.g. Perfusion Time, Age). There were many missing and irrelevant features and all features were only weakly predictive. Small sets of features were selected for each complication using the following procedures: (1) Select those 10 to 40 features experience and previous studies indicate are related to each complication, (2) Omit features if a univariate contingency table analysis shows the feature is not important, (3) Omit features that are missing for more than 5% of patients, (4) Order features by number of true positives, (5) Omit features that are similar to other features keeping the most predictive, and (7) Add features incrementally as a patient's hospital interaction progresses. This resulted in sets of from 3 to 11 features for the three complications. Figure 2 shows the 11 features selected to predict mortality. The first column lists the features, the second column presents a fraction equal to the number of complications when the feature was "high" divided by the number of times this feature was "high" (A threshold was assigned for continuous and categorical features that provided good separation), and the last column is the second column expressed as a percentage. Classifiers were provided identical sets of input features for all experiments. Continuous inputs to all classifiers were normalized to have zero mean and unit variance, categorical inputs ranged from -(D-1)/2 to (D-1)/2 in steps of 1.0, where D is the number of categories, and binary inputs were -0.5 or 0.5 . 3 PERFORMANCE COMPARISONS Risk prediction was evaluated by plotting and computing the area under receiver operating characteristic (ROC) curves and also by using chi-square tests to determine how accurately classifiers could stratify subjects into three risk categories. Automated experiments were performed using bootstrap sampling to explore the effect of varying the training step size J058 Richard P. Lippmann. Linda Kukolich, David Shahian 100 ~ 80 :~ ~ c(/) 60 Q) en ~ 40 " I HISTORY (68.6%) ?fl. 20 o o 20 40 60 80 100 0 20 40 60 % FALSE ALARMS (100 - Specificity) 80 100 Figure 3. Fifty preoperative bootstrap ROCs predicting mortality using an MLP classifier with two hidden nodes and the average ROC (left), and average ROCS for mortality using history, preoperative, and postoperative features (right). from 0.005 to 0.1; of using squared-error, cross-entropy, and maximum likelihood cost functions; of varying the number of hidden nodes from 1 to 8; and of stopping training after from 5 to 40 epochs. ROC areas varied little as parameters were varied. Risk stratification, which measures how well classifier outputs approximate posterior probabilities, improved substantially with a cross-entropy cost function (instead of squared error), with a smaller stepsize (0.01 instead of 0.05 or 0.1) and with more training epochs (20 versus 5 or 10). An MLP classifier with two hidden nodes provided good overall performance across complications and patient stages with a cross-entropy cost function, a stepsize of 0.01, momentum of 0.6, and stochastic gradient descent stopping after 20 epochs. A single-layer MLP provided good performance with similar settings, but stopping after 5 epochs. These settings were used for all experiments. The left side of Figure 3 shows the 50 bootstrap ROCs created using these settings for a two-hidden-node MLP when predicting mortality with preoperative features and the ROC created by averaging these curves. There is a large variability in these ROes due to the small amount of training data. The ROC area varies from 67% to 85% (cr=4.7) and the sensitivity with 20% false alarms varies from 30% to 79%. Similar variability occurs for other complications. The right side of Figure 3 shows average ROCs for mortality created using this MLP with history, preoperative, and postoperative features. As can be seen, the ROC area and prediction accuracy increases from 68.6% to 79.2% as more input features become available. Figure 4 shows ROC areas across all complications and patient stages. Only three and two patient stages are shown for stroke and renal failure because no extra features were added at the missing stages for these complications. ROC areas are low for all complications and range from 62% to 80%. ROC areas are highest using postoperative features, lowest using only history features, and increase as more features are added. ROC areas are highest for mortality (68 to 80%) and lower for stroke (62 to 71 %) and renal failure (62 to 67% ).The MLP classifier with two hidden nodes (MLP) always provided slightly higher ROC areas than logistic regression. The average increase with the MLP classifier was 2.7 percentage Predicting the Risk of Complications in Coronary Artery Bypass Operations 100 90 III Logistic MORTALITY -80 C ~ 70 ~ 60 8a: 1059 [3 Single?Layer MLP ? MLP ? MLP?Commillee 50 40 30 HISTORY POSTTEST PRE?OP PATIENT STAGE POST?OP 100 90 - 80 C ? w 70 100 STROKE 90 80 I=2a I=2a 70 ~ 60 8a: RENAL FAILURE 60 50 50 40 40 30 30 HISTORY POSTTEST PATIENT STAGE POST?OP HISTORY POST?OP PATIENT STAGE Figure 4. ROC areas across all complications and patient stages for logistic regression, single-layer MLP classifier, two-layer MLP classifier with two hidden nodes, and a committee classifier containing eight two-layer MLP classifiers trained using different random starting weights. points (the increase ranged from 0.3 to 5.5 points). The single-layer MLPclassifier also provided good performance. The average ROC area with the single-layer MLP was only 0.6 percentage points below that of the MLP with two hidden nodes. The committee using eight two-layer MLP classifiers performed no better than an individual two-layer MLP classifier. Classifier outputs were used to bin or stratify each patient into one of four risk levels (05%, 5-10%, and 10-100%) by treating the output as an estimate of the complication posterior probability. Figure 5 shows the accuracy of risk stratification for the MLP classifier for all complications. Each curve was obtained by averaging 50 individual curves obtained using bootstrap sampling as with the ROC curves. Individual curves were obtained by placing each patient into one of the three risk bins based on the MLP output. The x's represent the average MLP output for all patients in each bin. Open squares are the true percentage of patients in each bin who experienced a complication. The bars represent ?2 binomial deviations about the true patient percentages. Risk prediction is accurate if the x's are close to the squares and within the confidence intervals. As can be seen, risk prediction is accurate and close to the actual number of patients who experienced complications. It is difficult, however, to assess risk prediction given the limited numbers of patients in the two highest bins. For example, in Figure 5, the median number of patients with complications was only 2 out of 20 in the middle bin and 2 out of 13 in the upper bin. Good and similar risk stratification, as measured by a chi-square test, was provided by all classifiers. Differences between classifier predictions and true patient percentages were small and not statistically significant. 1060 Richard P. Lippmann, Linda Kukolich, David Shahian ~ ~------------------------------~ MORTALITY 0- PATIENT COUNT X - MLP OUTPUT 30 10 o ____.M____________________________ 40 ~ ~--------------------------r_--_, RENAL FAILURE 30 20 10 o ~ __ _w~ 0-5 ________ ~ __________ 5?10 ~ __ ~ 10-100 BIN PROBABILITY RANGE ('?o) Figure 5. Accuracy of MLP risk stratification for three complications using preoperative features. Open squares are true percentages of patients in each bin with a complication, x's are MLP predictions, bars represent ?2 binomial standard deviation confidence intervals. 4 CONFIDENCE MLP NETWORKS Estimating the confidence in the classification decision produced by a neural network is a critical issue that has received relatively little study. Not being able to provide a confidence measure makes it difficult for physicians and other professionals to accept the use of complex networks. Bootstrap sampling (Efron, 1993) was selected as an approach to generate confidence intervals for medical risk prediction because 1) It can be applied to any type of classifier, 2) It measures variability due to training algorithms, implementation differences, and limited training data, and 3) It is simple to implement and apply. As shown in the top half of Figure 6, 50 bootstrap sets of training data are created from the original training data by resampling with replacement. These bootstrap training sets are used to train 50 bootstrap MLP classifiers using the same architecture and training procedures that were selected for the risk prediction MLP. When a pattern is fed into these classifiers, their outputs provide an estimate of the distribution of the output of the risk prediction MLP. Lower and upper confidence bounds for any input are obtained by sorting these outputs and selecting the 10% and 90% cumulative levels. It is computationally expensive to have to maintain and query 50 bootstrap MLPs whenever confidence bounds are desired. A simpler approach is to train a single confidence MLP to replicate the confidence bounds predicted by the 50 bootstrap MLPs, as shown in the bot- Predicting the Risk of Complications in Coronary Artery Bypass Operations 1061 OUTPUT STATISTICS UPPER LIMIT / INPUT PATTERN RISK PREDICTION MLP CONFIDENCE MLP LOWER LIMIT Figure 6. A confidence MLP trained using 50 bootstrap MLPs produces upper and lower confidence bounds for a risk prediction MLP. tom half of Figure 6. The the confidence MLP is fed the input pattern and the output of the risk prediction MLP and produces at its output the confidence intervals that would have been produced by 50 bootstrap MLPs. The confidence MLP is a mapping or regression network that replaces the 50 bootstrap networks. It was found that confidence networks with one hidden layer, two hidden nodes, and a linear output could accurately reproduce the upper and lower confidence intervals created by 50 bootstrap two-layer MLP networks. The confidence network outputs were almost always within ?15% of the actual bootstrap bounds. Upper and lower bounds produced by these confidence networks for all patients using preoperative features predicting mortality are show in Figure 7. Bounds are high (? 10 percentage points) when the complication risk is near 20% and drop to lower values (?0.4 percentage points) when the risk is near 1%. This relatively simple approach makes it possible to create and replicate confidence intervals for many types of classifiers. 5 SUMMARY AND FUTURE PLANS MLP networks provided slightly better risk prediction than conventional logistic regression when used to predict the risk of death, stroke, and renal failure on 1257 patients who underwent coronary artery bypass operations. Bootstrap sampling was required to compare approaches and regularization provided by early stopping was an important component of improved performance. A simplified approach to generating confidence intervals for MLP risk predictions using an auxiliary "confidence MLP" was also developed. The confidence MLP is trained to reproduce the confidence bounds that were generated during training by 50 MLP networks trained using bootstrap samples. Current research is validating these results using larger data sets, exploring approaches to detect outlier patients who are so different from any training patient that accurate risk prediction is suspect, developing approaches to explaining which input features are important for an individual patient, and determining why MLP networks provide improved performance. Richard P. Lippmann, Linda Kukolich, David Shahian 1062 ~r------.------.-------r------' T 30 T '# ... T ... TT 025 I- J;.... 5 :J20 T T UPPER W .. (,) Z 15 W Q u:: o(,) Z 10 5 COMPLICATION RISK% Figure 7. Upper and lower confidence bounds for all patients and preoperative mortality risk predictions calculated using two MLP confidence networks. ACKNOWLEDGMENT This work was sponsored by the Department of the Air Force. The views expressed are those of the authors and do not reflect the official policy or position of the U.S. Government. We wish to thank Stephanie Moisakis and Anne Nilson at the Lahey Clinic and Yuchun Lee at Lincoln Laboratory for assistance in organizing and preprocessing the data. BIBLIOGRAPHY F. Edwards, R. Clark, and M. Schwartz. (1994) Coronary Artery Bypass Grafting: The Society of Thoracic Surgeons National Database Experience. In Annals Thoracic Surgery, Vol. 57, 12-19. Bradley Efron and Robert J. Tibshirani. (1993) An Introduction to the Bootstrap. Monographs on Statistics and Applied Probability 57, New York: Chapman and Hall (1993). T. Higgins, F. Estafanous, et. al. (1992) Stratification of Morbidity and Mortality Outcome by Preoperative Risk Factors in Coronary Artery Bypass Patients. In Journal of the American Medical Society, Vol. 267, No. 17,2344-2348. R. Lippmann, L. Kukolich, and E. Singer. (1993) LNKnet: Neural Network, Machine Learning, and Statistical Software for Pattern Classification. In Lincoln Laboratory Journal, Vol. 6, No.2, 249-268. Marshall Guillenno, Laurie W. Shroyer, et al. (1994) Bayesian-Logit Model for Risk Assessment in Coronary Artery Bypass Grafting, In Annals Thoracic Surgery, Vol. 57, 1492-5000. G. O'Conner, S. Plume, et. al. (1992) Multivariate Prediction of In-Hospital Mortality Associated with Coronary Artery Bypass Surgery. In Circulation, Vol. 85, No.6, 21102118. Statistical Sciences. (1993) S-PLUS Guide to Statistical and Mathematical Analyses, Version 3.2, Seattle: StatSci, a division of MathSoft, Inc.
956 |@word version:1 middle:1 replicate:2 logit:1 open:2 edema:1 initial:2 united:2 selecting:1 bradley:1 current:1 comparing:1 anne:1 treating:1 drop:1 sponsored:1 resampling:1 congestion:1 half:2 selected:6 record:1 provides:2 complication:27 node:10 sigmoidal:1 simpler:1 mathematical:1 become:1 replication:1 fitting:1 ray:2 expected:1 roughly:2 chi:2 little:2 actual:2 provided:11 estimating:1 linda:5 lowest:1 renal:8 substantially:1 developed:2 lexington:1 classifier:29 hit:1 schwartz:1 unit:2 medical:6 omit:3 before:1 positive:1 limit:2 plus:1 statsci:1 limited:2 range:2 statistically:1 acknowledgment:1 block:2 implement:1 bootstrap:23 procedure:3 area:12 evolving:1 confidence:35 pre:2 specificity:1 close:2 selection:1 ntg:1 risk:49 conventional:2 equivalent:1 demonstrated:1 missing:5 chronic:1 starting:1 immediately:2 higgins:2 population:1 annals:2 mathsoft:2 us:1 expensive:1 database:1 balloon:1 highest:3 disease:1 monograph:1 graft:1 trained:9 weakly:1 predictive:2 surgeon:1 division:1 train:2 query:1 outcome:2 larger:1 valued:1 statistic:2 postoperative:4 advantage:1 interaction:2 organizing:1 flexibility:1 lincoln:3 artery:13 billion:1 seattle:1 assessing:1 produce:2 generating:2 perfusion:2 measured:1 op:4 received:1 progress:1 edward:2 auxiliary:2 implemented:2 predicted:1 indicate:1 conner:2 stochastic:2 bin:9 government:1 adjusted:1 exploring:1 hall:1 mapping:1 predict:6 early:3 create:1 mit:1 always:2 r_:1 cr:1 varying:2 likelihood:2 detect:1 stopping:6 accept:1 hidden:14 reproduce:3 overall:1 among:1 classification:3 issue:1 denoted:1 plan:1 equal:1 sampling:6 stratification:5 identical:1 placing:1 chapman:1 future:1 richard:5 few:1 resulted:1 national:1 individual:5 replaced:1 shahian:5 replacement:1 maintain:1 mlp:54 interest:2 nl:1 accurate:3 experience:2 perfonnance:1 old:1 desired:1 classify:1 modeling:1 column:4 marshall:2 cost:5 deviation:2 subset:2 pump:1 predictor:1 varies:2 sensitivity:1 lee:1 physician:1 squared:2 reflect:1 mortality:16 containing:2 dr:1 american:1 potential:1 inc:1 performed:4 view:1 mlps:5 air:1 square:6 ass:2 accuracy:4 variance:1 who:6 characteristic:2 circulation:1 surgical:1 bayesian:2 raw:2 accurately:2 produced:3 none:1 j20:1 history:9 stroke:7 whenever:1 failure:8 associated:1 mi:2 efron:3 yuchun:1 feed:2 higher:1 tom:1 improved:5 evaluated:7 stage:8 assessment:1 incrementally:1 logistic:12 quality:1 effect:1 normalized:1 true:6 ranged:2 regularization:3 assigned:1 laboratory:3 death:2 iteratively:1 reweighted:1 assistance:1 during:2 criterion:1 tt:1 spending:1 sigmoid:2 common:1 urgency:1 significant:2 nonlinearity:1 had:1 operating:1 base:2 add:1 posterior:2 multivariate:1 recent:2 irrelevant:1 binary:2 seen:2 care:1 determine:1 cross:3 divided:1 post:5 prediction:28 regression:11 multilayer:3 patient:36 roe:1 represent:3 cell:1 interval:10 thoracic:3 diagram:2 median:1 extra:2 fifty:1 morbidity:1 subject:1 suspect:1 validating:1 member:1 near:2 iii:1 automated:1 affect:1 architecture:1 copd:1 stratify:2 lnknet:2 york:1 amount:1 category:2 reduced:1 generate:1 percentage:9 bot:1 estimated:1 per:1 tibshirani:1 vol:5 splus:1 procedural:1 threshold:1 four:1 blood:1 nilson:1 fraction:2 wood:1 almost:1 separation:1 ob:1 decision:1 layer:15 bound:10 fl:1 replaces:1 software:2 bibliography:1 ventricular:1 relatively:2 department:1 developing:1 across:4 slightly:3 cardiac:2 describes:1 smaller:2 stephanie:1 outlier:1 glm:1 computationally:1 count:1 committee:3 singer:1 fed:3 available:1 operation:11 eight:3 apply:1 stepsize:2 batch:1 professional:1 original:1 assumes:1 binomial:2 top:1 society:2 surgery:8 added:2 occurs:1 gradient:2 thank:1 street:1 difficult:2 robert:1 implementation:1 packed:1 policy:1 allowing:2 upper:9 descent:2 variability:3 varied:2 david:5 required:1 able:1 bar:2 below:1 pattern:6 perfonned:1 critical:1 force:1 predicting:8 created:5 categorical:4 health:1 epoch:4 understanding:1 determining:1 coronary:13 versus:1 age:2 clark:1 clinic:3 contingency:1 plotting:1 bypass:14 summary:1 last:1 keeping:1 side:2 guide:1 perceptron:1 explaining:1 underwent:3 curve:6 calculated:1 cumulative:1 forward:2 made:1 author:1 preprocessing:1 simplified:2 approximate:1 lippmann:7 grafting:2 status:1 receiver:1 continuous:4 iterative:1 why:1 table:1 obtaining:1 laurie:1 complex:2 official:1 alarm:2 en:1 roc:19 structurally:1 momentum:1 experienced:2 position:1 wish:1 pul:1 kukolich:6 burlington:1 offset:1 decay:1 list:1 false:2 sorting:1 entropy:3 led:2 explore:2 likely:1 univariate:1 expressed:2 preoperative:9 pulmonary:2 ma:2 replace:1 included:1 averaging:2 total:1 hospital:3 select:2
7,196
957
An Analog Neural Network Inspired by Fractal Block Coding Fernando J. Pineda Andreas G. Andreou The Applied Physics Laboratory The Johns Hopkins University Johns Hokins Road Laurel, MD 20723-6099 Dept. of Electrical & Computer Engineering The Johns Hopkins University 34th & Charles St. Baltimore, MD 21218 Abstract We consider the problem of decoding block coded data, using a physical dynamical system. We sketch out a decompression algorithm for fractal block codes and then show how to implement a recurrent neural network using physically simple but highly-nonlinear, analog circuit models of neurons and synapses. The nonlinear system has many fixed points, but we have at our disposal a procedure to choose the parameters in such a way that only one solution, the desired solution, is stable. As a partial proof of the concept, we present experimental data from a small system a 16-neuron analog CMOS chip fabricated in a 2m analog p-well process. This chip operates in the subthreshold regime and, for each choice of parameters, converges to a unique stable state. Each state exhibits a qualitatively fractal shape. 1. INTRODUCTION Sometimes, a nonlinear approach is the simplest way to solve a linear problem. This is true when computing with physical dynamical systems whose natural operations are nonlinear. In such cases it may be expensive, in terms of physical complexity, to linearize the dynamics. For example in neural computation active ion channels have highly non linear input-output behaviour (see Hille 1984). Another example is 796 Fernando Pineda. Andreas G. Andreou subthreshold CMOS VLSI technology 1. In both examples the physics that governs the operation of the active devices, gives rise to gain elements that have exponential transfer characteristics. These exponentials result in computing structures with non-linear dynamics. It is therefore worthwhile, from both scientific and engineering perspectives, to investigate the idea of analog computation by highly non-linear components. This paper, explores an approach for solving a specific linear problem with analog circuits that have nonlinear transfer functions. The computational task considered here is that of fractal block code decompression (see e.g. Jacquin, 1989). The conventional approach to decompressing fractal codes is essentially an excercise in solving a high-dimenional sparse linear system of equations by using a relaxation algorithm. The relaxation algorithm is performed by iteratively applying an affine transformation to a state vector. The iteration yields a sequence of state vectors that converges to a vector of decoded data. The approach taken in this paper is based on the observation that one can construct a physically-simple nonlinear dyanmical system whose unique stable fixed point coincides with the solution of the sparse linear system of equations. In the next section we briefly summarize the basic ideas behind fractal block coding. This is followed by a description of an analog circuit with physically-simple nonlinear neurons. We show how to set the input voltages for the network so that we can program the position of the stable fixed point. Finally , we present experimental results obtained from a test chip fabricated in a 2mm CMOS process. 2. FRACTAL BLOCK CODING IN A NUTSHELL Let the N-dimensional state vector I represent a one dimensional curve sampled on N points. An affine transformation of this vector is simply a transformation of the form I' = WI+B , where W is an NxN -element matrix and B is an N-component vector. This transformation can be iterated to produce a sequence of vectors I(O)... . ,I(n). The sequence converges to a unique final state 1* that is independent of the initial state 1(0) if the maximum eigenvalue A.max of the matrix W satisfies Amax <1. The uniqueness of the final state implies that to transmit the state r to a receiver, we can either transmit r directly, In or we can transmit Wand B and let the receiver perform the iteration to generate the latter case we say that Wand B constitute an encoding of the state 1*. For this encoding to be useful, the amount of data needed to transmit Wand B must be less than the amount of data needed to transmit r This is the case when Wand B are sparse and parameterized and when the total number of bits needed to transmit these parameters is less than the total number of bits needed to transmit the uncompressed state r. r Fractal block coding is a special case of the above approach. It amounts to choosing a lWe consider subthreshold analog VLSI., (Mead 1989; Andreou and Boahen, 1994). A simple subthreshold model is ~ = I~nfet) exp(K'Vgb)( exp( -vsb) -exp( -Vdb?) for NFETS, where 1C - 0.67 and I~ t) = 9.7 x 10-18 A. The voltage differences Vgb, ,vsb,and Vdb are in units of the thermal voltafje, Vth= 0.025V. We use a corresponding expression for PFETs of the from Ids = I~pfe exp( -K'Vgb)( exp(vsb) - exp(vdb?) where I~Pfet) =3.8xl0- 18 A. An Analog Neural Network Inspired by Fractal Block Coding 797 blocked structure for the matrix W. This structure forces large-scale features to be mapped into small-scale features. The result is a steady state r that represents a curve with self similar (actually self affine) features. As a concrete example of such a structure, consider the following transformation of the state I. for O~ i ~ N -1 2 (1) r i = w R12i - N + bR for N 2: ~ i ~ N-l This transformation has two blocks. The transformation of the first N/2 components of I depend on the parameters W Land b L while the transformation of the second N/2 components depend on the parameters WR, and bR . Consequently just four parameters completely specify this transformation. This transformation can be expressed as a single affine transformation as follows: wL /' 0 I'N/2-1 /' N12 /' N-l = 10 wL 1 N/2-1 1 WR wR bL + bL N/2 bR I N- 1 bR (2) The top and bottom halves of I I depend on the odd and even components of I respectively. This subsampling causes features of size I to be mapped into features of size 112. A subsampled copy of the state I with transformed intensities is copied into the top half of 1'. Similarly, a subsampled copy of the state I with transformed intensities is copied into the bottom half of 1'. If this transformation is iterated, the sequence of transformed vectors will converge provided the eigenvalues determined by WL and WR are all less than one (i.e. WL and WR < 1). Although this toy example has just four free parameters and is thus too trivial to be useful for actual compression applications, it does suffice to generate state vectors with fractal properties since at steady state, the top and bottom halves of I' differ from the entire curve by an affine transformation. In this paper we will not describe how to solve the inverse problem which consists of finding a parameterized affine transformation that produces a given final state T. We note, however, that it is a special (and simpler) case of the recurrent network training problem, since the problem is linear, has no hidden units and has only one fixed point. The reader is refered to (Pineda, 1988) or. for a least squares algorithm in the context of neural nets or to (Monroe and Dudbridge, 1992) for a least squares algorithm in the context of coding. 3. A CMOS NEURAL NETWORK MODEL Now that we have described the salient aspects of the fractal decompression problem, we tum to the problem of implementing an analog neural network whose nonlinear dynamics converges to the same fixed point as the linear system. Nonlinearity arises because we 798 Fernando Pineda, Andreas G. Andreou make no special effort to linearize the gain elements (controlled conductances and transconductances) of the implementation medium. In this section we first describe a simple neuron. Then we analyze the dynamics of a network composed of such neurons. Finally we describe how to program the fixed point in the actual physical network. 3.1 The analog Neuron \\(~t) woulgll ) like to create a neuron model that calculates the transformation I =al + b . Consider the circuit shown in figure 1. This has three functional sections which compute by adding and subtracting currents and. where voltages are "log" coded; this is the essence of the "current-mode" aproach in circuit design (Andreou et.al. 1994). The first section, receives an input voltage from a presynaptic neuron, converts it into a current I(in), and multiplies it by a weight a. The second section adds and subtracts the bias current b. The last section converts the output current into an output voltage and transmits it to the next neuron in the network. Since the transistors have exponential transfer characteristics, this voltage is logarithmically coded. The parameters a and b are set by external voltages. Theyarameter a, is set by a single external voltage Va while the bias parameter b = br-) - b( + is set by two external voltages vb (+) and vb H . Two voltages are used for b to account for both positive and negative bIas values since b(-?0 and b( +?0 . '{,(+) r---------, r---- -, r-----I I I I I I I I I I I I I I I I I I I I I I I I (ou I (in) I I I I I I II I : I I I I I (out) I I ~ I I I~--~---+----~~r--;--~--~V I I I I I I I I I I I I I I I I I I I I I I I I L ______ -' .I I I l : aiin~ V L ________ ------ ~-) Figure 1. The analog neuron has three sections. To derive the dynamical equations of the neuron, it is neccesary to add up all the currents and invoke Kirchoffs current law, which requires that lout ) _al in ) +b(+) -b(-) = Ic . (3) If we now assume a simple subthreshold model for the behavior of the FET's and PFETs in the neuron, we can obtain the following expression for the current across the capacitor: Q dl out ) lout) dt --- =1 c (4) An Analog Neural Network Inspired by Fractal Block Coding 799 where Q = Cl1cVth determines the characteristic time scale of the neuron 2. It immediately follows from the last two expressions that the dynamics of a single neuron is determined by the equation Q dJCout) = _/(out) (I(out) _ al(in) _ b). (5) dt Where b = M-) - M+) . This equation appears to have a quadratic nonlinearity on the r.h.s. In fact, the noninearity is even more complicated since, the cooeficients a, M+) and b( -) are not constants, but depend on I(out) (through v(out). Application of the simple subthreshold model, results in a multiplier gain that is a function of v( out) (and hence tout)) as well as Va . It is given by a( va' v<out?) = 2exp(- v~ {sinh( v~ - va) -sinh( v~ _v(out? J (6) Similarly, the currents b(+) and b H are given by and b( +) = I~fpet) exp( KVb(+) )( 1- exp( _v(out?) (7.a) b(-) = I~nfet) exp(KVb (_) )(I_exp(_v(Out?) (7.b) == vdd - va . 3.2 Network dynamics and Stability considerations respectively, where va With these results we conclude that, a network of neurons, in which each neuron receives input from only one other neuron, would have a dynamical equation of the form d['! = -1?(/? - a? ( I? ) 1 ?(?) - b?) Q_ dt !! ! I J! I (8) where the connectivity of the network is determined by the function j( i) . The fixed points of these highly nonlinear equations occur when the r.h.s. of (8) vanishes. This can only happen if either Ii = Oor if (Ii - aJj(i) - bi ) = 0 for each i. The local stability of each of these fixed points follows by examining the eigenvalues (A.) of the corresponding jacobian. The expression for the jacobian at a general point I is J'!k i = dF dlk =-Q[(/.! - a.J.(?) - b?)8? I J I I Ik + [,(1a~1I J.(.)I I - b~)8'k I I a.J.8 '(')k] . I I J I (9) Where the partial derivatives, a'i and b'j are with respect to Ii. At a fixed point the jacobian takes the form J ik { bi 8ik = Q -/i [ (1- a[lj(i) - b[)8ik - ai8j(i)k] if if Ii =0 (Ii - ailj(i) - b i ) = O? (10) 2C represents the total gate capacitance from all the transistors connected to the horizontal line of the neuron. For the 2J..l analog proc~s~, the gate capacitance is aprroximately 0.5 fF/J..l2 so a 10J..l x 10J..l FET has a charactenstlc charge of Q =2.959 x 10- 4 Coulombs at room temperature. 800 Fernando Pineda, Andreas G. Andreou There are two cases of interest. The first case is when no neurons have zero output. This is the "desired solution." In this case, the jacobian specializes to J ik =-QIi [(1-aiIj(i) -b[)Oik -aiOj(i)k]' (11) Where, from (6) and (7), it can be shown that the partial derivatives, a'i and b'i are both non-positive. It immediately follows, from Gerschgorin's theorem, that a sufficient condition that the eigenvalues be negative and that the fixed point be stable, is that lail <l. The second case is when at least one of the neurons has zero output. We call these fixed points the "spurious solutions." In this case some of the eigenvalues are very easy to calculate because terms of the form (bi -).) ,where Ii = 0, can be factored from the expression for det(J-.?J). Thus some eigenvalues can be made positive by making some of the bi positive. Accordingly, if all the bi satisfy bi >0 , some of the eigenvalues will necessarily be positive and the spurious solutions will be unstable. To summarize the above discussion, we have shown that by choosing bi >0 and lail <1 for all i, we can make the desired fixed point stable and the spurious fixed points unstable. Note that a sufficient condition for bi >0 is if b~ +) =O. It remains to show that the system must converge to the desired fixed point, i.e. that the system cannot oscillate or wander chaotically. To do this we consider the connectivity of the network we implemented in our test chip. This is shown schematically in figure 2. The first eight neurons receive input from the odd numbered neurons while the second eight neurons receive input from the even numbered neurons. The neurons on the lefthand side all share the weight, WL, while the neurons on the right share the weight WR. By tracing the connections, we find that there are two independent loops of neurons: loop #1 = {0,8,12,14,IS,7,3,1} and loop #2 = {2,9,4,1O,13,6,1l,S}. Figure 2. The connection topology for the test chip is determined by the matrix of equation (1). The neurons are labeled 0-15. By inspecting each loop, we see that it passes through either the left or right hand range an even number of times. Hence, if there are any inhibitory weights in a loop, there must be an even number of them. This is the "even loop criterion", and it suffices to prove that the network is globally asymptotically stable, (Hirsch, 1987). 3.3. Programming the fixed point The nonlinear circuit of the previous section converges to a fixed point which is the solution of the following system of transcendental equations * -bi(-) (Ii* ,vb<-?-O Ii* -ai(li* ,va)Ij(i) (12) An Analog Neural Network Inspired by Fractal Block Coding 801 where the coefficients ai and bi are given by equations (6) and (7b) respectively. Similarly, the iterated affine transformations converge to the solution of the following linear equations * - Bj Ii* - A/j(j) =0 (13) where the coefficients {Ai ,Bi } and the connectionsj(i) are obtained by solving the approximate inverse problem with the additional constraints that bi >0 and lail <1 for all i,. The requirement that the fixed points of the two systems be identical results in the conditions Aj = aj(lj* ,va) ) Bj -- b(-)(I* j j ,Vb(-) (14) These equations can be solved for the required input voltages Va, and v b (-). Thus we are able to construct a nonlinear dynamical system that converges to the same fixed point as a linear system. For this programming method to work, of course, the subthreshold model we have used to characterize the network must accurately model the physical properties of the neural network. 4. PRELIMINARY RESULTS As a first step towards realizing a working system, we fabricated a Tiny chip containing 16 neurons arranged in two groups of eight. The topology is the same as shown in figure 2. The neurons are similar to those in figure 1 except that the bias term in each block of 8 neurons has the form b = kb( -) + (7 - k )b( -) , where O::;k::;7 is the label of a particular neuron within a block. This form increases the complexity of the neurons, but also allows us to represent ramps more easily (see figure 3). We fabricated the chip through MOSIS in a 2~m p-well CMOS process. A switching layer allows us to change the connection topology at run-time. One of the four possible configurations corresponds to the toplogy of figure 2. Six external voltages {Va ,V H ' Vi) H ' Va ,Vb H ' Vi)H }parameterize the fixed points of the network. These are confrolfM blpote~tioIdeters~ There is multiplexing circuitry included on the chip that selects which neuron output is to be amplified by a sense-amp and routed off-chip. The neurons can be addressed individually by a 4-bit neuron address. The addressing and analog-to-digital conversion is performed by a Motorolla 68HCIIAI microprocessor. We have operated the chip at 5volts and at 2.6 volts. Figure 3. shows the scanned steady state output of one of the test chips for a particular choice of input parameters with vdd =5 volts. The curve in figure 3. exhibits the qualitatively self-similar features of a recursively generated object. We are able to see three generations of a ramp. At 2.5 volts we see a very similar curve. We find that the chip draws 16.3 ~ at 2.5 volts. This corresponds to a steady state power dissipation of 411lW. Simulations indicate that the chip is operating in the subthreshold regime when Vdd = 2.5 volts. Simulations also indicate that the chip settles in less than one millisecond. We are unable to perform quantitiative measurements with the first chip because of several layout errors. On the other hand, we have experimentally verified that the network is indeed stable and that network produces qualitative fractals. We explored the parameter space informatlly. At no time did we encounter anything but the desired solutions. Fernando Pineda, Andreas G. Andreou 802 O~--~~~------~--==:L----~----~----~- o 2 4 6 8 Neuron label 10 12 14 Figure 3 D/A output for chip #3 for a particular set of input voltages. We have already fabricated a larger design without the layout problems of the prototype. This second design has 32 pixeles and a richer set of permitted topologies. We expect to make quantitative measurements with this second design. In particular we hope to use it to decompress an actual block code. Acknowledgements The work described here is funded by APL !R&D as well as a grant from the National Science Foundation ECS9313934, Paul Werbos is the monitor. The authors would like to thank Robert Jenkins, Kim Strohbehn and Paul Furth for many useful conversations and suggestions. References Andreou, A.G. and Boahen, K.A. Neural Information Processing I: The Current-Mode approach, Analog VLSI: Signal and Information Processing, (eds: M Ismail and T. Fiez) MacGraw-Hill Inc., New York. Chapter 6 (1994). Hille, B., Ionic Channels of Excitable Membranes, Sunderland, MA, Sinauer Associates Inc. (1984). Hirsch, M. ,Convergence in Neural Nets, Proceedings of the IEEE ICNN, San Diego, CA, (1987). Jacquin, A. E., A Fractal Theory of iterated Markov operators with applications to digital image coding, Ph.D. Dissertation, Georgia Institute of Technology (1989). Mead, c., Analog VLSI and Neural Systems, Addison Wesley, (1989) Monroe, D.M. and Dudbridge, F. Fractal block coding of images, Electronics Letters, 28, pp. 1053-1055, (1992). Pineda, F.J., Dynamics and Architecuture for Neural Computation, Journal of Complexity, 4, 216-245 (1988).
957 |@word briefly:1 compression:1 simulation:2 ajj:1 recursively:1 electronics:1 configuration:1 initial:1 amp:1 current:10 must:4 john:3 transcendental:1 happen:1 shape:1 half:4 device:1 accordingly:1 realizing:1 dissertation:1 simpler:1 ik:5 qualitative:1 consists:1 prove:1 indeed:1 behavior:1 inspired:4 globally:1 actual:3 provided:1 suffice:1 circuit:6 medium:1 finding:1 transformation:16 fabricated:5 quantitative:1 charge:1 nutshell:1 unit:2 grant:1 pfe:1 positive:5 engineering:2 local:1 switching:1 encoding:2 id:1 mead:2 qii:1 bi:12 dlk:1 range:1 unique:3 block:15 implement:1 procedure:1 road:1 jacquin:2 numbered:2 oor:1 cannot:1 operator:1 context:2 applying:1 conventional:1 layout:2 immediately:2 factored:1 amax:1 stability:2 n12:1 transmit:7 diego:1 programming:2 decompress:1 associate:1 element:3 logarithmically:1 expensive:1 werbos:1 labeled:1 bottom:3 electrical:1 solved:1 parameterize:1 calculate:1 connected:1 boahen:2 vanishes:1 complexity:3 dynamic:7 vdd:3 depend:4 solving:3 chaotically:1 decompressing:1 completely:1 easily:1 chip:16 chapter:1 describe:3 choosing:2 whose:3 richer:1 larger:1 solve:2 say:1 ramp:2 final:3 pineda:7 sequence:4 eigenvalue:7 transistor:2 net:2 subtracting:1 loop:6 lefthand:1 amplified:1 tout:1 description:1 ismail:1 convergence:1 requirement:1 produce:3 aiin:1 cmos:5 converges:6 object:1 derive:1 recurrent:2 linearize:2 ij:1 odd:2 kirchoffs:1 implemented:1 implies:1 indicate:2 differ:1 kb:1 settle:1 implementing:1 behaviour:1 suffices:1 icnn:1 preliminary:1 inspecting:1 mm:1 considered:1 ic:1 exp:10 bj:2 circuitry:1 uniqueness:1 proc:1 label:2 individually:1 wl:5 create:1 hope:1 voltage:13 laurel:1 kim:1 sense:1 hille:2 entire:1 lj:2 hidden:1 vlsi:4 spurious:3 sunderland:1 transformed:3 selects:1 multiplies:1 special:3 construct:2 identical:1 represents:2 uncompressed:1 composed:1 national:1 subsampled:2 conductance:1 interest:1 highly:4 investigate:1 operated:1 behind:1 partial:3 desired:5 lwe:1 addressing:1 examining:1 too:1 characterize:1 st:1 explores:1 physic:2 invoke:1 decoding:1 off:1 hopkins:2 concrete:1 connectivity:2 containing:1 choose:1 external:4 derivative:2 toy:1 li:1 account:1 coding:10 coefficient:2 inc:2 satisfy:1 vi:2 performed:2 analyze:1 complicated:1 square:2 characteristic:3 subthreshold:8 yield:1 iterated:4 accurately:1 ionic:1 synapsis:1 ed:1 pp:1 proof:1 transmits:1 gain:3 sampled:1 conversation:1 lail:3 ou:1 actually:1 appears:1 wesley:1 disposal:1 tum:1 dt:3 permitted:1 specify:1 arranged:1 just:2 sketch:1 receives:2 horizontal:1 hand:2 working:1 nonlinear:11 mode:2 aj:2 scientific:1 concept:1 true:1 multiplier:1 hence:2 volt:6 laboratory:1 iteratively:1 self:3 essence:1 steady:4 anything:1 coincides:1 fet:2 q_:1 criterion:1 hill:1 dissipation:1 temperature:1 image:2 consideration:1 charles:1 functional:1 physical:5 analog:18 measurement:2 pfets:2 blocked:1 ai:3 similarly:3 decompression:3 nonlinearity:2 funded:1 furth:1 stable:8 operating:1 kvb:2 add:2 perspective:1 additional:1 converge:3 fernando:5 signal:1 ii:10 coded:3 controlled:1 calculates:1 va:11 basic:1 essentially:1 df:1 physically:3 iteration:2 sometimes:1 represent:2 ion:1 receive:2 schematically:1 baltimore:1 addressed:1 pass:1 capacitor:1 call:1 easy:1 topology:4 andreas:5 idea:2 prototype:1 br:5 det:1 expression:5 six:1 effort:1 routed:1 york:1 cause:1 constitute:1 oscillate:1 fractal:16 useful:3 governs:1 amount:3 ph:1 simplest:1 generate:2 inhibitory:1 millisecond:1 wr:6 group:1 four:3 salient:1 monitor:1 verified:1 asymptotically:1 relaxation:2 mosis:1 convert:2 wand:4 run:1 inverse:2 parameterized:2 letter:1 reader:1 draw:1 vb:5 bit:3 layer:1 followed:1 sinh:2 copied:2 quadratic:1 occur:1 scanned:1 constraint:1 multiplexing:1 aspect:1 aproach:1 pfet:1 membrane:1 across:1 wi:1 making:1 refered:1 taken:1 equation:12 remains:1 needed:4 addison:1 operation:2 jenkins:1 eight:3 worthwhile:1 coulomb:1 encounter:1 gate:2 top:3 subsampling:1 transconductances:1 bl:2 capacitance:2 already:1 md:2 exhibit:2 unable:1 mapped:2 thank:1 presynaptic:1 unstable:2 trivial:1 code:4 nfet:2 robert:1 negative:2 rise:1 implementation:1 design:4 perform:2 conversion:1 neuron:37 observation:1 markov:1 thermal:1 intensity:2 required:1 connection:3 andreou:8 address:1 able:2 dynamical:5 regime:2 lout:2 summarize:2 program:2 max:1 power:1 natural:1 force:1 technology:2 fiez:1 specializes:1 excitable:1 l2:1 acknowledgement:1 sinauer:1 nxn:1 law:1 wander:1 expect:1 apl:1 generation:1 suggestion:1 digital:2 foundation:1 affine:7 sufficient:2 tiny:1 share:2 land:1 course:1 last:2 copy:2 free:1 bias:4 side:1 institute:1 sparse:3 tracing:1 curve:5 author:1 qualitatively:2 made:1 san:1 subtracts:1 approximate:1 hirsch:2 active:2 receiver:2 conclude:1 channel:2 transfer:3 ca:1 necessarily:1 microprocessor:1 did:1 paul:2 ff:1 georgia:1 position:1 decoded:1 exponential:3 lw:1 jacobian:4 theorem:1 specific:1 explored:1 dl:1 adding:1 monroe:2 strohbehn:1 simply:1 vth:1 expressed:1 vsb:3 corresponds:2 satisfies:1 determines:1 ma:1 consequently:1 towards:1 room:1 change:1 experimentally:1 included:1 determined:4 except:1 operates:1 total:3 experimental:2 latter:1 arises:1 xl0:1 dept:1
7,197
958
A Lagrangian Formulation For Optical Backpropagation Training In Kerr-Type Optical Networks James E. Steck Mechanical Engineering Wichita State University Wichita, KS 67260-0035 Steven R. Skinner Electrical Engineering Wichita State University Wichita, KS 67260-0044 Alvaro A. Cruz-Cabrara Electrical Engineering Wichita State University Wichita, KS 67260-0044 Elizabeth C. Behrman Physics Department Wichita State University Wichita, KS 67260-0032 Abstract A training method based on a form of continuous spatially distributed optical error back-propagation is presented for an all optical network composed of nondiscrete neurons and weighted interconnections. The all optical network is feed-forward and is composed of thin layers of a Kerrtype self focusing/defocusing nonlinear optical material. The training method is derived from a Lagrangian formulation of the constrained minimization of the network error at the output. This leads to a formulation that describes training as a calculation of the distributed error of the optical signal at the output which is then reflected back through the device to assign a spatially distributed error to the internal layers. This error is then used to modify the internal weighting values. Results from several computer simulations of the training are presented, and a simple optical table demonstration of the network is discussed. Elizabeth C. Behrman 772 1 KERR TYPE MATERIALS Kerr-type optical networks utilize thin layers of Kerr-type nonlinear materials, in which the index of refraction can vary within the material and depends on the amount of light striking the material at a given location. The material index of refraction can be described by: n(x)=no+nzI(x), where 110 is the linear index of refraction, ~ is the nonlinear coefficient, and I(x) is the irradiance of a applied optical field as a function of position x across the material layer (Armstrong, 1962). This means that a beam of light (a signal beam carrying information perhaps) passing through a layer of Kerr-type material can be steered or controlled by another beam of light which applies a spatially varying pattern of intensity onto the Kerr-type material. Steering of light with a glass lens (having constance index of refraction) is done by varying the thickness of the lens (the amount of material present) as a function of position. Thus the Kerr effect can be loosely thought of as a glass lens whose geometry and therefore focusing ability could be dynamically controlled as a function of position across the lens. Steering in the Kerr material is accomplished by a gradient or change in the material index of refraction which is created by a gradient in applied light intensity. This is illustrated by the simple experiment in Figure 1 where a small weak probe beam is steered away from a straight path by the intensity gradient of a more powerful pump beam. lex) Pump I~ > /-.. . x Figure 1: Light Steering In Kerr Materials 2 OPTICAL NETWORKS USING KERR MATERIALS The Kerr optical network, shown in Figure 2, is made up of thin layers of the Kerr- type nonlinear medium separated by thick layers of a linear medium (free space) (Skinner, 1995). The signal beam to be processed propagates optically in a direction z perpendicular to the layers, from an input layer through several alternating linear and nonlinear layers to an output layer. The Kerr material layers perform the nonlinear processing and the linear layers serve as connection layers. The input (l(x)) and the weights (W\(x),W2(x) ... Wn(x)) are irradiance fields applied to the Kerr type layers, as functions of lateral position x, thus varying the A Lagrangian Formulation for Optical Backpropasation 773 refractive index profile of the nonlinear medium. Basically, the applied weight irradiences steer the signal beam via the Kerr effect discussed above to produce the correct output. The advantage of this type of optical network is that both neuron processing and weighted connections are achieved by uniform layers of the Kerr material. The all optical nature eliminates the need to physically construct neurons and connections on an individual basis. O(x,y) ? Plane Wave (Eo) Figure 2: Kerr Optical Neural Network Architecture If E;(ex) is the light entering the itlt nonlinear layer at lateral position ex, then the effect of the nonlinear layer is given by (1) where W;( ex) is the applied weight field. Transmission of light at lateral location ex at the beginning of the itlt linear layer to location p just before the i+ 1tit nonlinear layer is given by where c 3 ko =-- I 2!:lLI (2) OPTICAL BACK-PROPAGATION TRAINING Traditional feed-forward artificial neural networks composed of a finite number of discrete neurons and weighted connections can be trained by many techniques. Some of the most successful techniques are based upon the well known training method called backpropagation which results from minimizing the network output error, with respect to the network weights by a gradient descent algorithm. The optical network is trained using a form of continuous optical back-propagation which is developed for a nondiscrete network. Gradient descent is applied to minimize the error over the entire output region of the optical network. This error is a continuous distribution of error calculated over the output region. 774 Elizabeth C. Behrman Optical back-propagation is a specific technique by which this error distribution is optically propagated backward through the linear and nonlinear optical layers to produce error signals by which the light applied to the nonlinear layers is modified. Recall that this applied light Wi controls what serves as connection "weights" in the optical network. Optical backpropagation minimizes the error Lo over an output region 0 0 > a subdomain of the fmal or nth layer of the network, ~ where = 'Y r O(u'fJ '(uliu )c o (3) subject to the constraint that the propagated light, Ei( ex), satisfies the equations of forward propagation (1) and (2). O(P) = En+I(P) and is the network output, y is a scaling factor on the output intensity. Lo then is the squared error between the desired output value D and the average intensity 10 of the output distribution O( P). This constrained minimization problem is posed in a Lagrange formulation similar to the work of (Ie Cunn, 1988) for conventional feedforward networks and (pineda, 1987) for conventional recurrent networks; the difference being that for the optical network of this paper the Electric field E and the Lagrange multiplier are complex and also continuous in the spatial variable thus requiring the Lagrangian below. A Lagrangian is defmed as; L = 4, + + :t fA; t u ) [ EI+I(U) - fFI~)~ Ie -jctP ... )z 0. 0. JA/+~U) [Ei+~U) it. - 0. - fF~~)~/e-jC~13-?)Z dP ] ax (4) dP r ax 0. Taking the variation ofL with respect to E i, the Lagrange multipliers Ai, and using gradient descent to minimize L with respect to the applied weight fields Wi gives a set of equations that amount to calculating the error at the output and propagating the error optically backwards through the network. The pertinent results are given below. The distributed assigrunent of error on the output field is calculated by A1f+1(R.) = ... ~0 'Y ' (R.) [ D - 10 ] ... (5) This error is then propagated back through the nth or final linear optical layer by the equation ?c 6 (~) = ~ " 1t r z A + (u) e -jC,/..13-u) dx ) Co " 1 (6) which is used to update the "weight" light applied to the nth nonlinear layer. Optical backpropagation, through the ith nonlinear layer (giving AlP? followed by the linear layer (giving ~i-I(P? is performed according to the equations A Lagrangian Formulation for Optical Backpropagation 775 (7) This gives the error signal ~j'I(P) used to update the "weight" light distribution Wj.I(P) applied to the i-I nonlinear layer. The "weights" are updated based upon these errors according to the gradient descent rule Wi-(~) = w/,/d(P) +l'lt~)ktPNLin2W/"t~) 2 IM[ ~(~) 6,(~) e-~ANL.nZ<lw,ClId(p)f.IE.(I\)I'>] (8) where ,,;CP) is a learning rate which can be, but usually is not a function of layer number i and spatial position p. Figure 3 shows the optical network (thick linear layers and thin nonlinear layers) with the unifonn plane wave Eo, the input signal distribution I, forward propagation signals EI E2 .. . En' the weighting light distributions at the nonlinear layers WI W 2 .. , W n. Also shown are the error signal An+1 at the output and the back-propagated error signals ~n ... ~2 ~I for updating the nonlinear layers. Common nonlinear materials exist for which the material constants are such that the second term in the first of Equations 7 becomes small. Ignoring this second term gives an approximate fonn of optical backpropagation which amounts to calculating the error at the output of the network and then reversing its direction to optically propagate this error backward through the device. This can be easily seen by comparing Equations 6 and 7 (with the second tenn dropped) for optical back-propagation of the output error An with Equations I and 2 for the forward propagation of the signal E j. This means that the optical back-propagation training calculations potentially can be implemented in the same physical device as the forward network calculations. Equation (8) then becomes; Wi-(~) + = Wio/d(~) (2t'l,(~'>kot:.NL"2) w//d(~) [ (Et~) At~? - ~t~) ~(~)r] (9) which may be able to be implemented optically. 4 SIMULATION RESULTS To prove feasibility, the network was then trained and tested on several benchmark classification problems, two of which are discussed here. More details on these and other simulations of the optical network can be found in (Skinner, 1995). In the first (Using Nworks, 1991), iris species were classified into one of three categories: Setosa, Versicolor or Virginica. Classification was based upon length and width of the sepals and Elizabeth C. Behrman 776 petals. The network consisted of an input self-defocusing layer with an applied irradiance field which was divided into 4 separate Gaussian distributed input regions 25 microns in width followed by a linear layer. This pattern is repeated for 4 more groups composed of a nonlinear layer (with applied weights) followed by a linear layer. The final linear layer has three separate output regions 10 microns wide for binary classification as to species. The nonlinear layers were all 20 microns thick with n2=-.05 and the linear layers were 100 microns thick. The wavelength of applied light was 1 micron and the width of the network was 512 microns discretized into 512 pixels. This network was trained on a set of 50 training pairs to produce correct classification of all 50 training pairs. The network was then used to classify 50 additional pairs of test data which were not used in the training phase. The network classified 46 of these correctly for a 92 % accuracy level which is comparable to a standard feedforward network with discrete output region sigmoidal neurons. -\ \.. In the second problem, we tested the performance of the network on a set of data from a dolphin sonar discrimination experiment (Roitblat, 1991). In this study a dolphin was presented with one of three of different types objects (a tube, a sphere, and a cone), allowed to echolocate, and rewarded for choosing the correct one from a comparison array. The Fourier transforms of his click echoes, in the form of average amplitudes in each of 30 frequency bins, were then used as inputs for a neural network. Nine nonlinear layers were used along with 30 input regions and 3 \1l'Y). ? ? output plane tttttttttttT ALn ttl O(x,y) ?n(x,y) ? ? ? ? ? ? ? ? ? I .ili'Ln Wn(x,y) En(x,y) t t t t. t t t. t t t t ? 01 (X,y) ? ? ? ? ??? ? t?. t t t W1 (x,y) El(X,y) ttttttttttt I .ili'L1 T ALo ...----_ _----,1 ~ ~1 ____________I(_X_,y_)____________ Eo .ili'Lo ttttttttttt Plane Wave Figure 3: Optical Network Forward Data and Backward Error DataFlow A Lagrangian Formulatio1l for Optical Backpropagation 777 output regions, the remainder of the network physical parameters were the same as above for the iris classification. Half the data (13 sets of clicks) was used to train the network, with the other half of the data (14 sets) used to test the training. After training, classification of the test data set was 100% correct. 5 EXPERIMENTAL RESULTS As a proof of the concept, the optical neural network was constructed in the laboratory to be trained to perform various logic functions. Two thermal self-defocusing layers were used, one for the input and the other for a single layer of weighting. The nonlinear coefficient of the index of refraction (nJ was measured to be -3xlO'" cm21W. The nonlinear layers had a thickness (~NLo and ~NL.) of 630llm and were separated by a distance (~Lo) of 15cm. The output region was 100llm wide and placed 15cm (~L.) behind the weighting layer. The experiment used HeNe laser light to provide the input plane wave and the input and weighting irradiances. The spatial profiles of the input and weighting layers were realized by imaging a LCD spatial light modulator onto the respective nonlinear layers. The inputs were two bright or dark regions on a Gaussian input beam producing the intensity profile: whereIo= 12.5 mW/cm2, leo = 900llm, Xo = 600Ilffi, K. = 400Ilffi, and Qo and Q. are the logic inputs taking on a value of zero or one. The weight profile W.(x) = Ioexp[-(xIKo)2][1 +w.(x)] where w.(x) can range from zero to one and is found through training using an algorithm which probed the weighting mask in order update the training weights. Table 1 shows the experimental results for three different logic gates. Given is the normalized output before and after training. The network was trained to recognize a logic zero for a normalized output ~ 0.9 and a logic one or a normalized output ~ 1.1. An output value greater than I is considered a logic one and an output value less than one is a logic zero. RME is the root mean error. 6 CONCLUSIONS Work is in progress to improve the logic gate results by increasing the power of propagating signal beam as well as both the input and weighting beams. This will effectively increase the nonlinear processing capability of the network since a higher power produces more nonlinear effect. Also, more power will allow expansion of all of the beams thereby increasing the effective resolution of the thermal materials. Thisreduces the effect of heat transfer within the material which tends to wash out or diffuse benificial steep gradients in temperature which are what produce the gradients in the index of refraction. In addition, the use of photorefractive crystals for optical weight storage shows promise for being able to optically phase conjugate and backpropagate the output errror as well as implement the weight update rule for all optical network training. This appears to be simpler than optical networks using volume hologram weight storage because the Kerr network requires only planar hologram storage. 778 Elizabeth C. Behrman Inputs Start Finish NOR Change Output AND Start Finish Change Output Start XNOR Finish Change Output 0 0 1.001 1.110 .109 0 1 .802 .884 .998 .082 0 1.092 .757 -.241 .855 -.237 0 0 .998 1.084 .086 .880 .933 1 1 .053 0 1 0 .698 1 1 Rl\1E .807 .896 7.3% .089 0 1.440 -7.3% 1.124 0 -.254 0 .893 -.316 -16.4% 1 .994 7.3% .928 .035 1.073 .079 2.7% -4.6% 0 1 .772 .074 0 1.148 .894 0 16.4% Table 1: Preliminary Experimental Logic Gate Results References Armstrong, J.A., Bloembergen, N., Ducuing, J., and Pershan, P.S., (1962) "Interactions Between Light Waves in a Nonlinear Dielectric", Physical Review, Vol. 127, pp. 1918-1939. Ie Cun, Yann, (1988) itA Theoretical Framework for Back-Propagation", Proceedings of the 1988 Connectionist Models Summer School, Morgan Kaufmann, pp. 21-28. Pineda, F.J., (1987) "Generalization of backpropagation to recurrent and higher order neural networks", Proceedings of IEEE Conference on Neural information Processing Systems, November 1987, IEEE Press. Roitblat., Moore, Nachtigall, and Penner, (1991) "Natural dolphin echo recognition using an integrator gateway network," in Advances in Neural Processing Systems 3 Morgan Kaufmann, San Mateo, CA, 273-281. Skinner, S.R, Steck, J.E., Behnnan, E .C., (1995) "An Optical Neural Network Using Kerr Type Nonlinear Materials", To Appear in Applied Optics. "Using Nworks, (1991) An Extended Tutorial for NeuralWorks Professional /lIPlus and NeuralWorks Explorer, NeuralWare, Inc. Pittsburgh, PA, pg. UN-18.
958 |@word cm2:1 steck:2 simulation:3 propagate:1 pg:1 fonn:1 thereby:1 optically:6 comparing:1 dx:1 cruz:1 pertinent:1 update:4 discrimination:1 tenn:1 half:2 device:3 plane:5 beginning:1 ith:1 location:3 sigmoidal:1 simpler:1 along:1 constructed:1 prove:1 mask:1 nor:1 integrator:1 discretized:1 increasing:2 becomes:2 medium:3 what:2 ttl:1 cm:2 aln:1 minimizes:1 developed:1 nj:1 control:1 appear:1 producing:1 before:2 engineering:3 dropped:1 modify:1 tends:1 path:1 nz:1 k:4 mateo:1 dynamically:1 co:1 perpendicular:1 range:1 itlt:2 implement:1 backpropagation:8 thought:1 onto:2 storage:3 conventional:2 lagrangian:7 resolution:1 rule:2 array:1 his:1 variation:1 updated:1 pa:1 recognition:1 updating:1 steven:1 electrical:2 region:10 wj:1 lcd:1 trained:6 carrying:1 tit:1 serve:1 upon:3 basis:1 easily:1 various:1 leo:1 train:1 separated:2 laser:1 heat:1 effective:1 artificial:1 choosing:1 whose:1 dielectric:1 posed:1 interconnection:1 ability:1 echo:2 final:2 pineda:2 advantage:1 interaction:1 refractive:1 remainder:1 dolphin:3 transmission:1 produce:5 object:1 recurrent:2 propagating:2 measured:1 school:1 progress:1 implemented:2 direction:2 thick:4 correct:4 alp:1 material:21 bin:1 ja:1 assign:1 generalization:1 preliminary:1 im:1 considered:1 vary:1 constance:1 weighted:3 minimization:2 gaussian:2 modified:1 varying:3 derived:1 ax:2 bloembergen:1 glass:2 el:1 entire:1 pixel:1 classification:6 constrained:2 spatial:4 field:7 construct:1 having:1 thin:4 connectionist:1 x_:1 composed:4 recognize:1 individual:1 geometry:1 phase:2 nl:2 light:18 behind:1 respective:1 loosely:1 desired:1 theoretical:1 classify:1 steer:1 penner:1 pump:2 uniform:1 successful:1 thickness:2 alvaro:1 ie:4 physic:1 w1:1 squared:1 tube:1 nachtigall:1 alo:1 steered:2 coefficient:2 inc:1 jc:2 depends:1 performed:1 root:1 wave:5 start:3 capability:1 minimize:2 bright:1 accuracy:1 kaufmann:2 weak:1 roitblat:2 basically:1 lli:1 straight:1 classified:2 frequency:1 pp:2 james:1 refraction:7 e2:1 proof:1 propagated:4 neuralworks:2 dataflow:1 recall:1 amplitude:1 irradiance:3 back:10 focusing:2 feed:2 appears:1 higher:2 reflected:1 planar:1 formulation:6 done:1 fmal:1 just:1 ei:4 qo:1 nonlinear:29 propagation:10 perhaps:1 effect:5 requiring:1 multiplier:2 consisted:1 concept:1 normalized:3 skinner:4 spatially:3 alternating:1 entering:1 laboratory:1 moore:1 illustrated:1 xnor:1 self:3 defmed:1 width:3 iris:2 crystal:1 cp:1 l1:1 fj:1 temperature:1 subdomain:1 common:1 physical:3 rl:1 volume:1 discussed:3 ai:1 had:1 gateway:1 virginica:1 rewarded:1 binary:1 accomplished:1 seen:1 morgan:2 additional:1 greater:1 steering:3 hologram:2 eo:3 signal:12 calculation:3 nlo:1 sphere:1 y_:1 divided:1 controlled:2 feasibility:1 ko:1 physically:1 ofl:1 achieved:1 beam:11 addition:1 w2:1 eliminates:1 subject:1 petal:1 mw:1 backwards:1 feedforward:2 wn:2 finish:3 architecture:1 modulator:1 click:2 passing:1 nine:1 sepal:1 amount:4 transforms:1 dark:1 processed:1 category:1 behrman:5 exist:1 unifonn:1 tutorial:1 nzi:1 correctly:1 discrete:2 probed:1 promise:1 vol:1 group:1 utilize:1 backward:3 imaging:1 cone:1 micron:6 powerful:1 striking:1 yann:1 scaling:1 comparable:1 layer:47 followed:3 summer:1 optic:1 constraint:1 diffuse:1 rme:1 fourier:1 optical:40 department:1 according:2 conjugate:1 describes:1 across:2 elizabeth:5 wi:5 cun:1 defocusing:3 xo:1 ln:1 equation:8 ffi:1 kerr:19 serf:1 probe:1 away:1 professional:1 gate:3 calculating:2 giving:2 lex:1 realized:1 fa:1 traditional:1 gradient:9 dp:2 distance:1 separate:2 lateral:3 length:1 index:8 demonstration:1 minimizing:1 steep:1 potentially:1 perform:2 neuron:5 benchmark:1 finite:1 descent:4 november:1 thermal:2 extended:1 intensity:6 pair:3 mechanical:1 connection:5 able:2 below:2 pattern:2 usually:1 kot:1 power:3 natural:1 explorer:1 nth:3 improve:1 created:1 review:1 ita:1 propagates:1 lo:4 placed:1 free:1 allow:1 wide:2 taking:2 distributed:5 calculated:2 versicolor:1 forward:7 made:1 san:1 approximate:1 logic:9 pittsburgh:1 llm:3 anl:1 continuous:4 un:1 sonar:1 table:3 nature:1 transfer:1 ca:1 ignoring:1 expansion:1 complex:1 electric:1 profile:4 n2:1 repeated:1 allowed:1 en:3 ff:1 position:6 lw:1 weighting:8 specific:1 effectively:1 setosa:1 wash:1 backpropagate:1 lt:1 wavelength:1 lagrange:3 xlo:1 applies:1 satisfies:1 change:4 reversing:1 photorefractive:1 lens:4 called:1 specie:2 ili:3 wio:1 experimental:3 internal:2 a1f:1 armstrong:2 tested:2 ex:5
7,198
959
Learning Many Related Tasks at the Same Time With Backpropagation Rich Caruana School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 caruana@cs.cmu.edu Abstract Hinton [6] proposed that generalization in artificial neural nets should improve if nets learn to represent the domain's underlying regularities . Abu-Mustafa's hints work [1] shows that the outputs of a backprop net can be used as inputs through which domainspecific information can be given to the net . We extend these ideas by showing that a backprop net learning many related tasks at the same time can use these tasks as inductive bias for each other and thus learn better. We identify five mechanisms by which multitask backprop improves generalization and give empirical evidence that multi task backprop generalizes better in real domains. 1 INTRODUCTION You and I rarely learn things one at a time, yet we often ask our programs to-it must be easier to learn things one at a time than to learn many things at once. Maybe not. The things you and I learn are related in many ways . They are processed by the same sensory apparatus, controlled by the same physical laws, derived from the same culture, ... Perhaps it is the similarity between the things we learn that helps us learn so well. What happens when a net learns many related functions at the same time? Will the extra information in the teaching signal of the related tasks help it learn better? Section 2 describes five mechanisms that improve generalization in backprop nets trained simultaneously on related tasks. Section 3 presents empirical results from a road-following domain and an object-recognition domain where backprop with multiple tasks improves generalization 10-40%. Section 4 briefly discusses when and how to use multitask backprop . Section 5 cites related work and Section 6 outlines directions for future work. 658 2 Rich Caruana MECHANISMS OF MULTITASK BACKPROP We identified five mechanisms that improve generalization in backprop nets trained simultaneously on multiple related tasks. The mechanisms all derive from the summing of error gradient terms at the hidden layer from the different tasks. Each exploits a different relationship between the tasks. 2.1 Data Amplification Data amplification is an effective increase in sample size due to extra information in the training signal of related tasks. There are two types of data amplification. 2.1.1 Statistical Data Amplification Statistical amplification, occurs when there is noise in the training signals. Consider two tasks, T and T', with independent noise added to their training signals, that both benefit from computing a feature F of the inputs. A net learning both T and T' can, if it recognizes that the two tasks share F, use the two training signals to learn F better by averaging F through the noise. The simplest case is when T T', i.e., when the two outputs are independently corrupted versions of the same signal. = 2.1.2 Blocking Data Amplification The 2nd form of data amplification occurs even if there is no noise. Consider two tasks, T and T', that use a common feature F computable from the inputs, but each uses F for different training patterns. A simple example is T = A OR F and T' = NOT(A) OR F. T uses F when A = 0 and provides no information about F when A = 1. Conversely, T' provides information about F only when A = 1. A net learning just T gets information about F only on training patterns for which A 0, but is blocked when A = 1. But a net learning both T and T' at the same time gets information about F on every training pattern; it is never blocked. It does not see more training patterns, it gets more information for each pattern. If the net learning both tasks recognizes the tasks share F, it will see a larger sample of F. Experiments with blocked functions like T and T' (where F is a hard but learnable function of the inputs such as parity) indicate backprop does learn common subfeatures better due to the larger effective sample size. = 2.2 Attribute Selection Consider two tasks, T and T', that use a common subfeature F. Suppose there are many inputs to the net, but F is a function of only a few of the inputs. A net learning T will,_if there is limited training data and/or significant noise, have difficulty distinguishing inputs relevant to F from those irrelevant to it. A net learning both T and T', however, will better select the attributes relevant to F because data amplification provides better training signals for F and that allows it to better determine which inputs to use to compute F. (Note: data amplification occurs even when there is no attribute selection problem. Attribute selection is a consequence of data amplification that makes data amplification work better when a selection problem exists.) We detect attribute selection by looking for connections to relevant inputs that grow stronger compared to connections for irrelevant inputs when multiple tasks are trained on the net. Learning Many Related Tasks at the Same Time with Backpropagation 2.3 659 Eavesdropping Consider a feature F, useful to tasks, T and T', that is easy to learn when learning T, but difficult to learn when learning T' because T' uses F in a more complex way. A net learning T will learn F, but a net learning just T' may not. If the net learning T' also learns T, T' can eavesdrop on the hidden layer learned for T (e.g., F) and thus learn better. Moreover, once the connection is made between T' and the evolving representation for F, the extra information from T' about F will help the net learn F better via the other mechanisms. The simplest case of eavesdropping is when T = F. Abu-Mostafa calls these catalytic hints[l]. In this case the net is being told explicitly to learn a feature F that is useful to the main task. Eavesdropping sometimes causes non-monotonic generalization curves for the tasks that eavesdrop on other tasks. This happens when the eavesdropper begins to overtrain, but then finds something useful learned by another task, and begins to perform better as it starts using this new information. 2.4 Representation Bias Because nets are initialized with random weights, backprop is a stochastic search procedure; multiple runs rarely yield identical nets. Consider the set of all nets (for fixed architecture) learnable by backprop for task T. Some of these generalize better than others because they better "represent" the domain's regularities. Consider one such regularity, F, learned differently by the different nets. Now consider the set of all nets learnable by backprop for another task T' that also learns regularity F. If T and T' are both trained on one net and the net recognizes the tasks share F, search will be biased towards representations of F near the intersection of what would be learned for T or T' alone. We conjecture that representations of F near this intersection often better capture the true regularity of F because they satisfy more than one task from the domain. Representations of F Findable by Backprop A form of representation bias that is easier to experiment with occurs when the representations for F sampled by the two tasks are different minima. Suppose there are two minima, A and B, a net can find for task T. Suppose a net learning task T' also has two minima, A and C. Both share the minima at A (i.e., both would perform well if the net entered that region of weight space), but do not overlap at Band C. We ran two experiments. In the first, we selected the minima so that nets trained on T alone are equally likely to find A or B, and nets trained on T' alone are equally likely to find A or C. Nets trained on both T and T' usually fall into A for both tasks. 1 Tasks prefer hidden layer representations that other tasks prefer. In the second experiment we selected the minima so that T has a strong preference lIn these experiments the nets have sufficient capacity to find independent minima for the tasks. They are not forced to share the hidden layer representations. But because the initial weights are random, they do initially share the hidden layer and will separate the tasks (i.e., use independent chunks of the hidden layer for each task) only if learning causes them to. 660 Rich Caruana for B over A: a net trained on T always falls into B. T', however, still has no preference between A or C. When both T and T' are trained on one net, T falls into B as expected: the bias from T' is unable to pull it to A. Surprisingly, T' usually falls into C, the minima it does not share with T! T creates a "tide" in the hidden layer representation towards B that flows away from A. T' has no preference for A or C, but is subject to the tide created by T. Thus T' usually falls into C; it would have to fight the tide from T to fall into A. Tasks prefer NOT to use hidden layer representations that other tasks prefer NOT to use. 2.5 How the Mechanisms are Related The "tide" mentioned while discussing representation bias results from the aggregation of error gradients from multiple tasks at the hidden layer. It is what makes the five mechanisms tick. It biases the search trajectory towards better performing regions of weight space. Because the mechanisms arise from the same underlying cause, it easy for them to act in concert. Their combined effect can be substantial. Although the mechanisms all derive from gradient summing, they are not the same. Each emphasizes a different relationship between tasks and has different effects on what is learned. Changes in architecture, representation, and the learning procedure affect the mechanisms in different ways. One particularly noteworthy difference between the mechanisms is that if there are minima, representation bias affects learning even with infinite sample size. The other mechanisms work only with finite sample size: data amplification (and thus attribute selection) and eavesdropping are beneficial only when the sample size is too small for the training signal for one task to provide enough information to the net for it to learn good models. 3 EMPIRICAL RESULTS Experiments on carefully crafted test problems verify that each of the mechanisms can work. 2 These experiments, however, do not indicate how effective multitask backprop is on real problems: tweaking the test problems alters the size of the effects. Rather than present results for contrived problems, we present a more convincing demonstration of multi task backprop by testing it on two realistic domains. 3.1 1D-ALVINN ID-ALVINN uses a road image simulator developed by Pomerleau. It was modified to generate I-D road images comprised of a single 32-pixel horizontal scan line instead of the original 2-D 30x32-pixel image. This was done to speed learning to allow thorough experimentation. ID-ALVINN retains much of the complexity of the original 2-D domain-the complexity lost is road curvature and that due to the smaller input (960 pixels vs. 32 pixels). The principle task in ID-ALVINN is to predict steering direction. Eight additional tasks were used for multitask backprop: ? ? ? ? whether the road is one or two lanes location of left edge of road location of road center intensity of region bordering road ? ? ? ? location of centerline (2-lane roads only) location of right edge of road intensity of road surface intensity of centerline (2-lane roads only) 2We have yet to determine how to directly test the hypothesis that representations for F in the intersection of T and T' perform better. Testing this requires interpreting representations learned for real tasks; experiments on test problems demonstrate only that search is biased towards the intersection, not that the intersection is the right place to be. Learning Many Related Tasks at the Same Time with Backpropagation 661 Table 1 shows the performance of single and multitask backprop (STB and MTB, respectively) on 1D-ALVINN using nets with one hidden layer. The MTB net has 32 inputs, 16 hidden units, and 9 outputs. The 36 STB nets have 32 inputs, 2, 4, 8 or 16 hidden units, and 1 output. A similar experiment using nets with 2 hidden layers containing 2, 4, 8, 16, or 32 hidden units per layer for STB and 32 hidden units per layer for MTB yielded comparable results. The size of the MTB nets is not optimized in either experiment. Table 1: Performance of STB and MTB with One Hidden Layer on 1D-ALVINN II TASK 1 or 2 Lanes Left Edge Right Edge Line Center Road Center Road Greylevel Edge Greylevel Line Greylevel Steering ROOT-MEAN SQUARED ERROR ON TEST SET Single Task Backprop MTB % Change % Change Best STB Mean STB 16HU 2HU T4HU I 8HU f 16HU .201 .209 .207 .178 .156 14.1 27.4 .069 .071 .073 .073 .062 11.3 15.3 .076 .062 .058 .056 .051 9.8 23.5 .152 .152 .151 .153 .152 0.7 0.8 .038 .037 .039 .042 .034 8.8 14.7 .055 .055 .054 .038 42.1 43.4 .054 .038 .037 .038 .039 .038 -2.6 0.0 .054 .054 .054 .054 0.0 0.0 .054 .069 .093 .087 .072 .058 19.0 38.4 The entries under the STB and MTB headings are the peak generalization error for nets of the specified size. The italicized STB entries are the STB runs that yielded best performance. The last two columns compare STB and MTB. The first is the percent difference between MTB and the best STB run. Positive percentages indicate MTB performs better. This test is biased towards STB because it compares a single run of MTB on an unoptimized net size with several independent runs of STB that use different random seeds and are able to find near-optimal net size. The last column is the percent difference between MTB and the average STB. Note that on the important steering task, MTB outperforms STB 20-40%. 3.2 ID-DOORS To test multitask backprop on a real problem, we created an object recognition domain similar in some respects to 1D-ALVINN. In 1D-DOORS the main tasks are to locate doorknobs and recognize door types (single or double) in images of doors collected with a robot-mounted camera. Figure 1 shows several door images. As with 1D-ALVINN, the problem was simplified by using horizontal stripes from image, one for the green channel and one for the blue channel. Each stripe is 30 pixels wide (accomplished by applying Gaussian smoothing to the original 150 pixel wide image) and occurs at the vertical location of the doorknob. 10 tasks were used: ? ? ? ? ? horizontal location of doorknob horizontal location of doorway center horizontal location of left door jamb width of left door jamb horizontal location of left edge of door ? ? ? ? ? single or double door width of doorway horizontal location of right door jamb width of right door jamb horizontal location of right edge of door The difficulty of 1D-DOORS precludes running as exhaustive a set of experiments as with 1D-ALVINN: runs were done only for the two most important and difficult tasks: doorknob location and door type. STB was tested on nets using 6, 24, and 96 hidden units. MTB was tested on a net with 120 hidden units. The results 662 Rich Caruana Figure 1: Sample Doors from the ID-DOORS Domain are in Table 2. STB generalizes 35-45% worse than MTB on these tasks. Less thorough experiments on the other eight tasks suggest MTB probably always yields performance equal to or better than STB. Table 2: Performance of STB and MTB on ID-DOORS. RMS ERROR ON TEST SET TASK Doorknob Loc Door Type 4 DISCUSSION In our experience, multitask backprop usually generalizes better than single task backprop. The few cases where STB has been better is on simpler tasks, and there the difference between MTB and STB was small. Multitask backprop appears to provide the most benefit on hard tasks. MTB also usually learns in fewer epochs than STB. When all tasks must be learned, MTB is computationally more efficient than training single nets. When few tasks are important, however, STB is usually more efficient (but also less accurate). Tasks do not always learn at the same rate. It is important to watch the training curve of each MTB task individually and stop training each task when its performance peaks. The easiest way to do this to take a snapshot of the net when performance peaks on a task of interest. MTB does not mean one net should be used to predict all tasks, only that all tasks should be trained on one net so they may benefit each other. Do not treat tasks as one task just because they are being trained on one net! Balancing tasks (e.g., using different learning rates for different outputs) sometimes helps tasks learn at similar rates, thus maximizing the potential benefits of MTB. Also, because the training curves for MTB are often more complex due to interactions between tasks (MTB curves are frequently multimodal), it is important to train MTB nets until all tasks appear to be overtraining. Restricting the capacity of MTB nets to force sharing or prevent overtraining usually hurts performance instead of helping it. MTB does not depend on restricted net capacity. We created the extra tasks in ID-ALVINN and ID-DOORS specifically because we thought they would improve performance on the important tasks. Multitask backprop can be used in other ways. Often the world gives us related tasks to learn. For example, the Calendar Apprentice System (CAP)[4] learns to predict the Location, Time_Of _Day, Day_Of _Week, and Duration of the meetings it schedules. These tasks are functions of the same data, share many common features, Learning Many Related Tasks at the Same Time with Backpropagation 663 and would be easy to learn together. Sometimes the world gives us related tasks in mysterious ways. For example, in a medical domain we are examining where the goal is to predict illness severity, half of the lab tests are cheap and routinely measured before admitting a patient (e.g., blood pressure, pulse, age). The rest are expensive tests requiring hospitalization. Users tell us it would be useful to predict if the severity of the illness warrants admission (and further testing) using just the pre-admission tests. Rather than ignore the most diagnostically useful information in the database, we use the expensive tests as additional tasks the net must learn. They are not very predictable from the simple pre-admission tests, but providing them to the net as outputs helps it learn illness severity better. Multitask backprop is one way of providing to a net information that at run time would only be available in the future. The training signals are needed only for the training set because they are outputs-not inputs-to the net. 5 RELATED WORK Training nets with many outputs is not new; NETtalk [9] used one net to learn phonemes and stress. This approach was natural for NETtalk where the goal was to control a synthesizer that needed both phoneme and stress commands at the same time. No analysis, however, was made of the advantages of using one net for all the tasks3 , and the different outputs were not treated as independent tasks. For example, the NETtalk stress task overtrains badly long before the phoneme task is learned well, but NETtaik did not use different snapshots of the net for different tasks. NETtalk also made no attempt to balance tasks so that they would learn at a similar rate, or to add new tasks that might improve learning but which would not be useful for controlling the synthesizer. Work has been done on serial transfer between nets [8]. Improved learning speed was reported, but not improved generalization. The key difficulties with serial transfer are that it is difficult to scale to many tasks, it is hard to prevent catastrophic interference from erasing what was learned previously, the learning sequence must be defined manually, and serial learning precludes mutual benefit between tasks. This work is most similar to catalytic hints [1][10] where extra tasks correspond to important learnable features of a main task. This work extends hints by showing that tasks can be related in more diverse ways, by expanding the class of mechanisms responsible for multitask backprop, by showing that capacity restriction is not an important mechanism for multitask backprop [2], and by demonstrating that creating many new related tasks may be an efficient way of providing domain-specific inductive bias to backprop nets. 6 FUTURE WORK We used vanilla backprop to show the benefit of training many related tasks on one net. Additional techniques may enhance the effects. Regularization and incremental net growing procedures might improve performance by promoting sharing without restricting capacity. New techniques may also be necessary to enhance the benefit of multitask backprop. Automatic balancing of task learning rates would make MTB easier to use. It would also be valuable to know when the different MTB mechanisms are working-they might be useful in different kinds of domains and might benefit from different regularization or balancing techniques. Finally, although MTB usually seems to help and rarely hurts, the only way to know it 3S ee [5] for evidence that NETtalk is harder to learn using separate nets. 664 Rich Caruana helps is to try it. It would be better to have a predictive theory of how tasks should relate to benefit MTB , particularly if new tasks are to be created only to provide a multitask benefit for the other important tasks in the domain. 7 SUMMARY Five mechanisms that improve generalization performance on nets trained on multiple related tasks at the same time have been identified. These mechanisms work without restricting net capacity or otherwise reducing the net's VC-dimension. Instead, they exploit backprop's ability to combine the error terms for related tasks into an aggregate gradient that points towards better underlying represen tations. Multitask backprop was tested on a simulated domain, ID-ALVINN, and on a real domain, ID-DOORS . It improved generalization performance on hard tasks in these domains 20-40% compared with the best performance that could be obtained from multiple trials of single task backprop. Acknowledgements Thanks to Tom Mitchell, Herb Simon, Dean Pomerleau, Tom Dietterich, Andrew Moore, Dave Touretzky, Scott Fahlman, Sebastian Thrun, Ken Lang, and David Zabowski for suggestions that have helped shape this work. This research is sponsored in part by the Advanced Research Projects Agency (ARPA) under grant no. F33615-93-1-1330. References [1] Y.S. Abu-Mostafa, "Learning From Hints in Neural Networks," Journal of Complexity 6:2, pp. 192-198,1989. [2] Y.S. Abu-Mostafa, "Hints and the VC Dimension," Neural Computation, 5:2, 1993. [3] R. Caruana, "Multitask Connectionist Learning," Proceedings of the 1993 Connectionist Models Summer School, pp. 372-379, 1993. [4] L. Dent, J. Boticario, J. McDermott, T. Mitchell, and D. Zabowski, "A Personal Learning Apprentice," Proceedings of 1992 National Conference on Artificial Intelligence, 1992. [5] T.G. Dietterich, H. Hild, and G. Bakiri, "A Comparative Study of ID3 and Backpropagation for English Text-to-speech Mapping," Proceedings of the Seventh International Conference on Artificial Intelligence, pp. 24-31, 1990. [6] G.E. Hinton, "Learning Distributed Representations of Concepts," Proceedings of the Eight International Conference of The Cognitive Science Society, pp. 112, 1986. [7] D.A. Pomerleau, "Neural Network Perception for Mobile Robot Guidance," Carnegie Mellon University: CMU-CS-92-115, 1992. [8] L.Y. Pratt, J. Mostow, and C.A. Kamm, "Direct Transfer of Learned Information Among Neural Networks," Proceedings of AAAI-91, 1991. [9] T.J. Sejnowski and C.R. Rosenberg, "NETtalk: A Parallel Network that Learns to Read Aloud ," John Hopkins: JHU/EECS-8'6/01, 1986. [10] S.C. Suddarth and A.D.C. Holden, "Symbolic-neural Systems and the Use of Hints for Developing Complex Systems," International Journal of MaxMachine Studies 35:3, pp. 291-311, 1991.
959 |@word multitask:17 trial:1 version:1 briefly:1 stronger:1 seems:1 nd:1 hu:4 pulse:1 pressure:1 harder:1 initial:1 loc:1 outperforms:1 lang:1 yet:2 synthesizer:2 must:4 john:1 realistic:1 shape:1 cheap:1 concert:1 sponsored:1 v:1 alone:3 half:1 selected:2 fewer:1 intelligence:2 provides:3 location:13 preference:3 centerline:2 simpler:1 five:5 admission:3 direct:1 combine:1 expected:1 overtrain:1 frequently:1 growing:1 multi:2 simulator:1 kamm:1 begin:2 project:1 underlying:3 moreover:1 what:5 easiest:1 kind:1 developed:1 thorough:2 every:1 act:1 control:1 unit:6 medical:1 grant:1 appear:1 positive:1 before:2 treat:1 apparatus:1 consequence:1 id:10 noteworthy:1 eavesdropping:4 might:4 conversely:1 limited:1 camera:1 responsible:1 testing:3 lost:1 backpropagation:5 procedure:3 empirical:3 evolving:1 jhu:1 thought:1 pre:2 road:14 tweaking:1 suggest:1 symbolic:1 get:3 selection:6 applying:1 restriction:1 dean:1 center:4 maximizing:1 independently:1 duration:1 x32:1 pull:1 hurt:2 controlling:1 suppose:3 user:1 us:4 distinguishing:1 hypothesis:1 pa:1 recognition:2 particularly:2 expensive:2 stripe:2 database:1 blocking:1 capture:1 region:3 valuable:1 ran:1 mentioned:1 substantial:1 predictable:1 agency:1 complexity:3 personal:1 trained:12 depend:1 predictive:1 creates:1 multimodal:1 differently:1 routinely:1 train:1 forced:1 effective:3 sejnowski:1 artificial:3 tell:1 aggregate:1 exhaustive:1 aloud:1 larger:2 otherwise:1 precludes:2 calendar:1 ability:1 id3:1 advantage:1 sequence:1 net:72 interaction:1 relevant:3 entered:1 amplification:12 regularity:5 contrived:1 double:2 comparative:1 incremental:1 object:2 help:7 derive:2 andrew:1 mostow:1 measured:1 school:2 strong:1 c:2 indicate:3 direction:2 attribute:6 stochastic:1 vc:2 backprop:33 generalization:10 dent:1 helping:1 hild:1 seed:1 mapping:1 predict:5 week:1 mostafa:3 individually:1 always:3 gaussian:1 modified:1 rather:2 mobile:1 command:1 rosenberg:1 derived:1 catalytic:2 detect:1 holden:1 initially:1 hidden:18 fight:1 unoptimized:1 pixel:6 among:1 smoothing:1 mutual:1 equal:1 once:2 never:1 manually:1 identical:1 warrant:1 future:3 others:1 connectionist:2 hint:7 few:3 simultaneously:2 recognize:1 national:1 attempt:1 interest:1 hospitalization:1 admitting:1 accurate:1 edge:7 necessary:1 experience:1 culture:1 initialized:1 guidance:1 arpa:1 column:2 herb:1 caruana:7 retains:1 entry:2 comprised:1 examining:1 seventh:1 too:1 reported:1 corrupted:1 eec:1 combined:1 chunk:1 thanks:1 peak:3 international:3 told:1 enhance:2 together:1 hopkins:1 squared:1 aaai:1 containing:1 worse:1 cognitive:1 creating:1 potential:1 satisfy:1 explicitly:1 helped:1 root:1 lab:1 try:1 start:1 aggregation:1 parallel:1 simon:1 phoneme:3 yield:2 identify:1 correspond:1 generalize:1 emphasizes:1 trajectory:1 dave:1 overtraining:2 mtb:32 touretzky:1 sharing:2 sebastian:1 mysterious:1 pp:5 sampled:1 stop:1 ask:1 mitchell:2 cap:1 improves:2 schedule:1 carefully:1 appears:1 day:1 tom:2 improved:3 done:3 just:4 until:1 working:1 horizontal:8 bordering:1 perhaps:1 effect:4 dietterich:2 verify:1 true:1 concept:1 requiring:1 inductive:2 regularization:2 read:1 moore:1 nettalk:6 width:3 stress:3 outline:1 demonstrate:1 performs:1 interpreting:1 percent:2 image:7 common:4 physical:1 extend:1 illness:3 mellon:2 blocked:3 significant:1 automatic:1 vanilla:1 teaching:1 robot:2 similarity:1 surface:1 add:1 something:1 curvature:1 irrelevant:2 discussing:1 meeting:1 accomplished:1 mcdermott:1 tide:4 minimum:9 additional:3 steering:3 determine:2 signal:9 ii:1 multiple:7 suddarth:1 long:1 lin:1 serial:3 equally:2 controlled:1 patient:1 cmu:2 represent:2 sometimes:3 grow:1 extra:5 biased:3 rest:1 probably:1 subject:1 thing:5 stb:23 flow:1 call:1 ee:1 near:3 door:20 easy:3 enough:1 subfeatures:1 pratt:1 affect:2 architecture:2 identified:2 idea:1 computable:1 whether:1 rms:1 speech:1 cause:3 useful:7 maybe:1 band:1 processed:1 ken:1 simplest:2 generate:1 percentage:1 alters:1 per:2 blue:1 diverse:1 carnegie:2 abu:4 key:1 demonstrating:1 blood:1 prevent:2 run:7 you:2 place:1 extends:1 prefer:4 comparable:1 layer:14 summer:1 yielded:2 badly:1 represen:1 lane:4 speed:2 f33615:1 performing:1 conjecture:1 developing:1 describes:1 beneficial:1 smaller:1 happens:2 restricted:1 interference:1 computationally:1 previously:1 discus:1 mechanism:19 needed:2 know:2 generalizes:3 available:1 experimentation:1 eight:3 nettaik:1 promoting:1 away:1 apprentice:2 original:3 running:1 recognizes:3 exploit:2 bakiri:1 society:1 added:1 occurs:5 gradient:4 separate:2 unable:1 diagnostically:1 capacity:6 simulated:1 thrun:1 italicized:1 collected:1 eavesdropper:1 relationship:2 providing:3 convincing:1 demonstration:1 balance:1 zabowski:2 difficult:3 relate:1 pomerleau:3 perform:3 vertical:1 snapshot:2 finite:1 hinton:2 looking:1 severity:3 locate:1 intensity:3 david:1 specified:1 connection:3 optimized:1 learned:10 able:1 usually:8 pattern:5 scott:1 perception:1 program:1 green:1 overlap:1 difficulty:3 force:1 natural:1 treated:1 advanced:1 improve:7 tations:1 created:4 text:1 epoch:1 acknowledgement:1 law:1 suggestion:1 mounted:1 age:1 sufficient:1 principle:1 share:8 balancing:3 erasing:1 summary:1 surprisingly:1 parity:1 last:2 fahlman:1 heading:1 english:1 bias:8 tick:1 allow:1 fall:6 wide:2 benefit:10 distributed:1 curve:4 dimension:2 world:2 rich:5 sensory:1 domainspecific:1 made:3 simplified:1 eavesdrop:2 ignore:1 mustafa:1 summing:2 pittsburgh:1 doorway:2 search:4 table:4 learn:27 channel:2 transfer:3 expanding:1 alvinn:11 complex:3 domain:17 did:1 main:3 noise:5 arise:1 crafted:1 learns:6 specific:1 showing:3 learnable:4 evidence:2 exists:1 restricting:3 easier:3 intersection:5 likely:2 watch:1 monotonic:1 cite:1 goal:2 towards:6 hard:4 change:3 infinite:1 specifically:1 reducing:1 averaging:1 catastrophic:1 rarely:3 select:1 scan:1 tested:3
7,199
960
Non-linear Prediction of Acoustic Vectors Using Hierarchical Mixtures of Experts S.R.Waterhouse A.J.Robinson Cambridge University Engineering Department, Trumpington St ., Cambridge, CB2 1PZ, England. Tel: [+44] 223 332800, Fax: [+44] 223 332662, Email: srwlO01.ajr@eng.cam.ac.uk URL: http://svr-www.eng.cam.ac.ukr srw1001 Abstract In this paper we consider speech coding as a problem of speech modelling. In particular, prediction of parameterised speech over short time segments is performed using the Hierarchical Mixture of Experts (HME) (Jordan & Jacobs 1994). The HME gives two advantages over traditional non-linear function approximators such as the Multi-Layer Percept ron (MLP); a statistical understanding of the operation of the predictor and provision of information about the performance of the predictor in the form of likelihood information and local error bars. These two issues are examined on both toy and real world problems of regression and time series prediction. In the speech coding context, we extend the principle of combining local predictions via the HME to a Vector Quantization scheme in which fixed local codebooks are combined on-line for each observation. 1 INTRODUCTION We are concerned in this paper with the application of multiple models, specifically the Hierarchical Mixtures of Experts, to time series prediction, specifically the problem of predicting acoustic vectors for use in speech coding. There have been a number of applications of multiple models in time series prediction. A classic example is the Threshold Autoregressive model (TAR) which was used by Tong & 836 S. R. Waterhouse, A. J. Robinson Lim (1980) to predict sunspot activity. More recently, Lewis, Kay and Stevens (in Weigend & Gershenfeld (1994)) describe the use of Multivariate and Regression Splines (MARS) to the prediction of future values of currency exchange rates. Finally, in speech prediction, Cuperman & Gersho (1985) describe the Switched Inter-frame Vector Prediction (SIVP) method which switches between separate linear predictors trained on different statistical classes of speech. The form of time series prediction we shall consider in this paper is the single step prediction fI(t) of a future quantity y(t) , by considering the previous samples. This may be viewed as a regression problem over input-output pairs {x t), y(t)}~ where x(t) is the lag vector (y(t-I), y(t-2), ... , y(t- p ?. We may perform this regression using standard linear models such as the Auto-Regressive (AR) model or via nonlinear models such as connectionist feed-forward or recurrent networks. The HME overcomes a number of problems associated with traditional connectionist models via its architecture and statistical framework. Recently, Jordan & Jacobs (1994) and Waterhouse & Robinson (1994) have shown that via the EM algorithm and a 2nd order optimization scheme known as Iteratively Reweighted Least Squares (IRLS), the HME is faster than standard Multilayer Perceptrons (MLP) by at least an order of magnitude on regression and classification tasks respectively. Jordan & Jacobs also describe various methods to visualise the learnt structure of the HME via 'deviance trees' and histograms of posterior probabilities. In this paper we provide further examples of the structural relationship of the trained HME and the input-output space in the form of expert activation plots. In addition we describe how the HME can be extended to give local error bars or measures of confidence in regression and time series prediction problems. Finally, we describe the extension of the HME to acoustic vector prediction, and a VQ coding scheme which utilises likelihood information from the HME. c: 2 HIERARCHICAL MIXTURES OF EXPERTS The HME architecture (Figure 1) is based on the principle of 'divide and conquer' in which a large, hard to solve problem is broken up into many, smaller, easier to solve problems. It consists of a series of 'expert networks' which are trained on different parts of the input space. The outputs of the experts are combined by a 'gating network' which is trained to stochastically select the expert which is performing best at solving a particular part of the problem. The operation of the HME is as follows: the gating networks receive the input vectors x(t) and produce as outputs probabilities P(mi/.x(t), 7'/j) for each local branch mj of assigning the current input to the different branches, where T/j are the gating network parameters. The expert networks sit at the leaves of the tree and each output a vector flJt) given input vector x(t) and parameters Bj . These outputs are combined in a weighted sum by P(mjlX<t), T/j) to give the overall output vector for this region. This procedure continues recursively upwards to the root node. In time series prediction, each expert j is a linear single layer network with the form: flY) = B; x (t) where B; is matrix and form to an AR model. x(t) is the lag vector discussed earlier, which is identical in Non-Linear Prediction of Acoustic Vectors Using Hierarchical Mixtures of Experts x x x 837 x Figure 1: The Hierarchical Mixture of Experts. 2.1 Error bars via HME Since each expert is an AR model, it follows that the output of each expert y(t) is the expected value of the observations y(t) at each time t. The conditional likelihood of yet) given the input and expert mj is P(y(t) Ix (t), mj, Bj) = 12:Cj I exp ( - ~ (y - yy?)T Cj(y - yjt))) where Cj is the covariance matrix for expert mj which is updated during training as: C = _1_ "'" h(t)(y(r) _ y(t)l (y(t) _ y~t)) J ' " h~t) L..J J ] ] L.Jt J t where hy) are the posterior probabilities I of each expert mj' Taking the moments of the overall likelihood of the HME gives the output of the HME as the conditional expected value of the target output yct), yet) = E(yct)lxct) , 0, M) = 2: P(mjlxct), l1j)E(y(t)lx ct), ej,mj) =2: gY)iJ/t), j j Where M represents the overall HME model and Taking the second central moment of yct) gives, C = = e the overall set of parameters. I E?y(t) - yy?)2 xct), 0, M) 2: P(mJlx(t), l1j)E?y(t) 2: gjt)(Cj + yjt). iJj(t)T), yjt))2I x (t), ej, mj) j = j lSee (Jordan & Jacobs 1994) for a fuller discussion of posterior probabilities and likelihoods in the context of the HME. 838 S. R. Waterhouse, A. J. Robinson which gives, in a direct fashion, the covariance of the output given the input and the model. If we assume that the observations are generated by an underlying model, which generates according to some function f(x(t)) and corrupted by zero mean normally distributed noise n(x) with constant covariance 1:, then the covariance of y(t) is given by, V(y(t)) =V(t o) + 1:, so that the covariance computed by the method above, V(y(t)) , takes into account the modelling error as well as the uncertainty due to the noise. Weigend & Nix (1994) also calculate error bars using an MLP consisting of a set of tanh hidden units to estimate the conditional mean and an auxiliary set of tanh hidden units to estimate the variance, assuming normally distributed errors. Our work differs in that there is no assumption of normality in the error distribution, rather that the errors of the terminal experts are distributed normally, with the total error distribution being a mixture of normal distributions. 3 SIMULATIONS In order to demonstrate the utility of our approach to variance estimation we consider one toy regression problem and one time series prediction problem. 3.1 Toy Problem: Computer generated data 2,------..--...., 0 . 08,------~-..., 1.5 O. OS N ~ fO.04 ~ -0.02 2 x ~-0.5 -1 -0.2 -1.5 -0.4 0.8 Q) go.S -2 -O.S -2.5 -0.8 -3 '------'------' o 2 x .~ ~0.4 -1 ' - - - - - - ' - - - - - " o 2 x x 2 Figure 2: Performance on the toy data set of a 5 level binary HME. (a) training set (dots) and underlying function f(x) (solid), (b) underlying function (solid) and prediction y(x) (dashed), (c) squared deviation of prediction from underlying function, (d) true noise variance (solid) and variance of prediction (dashed). By way of comparison, we used the same toy problem as Weigend & Nix (1994) which consists of 1000 training points and 10000 separate evaluation points from Non-Linear Prediction of Acoustic Vectors Using Hierarchical Mixtures of Experts 839 the function g(x) where g(x) consists of a known underlying function f(x) corrupted by normally distributed noise N(O, (J2(X)) , f(x) = sin(2.5x) x sin(l. 5x), (J2(x) = 0.01 + O. 25 x [1 - sin(2. 5x)f. As can be seen by Figure 2, the HME has learnt to approximate both the underlying function and the additive noise variance. The deviation of the estimated variance from the "true" noise variance may be due to the actual noise variance being lower than the maximum denoted by the solid line at various points. 3.2 Sunspots 1920 1930 1940 1950 1960 1970 1980 Year 8 6 t 8.4 )( w 2 . . -.-- . ? ? ? ? ? .. ? ... II ? l- ? - ? I ? - ? - .. - ? -? ? ? ? OL-------~--------~--------~--------L-------~--------~ 1920 1930 1940 1950 1960 1970 1980 Year Figure 3: Performance on the Sunspots data set. (a) Actual Values (x) and predicted values (0) with error bars. (b) Activation of the expert networks; bars wide in the vertical axis indicate strong activation. Notice how expert 7 concentrates on the lulls in the series while expert 2 deals with the peaks. I METHOD I MLP TAR HME NMSE ' Train 1700-1920 0.082 0.097 0.061 Test 1921-1955 1956-1979 0.086 0.35 0.097 0.28 0.089 0.27 Table 1: Results of single step prediction on the Sunspots data set using a mixture of 7 experts (104 parameters) and a lag vector of 12 years. NMSE' is the NMSE normalised by the variance of the entire record 1700 to 1979. 840 S. R. Waterhouse, A. J. Robinson The Sunspots2 time series consists of yearly sunspot activity from 1700 to 1979 and was first tackled using connectionist models by Weigend, Huberman & Rumelhart (1990) who used a 12-8-1 MLP (113 parameters) . Prior to this work, the TAR was used by Tong (1990). Our results, which were obtained using a random leave 10% out cross validation method, are shown in Table 1. We are considering only single step prediction on this problem, which involves prediction of the next value based on a set of previous values of the time series. Our results are evaluated in terms of Normalised Mean Squared Error (NMSE) (Weigend et al. 1990), which is defined as the ratio of the variance of the prediction on the test set to the variance of the test set itself. The HME outperforms both the TAR and the MLP on this problem, and additionally provides both information about the structure of the network after training via the expert activation plot and error bars of the predictions, as shown in Figure 3. Further improvements may be possible by using likelihood information during cross validation so that a joint optimisation of overall error and variance is achieved. 4 SPEECH CODING USING HME In the standard method of Linear Predictive Coding (LPC) (Makhoul 1975), speech is parametrised into a set of vectors of duration one frame (around 10 ms). Whilst simple scalar quantization of the LPC vectors can achieve bit rates of around 2400 bits per second (bps), Yong, Davidson & Gersho (1988) have shown that simple linear prediction of Line Spectral Pairs (LSP) (Soong & Juang 1984) vectors followed by Vector Quantization (VQ) (Abut , Gray & Rebolledo 1984) of the error vectors can yield bit rates of around 800 bps. In this paper we describe a speech coding framework which uses the HME in two stages. Firstly, the HME is used to perform prediction of the acoustic vectors. The error vectors are then quantized efficiently by using a VQ scheme which utilises the likelihood information derived from the HME . 4.1 Mixing VQ codebooks ia Gating networks In a VQ scheme using a Euclidean distance measure , there is an implicit assumption that the inputs follow a Gaussian probability density function (pdf). This is satisfied if we quantize the residuals from a linear predictor , but not the residuals from an HME which follow a mixture of Gaussians pdf. A more efficient method is therefore to generate separate VQ code books for each expert in the HME and combine them via the priors on each expert from the gating networks. The code book for the overall residual vectors on the test set is then generated at each time dynamically by choosing the first D x gjt) codes, where D is the size of the expert codebooks and gY) is the prior on each expert. 2 Available via anonymous DataSunspots.Yearly ftp at fip.cs.colorado.edu III jpub jTime-Series as Non-Linear Prediction of Acoustic Vectors Using Hierarchical Mixtures of Experts 4.2 841 Results of Speech Coding Evaluations Initial experiments were performed using 23 Mel scale log energy frequency bins as acoustic vectors and using single variances Cj = (J;I as expert network covariance matrices. The results of training over 100 ,000 frames and evaluation over a further 100 ,000 frames on the Resource Management (RM) corpus are shown in Table 2 and Figure 4 which shows the good specialisation of the HME in this problem. METHOD Prediction Train 12.07 18.1 20.20 Linear 1 level HME 2 level HME Gain (dB) Test 10.95 15.55 16.39 Table 2: Prediction of Acoustic Vectors using linear prediction and binary branching HMEs with 1 and 2 levels. Prediction gain (Cuperman & Gersho 1985) is the ratio of the signal variance to prediction error variance. (a) Spectogram 20 !!! ::3 c;; 15 Cf ~ 10 ::3 8 < . 5 o 10 20 30 40 50 60 70 80 90 100 Frame (b) Gating Network Decisions 71--- - - - I - -__.- I 6~----. . . .~.---~--_ _--------_=--~~--? ~5~~---------1~---.I---1IHI~I~?--. . . .~.~.----~4~~~----------------~~~~---------- w3~~---------~-----. .~__------_.----------------- - 2.-.---- - -- - - -11-1...---11---1- --11-1 ? ? I II . ....- I ? 1~--------------~~~~--------------------------------~ 10 20 30 40 50 60 70 80 90 100 Frame Figure 4: The behaviour of a mixture of 7 experts at predicting Mel-scale log energy frequency bins over 100 16ms fram es. The top figure is a spectrogram of the speech and the lower figure is an expert activation plot, showmg the gating network decisions. We have conducted further experiments using LSPs and cepstrals as acoustic vectors , and using diagonal expert network covariance matrices, on a very large speech corpus. However, initial experiments show only a small improvement in gain over a single linear predictor and further investigation is underway. We have also coded acoustic vectors using 8 bits per frame with frame lengths of 12.5 ms, passing power, pitch and degree of voicing as side band information , without appreciable distortion over simple LPC coding. A full system will include prediction of all acoustic 842 S. R. Waterhouse, A. J. Robinson parameters and we anticipate further reductions on this initial figure with future developments. 5 CONCLUSION The aim of speech coding is the efficient coding of the speech signal with little perceptual loss. This paper has described the use of the HME for acoustic vector prediction. We have shown that the HME can provide improved performance over a linear predictor and in addition it provides a time varying variance for the prediction error. The decomposition of the linear prediction problem into a solution via a mixture of experts also allows us to construct a VQ codebook on the fly by mixing the codebooks of the various experts. We expect that the direct computation of the time varying nature of the prediction accuracy will find many applications. Within the acoustic vector prediction problem we would like to exploit this information by exploring the continuum between the fixed bit rate coder described here and a variable bit rate coder that produces constant spectral distortion. Acknowledgements This work was funded in part by Hewlett Packard Laboratories, UK. Steve Waterhouse is supported by an EPSRC Research Students hip and Tony Robinson was supported by a EPSRC Advanced Research Fellowship. References Abut, H., Gray, R. M. & Rebolledo, G. (1984), 'Vector quantization of speech and speechlike waveforms', IEEE Transactions on Acoustics, Speech, and Signal Processing. Cuperman, V. & Gersho, A. (1985), 'Vector predictive coding of speech at 16 kblt/s', IEEE Transactions on Communications COM-33, 685-696. Jordan, M. I. & Jacobs, R. A. (1994), 'Hierarchical Mixtures of Experts and the EM algorithm', Neural Computation 6, 181-214. Makhoiil, J. (1975), 'Linear prediction: A tutorial review', Proceedings of the IEEE 63(4) 561-580. Soong~ F. k. & Juang, B. H. (1984), Line spectrum pair (LSP) and speech data compressIon. Tong, H. (1990), Non-linear Time Series: a dynamical systems approach, Oxford Universi~y Press. Tong, H. & Lim, K. (1980), 'Threshold autoregression, limit cycles and cyclical data', Journal of Royal Statistical Society. Waterhouse, S. R. & Robinson, A. J. (1994), Classification using hierarchical mixtures of experts, in 'IEEE Workshop on Neural Networks for SigI!al Processing'. Weigend, A. S. & Gershenfeld, N. A. (1994), Time Series Prediction: Forecasting the Future and Understanding the Past) Addison-Wesley. Weigend, A. S. & Nix, D. A. (1994), Predictions with confidence intervals (local error bars), Technical Report CU-CS-724-94, Department of Computer Science and Institute of Coznitive Science, University of Colorado, Boulder, CO 80309-0439. Weigend, A. S., Huberman, B. A. & Rumelhart, D. E. (1990), 'Predicting the future: a connectionist approach', International Journal of Neural Systems 1, 193-209. Yong, M., Davidson, G. & Gersho, A. (1988), Encoding of LPC spectral parameters using switched-adaptive interframe vector prediction, in 'Proceedings of the IEEE International Conference on Acoustics Speech, and Signal Processing', pp. 402-405.
960 |@word cu:1 compression:1 yct:3 nd:1 simulation:1 eng:2 jacob:5 covariance:7 decomposition:1 solid:4 recursively:1 reduction:1 moment:2 initial:3 series:14 outperforms:1 past:1 current:1 com:1 activation:5 assigning:1 yet:2 additive:1 fram:1 plot:3 leaf:1 short:1 record:1 regressive:1 provides:2 quantized:1 node:1 ron:1 lx:1 codebook:1 firstly:1 universi:1 direct:2 consists:4 combine:1 inter:1 expected:2 multi:1 ol:1 terminal:1 actual:2 little:1 considering:2 underlying:6 coder:2 whilst:1 rm:1 uk:2 normally:4 unit:2 engineering:1 local:6 limit:1 encoding:1 oxford:1 examined:1 dynamically:1 co:1 ihi:1 differs:1 cb2:1 procedure:1 confidence:2 deviance:1 svr:1 context:2 www:1 go:1 duration:1 kay:1 classic:1 lsee:1 updated:1 target:1 colorado:2 us:1 rumelhart:2 continues:1 epsrc:2 fly:2 calculate:1 region:1 cycle:1 broken:1 cam:2 trained:4 solving:1 segment:1 predictive:2 joint:1 various:3 train:2 describe:6 choosing:1 lag:3 solve:2 distortion:2 itself:1 advantage:1 j2:2 combining:1 mixing:2 achieve:1 juang:2 produce:2 leave:1 ftp:1 recurrent:1 ac:2 ij:1 strong:1 auxiliary:1 predicted:1 involves:1 indicate:1 c:2 concentrate:1 waveform:1 stevens:1 bin:2 exchange:1 behaviour:1 anonymous:1 investigation:1 anticipate:1 extension:1 exploring:1 around:3 normal:1 exp:1 predict:1 bj:2 continuum:1 estimation:1 tanh:2 weighted:1 gaussian:1 aim:1 rather:1 ej:2 tar:4 varying:2 derived:1 improvement:2 modelling:2 likelihood:7 entire:1 hmes:1 hidden:2 issue:1 classification:2 overall:6 denoted:1 development:1 construct:1 fuller:1 identical:1 represents:1 future:5 connectionist:4 spline:1 report:1 consisting:1 mlp:6 evaluation:3 mixture:15 hewlett:1 parametrised:1 tree:2 divide:1 euclidean:1 hip:1 earlier:1 ar:3 deviation:2 abut:2 predictor:6 conducted:1 ukr:1 corrupted:2 learnt:2 combined:3 st:1 density:1 peak:1 international:2 squared:2 central:1 satisfied:1 management:1 gjt:2 stochastically:1 book:2 expert:37 lull:1 toy:5 account:1 hme:32 gy:2 coding:12 student:1 performed:2 root:1 square:1 l1j:2 accuracy:1 variance:16 who:1 percept:1 efficiently:1 yield:1 irls:1 fo:1 email:1 energy:2 frequency:2 pp:1 associated:1 mi:1 gain:3 lim:2 provision:1 cj:5 feed:1 steve:1 wesley:1 follow:2 improved:1 evaluated:1 mar:1 parameterised:1 stage:1 implicit:1 nonlinear:1 o:1 gray:2 true:2 iteratively:1 laboratory:1 deal:1 reweighted:1 sin:3 during:2 branching:1 mel:2 m:3 pdf:2 demonstrate:1 upwards:1 recently:2 fi:1 lsp:2 visualise:1 extend:1 discussed:1 cambridge:2 funded:1 dot:1 multivariate:1 posterior:3 binary:2 approximators:1 seen:1 utilises:2 spectrogram:1 dashed:2 ii:2 branch:2 multiple:2 currency:1 full:1 signal:4 technical:1 faster:1 england:1 cross:2 yjt:3 coded:1 prediction:43 pitch:1 regression:7 multilayer:1 optimisation:1 histogram:1 achieved:1 receive:1 addition:2 fellowship:1 interval:1 db:1 xct:1 jordan:5 structural:1 iii:1 concerned:1 switch:1 w3:1 architecture:2 codebooks:4 utility:1 url:1 forecasting:1 speech:20 passing:1 band:1 http:1 generate:1 tutorial:1 notice:1 estimated:1 per:2 yy:2 shall:1 threshold:2 gershenfeld:2 sum:1 year:3 weigend:8 uncertainty:1 decision:2 bit:6 layer:2 ct:1 followed:1 tackled:1 activity:2 hy:1 yong:2 generates:1 performing:1 department:2 trumpington:1 according:1 makhoul:1 smaller:1 em:2 soong:2 boulder:1 resource:1 vq:7 addison:1 gersho:5 available:1 operation:2 gaussians:1 autoregression:1 hierarchical:10 spectral:3 voicing:1 ajr:1 top:1 cf:1 include:1 tony:1 yearly:2 exploit:1 conquer:1 society:1 quantity:1 traditional:2 diagonal:1 distance:1 separate:3 assuming:1 code:3 length:1 relationship:1 ratio:2 perform:2 vertical:1 observation:3 nix:3 extended:1 communication:1 frame:8 pair:3 acoustic:16 robinson:8 bar:8 dynamical:1 lpc:4 packard:1 royal:1 ia:1 power:1 predicting:3 residual:3 advanced:1 normality:1 scheme:5 axis:1 auto:1 fax:1 prior:3 understanding:2 acknowledgement:1 review:1 waterhouse:8 underway:1 loss:1 expect:1 validation:2 switched:2 degree:1 principle:2 supported:2 side:1 normalised:2 institute:1 wide:1 taking:2 distributed:4 world:1 autoregressive:1 forward:1 adaptive:1 transaction:2 approximate:1 fip:1 overcomes:1 corpus:2 davidson:2 spectrum:1 table:4 additionally:1 mj:7 nature:1 tel:1 quantize:1 noise:7 nmse:4 sunspot:5 fashion:1 tong:4 perceptual:1 ix:1 jt:1 gating:7 pz:1 specialisation:1 sit:1 workshop:1 quantization:4 magnitude:1 easier:1 ijj:1 scalar:1 cyclical:1 lewis:1 conditional:3 viewed:1 srw1001:1 appreciable:1 hard:1 specifically:2 huberman:2 total:1 e:1 perceptrons:1 bps:2 select:1