Unnamed: 0
int64
0
7.24k
id
int64
1
7.28k
raw_text
stringlengths
9
124k
vw_text
stringlengths
12
15k
700
1,640
Approximate inference algorithms for two-layer Bayesian networks AndrewY. Ng Computer Science Division UC Berkeley Berkeley, CA 94720 ang@cs.berkeley.edu Michael I. Jordan Computer Science Division and Department of Statistics UC Berkeley Berkeley, CA 94720 jordan@cs.berkeley.edu Abstract We present a class of approximate inference algorithms for graphical models of the QMR-DT type. We give convergence rates for these algorithms and for the Jaakkola and Jordan (1999) algorithm, and verify these theoretical predictions empirically. We also present empirical results on the difficult QMR-DT network problem, obtaining performance of the new algorithms roughly comparable to the Jaakkola and Jordan algorithm. 1 Introduction The graphical models formalism provides an appealing framework for the design and analysis of network-based learning and inference systems. The formalism endows graphs with a joint probability distribution and interprets most queries of interest as marginal or conditional probabilities under this joint. For a fixed model one is generally interested in the conditional probability of an output given an input (for prediction), or an input conditional on the output (for diagnosis or control). During learning the focus is usually on the likelihood (a marginal probability), on the conditional probability of unobserved nodes given observed nodes (e.g., for an EM or gradient-based algorithm), or on the conditional probability of the parameters given the observed data (in a Bayesian setting). In all of these cases the key computational operation is that of marginalization. There are several methods available for computing marginal probabilities in graphical models, most of which involve some form of message-passing on the graph. Exact methods, while viable in many interesting cases (involving sparse graphs), are infeasible in the dense graphs that we consider in the current paper. A number of approximation methods have evolved to treat such cases; these include search-based methods, loopy propagation, stochastic sampling, and variational methods. Variational methods, the focus of the current paper, have been applied successfully to a number of large-scale inference problems. In particular, Jaakkola and Jordan (1999) developed a variational inference method for the QMR-DT network, a benchmark network involving over 4,000 nodes (see below). The variational method provided accurate approximation to posterior probabilities within a second of computer time. For this difficult A. Y. Ng and M. 1. Jordan 534 inference problem exact methods are entirely infeasible (see below), loopy propagation does not converge to correct posteriors (Murphy, Weiss, & Jordan, 1999), and stochastic sampling methods are slow and unreliable (Jaakkola & Jordan, 1999). A significant step forward in the understanding of variational inference was made by Kearns and Saul (1998), who used large deviation techniques to analyze the convergence rate of a simplified variational inference algorithm. Imposing conditions on the magnitude of the weights in the network, they established a 0 ( Jlog N / N) rate of convergence for the error of their algorithm, where N is the fan-in. In the current paper we utilize techniques similar to those of Kearns and Saul to derive a new set of variational inference algorithms with rates that are faster than 0 ( Jlog N / N). Our techniques also allow us to analyze the convergence rate of the Jaakkola and Jordan (1999) algorithm. We test these algorithms on an idealized problem and verify that our analysis correctly predicts their rates of convergence. We then apply these algorithms to the difficult the QMR-DT network problem. 2 Background 2.1 The QMR-DT network The QMR-DT (Quick Medical Reference, Decision-Theoretic) network is a bipartite graph with approximately 600 top-level nodes di representing diseases and approximately 4000 lower-level nodes Ii representing findings (observed symptoms). All nodes are binaryvalued. Each disease is given a prior probability P(d i = 1), obtained from archival data, and each finding is parameterized as a "noisy-OR" model: P(h = lid) = 1- e-(lio-L:jE"i (lijd j , where 7T'i is the set of parent diseases for finding h and where the parameters obtained from assessments by medical experts (see Shwe, et aI., 1991). Letting Zi = OiQ Oij are + I:jE 1l'i Oijdj , we have the following expression for the likelihood I : (1) where the sum is a sum across the approximately 2600 configurations of the diseases. Note that the second product, a product over the negative findings, factorizes across the diseases dj ; these factors can be absorbed into the priors P (dj ) and have no significant effect on the complexity of inference. It is the positive findings which couple the diseases and prevent the sum from being distributed across the product. Generic exact algorithms such as the junction tree algorithm scale exponentially in the size of the maximal clique in a moralized, triangulated graph. Jaakkola and Jordan (1999) found cliques of more than 150 nodes in QMR-DT; this rules out the junction tree algorithm. Heckerman (1989) discovered a factorization specific to QMR-DT that reduces the complexity substantially; however the resulting algorithm still scales exponentially in the number of positive findings and is only feasible for a small subset of the benchmark cases. I In this expression, the factors P( dj) are the probabilities associated with the (parent-less) disease nodes, the factors (1 - e - Zi) are the probabilities of the (child) finding nodes that are observed to be in their positive state, and the factors e -Zi are the probabilities of the negative findings. The resulting product is the joint probability P(f, d), which is marginalized to obtain the likelihood P(f). Approximate Inference Algorithms for Two-Layer Bayesian Networks 2.2 535 The Jaakkola and Jordan (JJ) algorithm Jaakkola and Jordan (1999) proposed a variational algorithm for approximate inference in the QMR-DT setting. Briefly, their approach is to make use of the following variational inequality: where Ci is a deterministic function of Ai. This inequality holds for arbitrary values of the free "variational parameter" Ai. Substituting these variational upper bounds for the probabilities of positive findings in Eq. (1), one obtains a factorizable upper bound on the likelihood. Because of the factorizability, the sum across diseases can be distributed across the joint probability, yielding a product of sums rather than a sum of products. One then minimizes the resulting expression with respect to the variational parameters to obtain the tightest possible variational bound. 2.3 The Kearns and Saul (KS) algorithm A simplified variational algorithm was proposed by Kearns and Saul (1998), whose main goal was the theoretical analysis of the rates of convergence for variational algorithms. In their approach, the local conditional probability for the finding Ii is approximated by its value at a point a small distance Ci above or below (depending on whether upper or lower bounds are desired) the mean input E[Zi]. This yields a variational algorithm in which the values Ci are the variational parameters to be optimized. Under the assumption that the weights Oij are bounded in magnitude by T / N, where T is a constant and N is the number of parent ("disease") nodes, Kearns and Saul showed that the error in likelihood for their algorithm converges at a rate of O( Vlog N / N). 3 Algorithms based on local expansions Inspired by Kearns and Saul (1998), we describe the design of approximation algorithms for QMR-DT obtained by expansions around the mean input to the finding nodes. Rather than using point approximations as in the Kearns-Saul (KS) algorithm, we make use of Taylor expansions. (See also Plefka (1982), and Barber and van de Laar (1999) for other perturbational techniques.) Consider a generalized QMR-DT architecture in which the noisy-OR model is replaced by a general function 'IjJ( z) : R -t [0, 1] having uniformly bounded derivatives, i.e., \'IjJ(i) (z) \ :::; B i . Define F(Zl, . .. , ZK) = rr~l ('IjJ(zi))fi rr~l (1 - 'IjJ(Zd)l-fi so that the likelihood can be written as (2) P(f) = E{z;}[F(Zl"" ,ZK)]. Also define /-ti = E[Zi] = Ow + 2:7=1 OijP(dj = 1). A simple mean-field-like approximation can be obtained by evaluating F at the mean values P(f) ~ F(/-t1, ... ,/-tK). We refer to this approximation as "MF(O)." (3) Expanding the function F to second order, and defining (i = Zi - /-ti, we have: K P(f) E{fi} rF(jl) + L L 11 =1 1 K Fi1 (J1)(i 1 + 21 K .L L Fid2 (J1)Eh (i2 + 11 =112=1 (4) A. Y. Ng and M I. Jordan 536 where the subscripts on F represent derivatives. Dropping the remainder term and bringing the expectation inside, we have the "MF(2)" approximation: 1 P(f) ~ F(il) + 2 K K LL Fili2(Jt)E[EilEi2] i 1 =1 i2=1 More generally, we obtain a "MF(i)" approximation by carrying out a Taylor expansion to i-th order. 3.1 Analysis In this section, we give two theorems establishing convergence rates for the MF( i) family of algorithms and for the Jaakkola and Jordan algorithm. As in Kearns and Saul (1998), our results are obtained under the assumption that the weights are of magnitude at most O(lIN) (recall that N is the number of disease nodes). For large N, this assumption of "weak interactions" implies that each Zi will be close to its mean value with high probability (by the law of large numbers), and thereby gives justification to the use of local expansions for the probabilities of the findings. Due to space constraints, the detailed proofs of the theorems given in this section are deferred to the long version of this paper, and we will instead only sketch the intuitions for the proofs here. Theorem 1 Let K (the number offindings) be fixed, and suppose IfJ ij I :::; ~ for all i, j for some fixed constantT. Then the absolute error ofthe MF(k) approximation is 0 (N(!:!1)/2) for k odd and 0 (N(k72+1) ) for k even. Proof intuition. First consider the case of odd k. Since IfJ ij I :::; ~, the quantity Ei = Zi J-Li = 2: j fJ ij (dj - E[ dj ]) is like an average of N random variables, and hence has standard deviation on the order 11m. Since MF(k) matches F up to the k-th order derivatives, we find that when we take a Taylor expansion ofMF(k)'s error, the leading non-zero term is the k + 1-st order term, which contains quantities such as 10:+1. Now because Ei has standard deviation on the order 11m, it is unsurprising that E[E:+l] is on the order 1IN(k+l)/2, which gives the error of MF(k) for odd k. For k even, the leading non-zero term in the Taylor expansion of the error is a k + 1-st order term with quantities such as 10:+1. But if we think of Ei as converging (via a central limit theorem effect) to a symmetric distribution, then since symmetric distributions have small odd central moments, E[ 10:+1] would be small. This means that for k even, we may look to the order k + 2 term for the error, which leads to MF(k) having the the same big-O error as MF(k + 1). Note this is also consistent with how MF(O) and MF(l) always give the same estimates and hence have the same absolute error. 0 A theorem may also be proved for the convergence rate of the Jaakkola and Jordan (JJ) algorithm. For simplicity, we state it here only for noisy-OR networks. 2 A closely related result also holds for sigmoid networks with suitably modified assumptions; see the full paper. Theorem 2 Let K befixed, and suppose 'Ij;(z) = 1-e- z is the noisy-ORfunction. Suppose further that 0 :::; fJ ij :::; ~ for all i, j for some fixed constant T, and that J-Li ~ J-Lmin for all i, for some fixed J-Lmin > O. Then the absolute error of the JJ approximation is 0 (~ ). 2Note in any case that 11 can be applied only when 1/J is log-concave, such as in noisy-OR networks (where incidentally all weights are non-negative). 537 Approximate Inference Algorithms for Two-Layer Bayesian Networks The condition of some Pmin lowerbounding the Pi'S ensures that the findings are not too unlikely; for it to hold, it is sufficient that there be bias ("leak") nodes in the network with weights bounded away from zero. Proof intuition. Neglecting negative findings, (which as discussed do not need to be handled variationally,) this result is proved for a "simplified" version of the JJ algorithm, that always chooses the variational parameters so that for each i, the exponential upperbound on '1/J(Zi) is tangent to '1/J at Zi = Pi. (The "normal" version of JJ can have error no worse than this simplified one.) Taking a Tay lor expansion again of the approximation's error, we find that since the upperbound has matched zeroth and first derivatives with F, the error is As discussed in the MF(k) proof outline, a second order term with quantities such as this quantity has expectation on the order 1jN, and hence JJ's error is O(ljN). 0 f.t. To summarize our results in the most useful cases, we find that MF(O) has a convergence rate of O(ljN) , both MF(2) and MF(3) have rates of O(ljN2) , and JJ has a convergence rate of O(ljN). 4 Simulation results 4.1 Artificial networks We carried out a set of simulations that were intended to verify the theoretical results presented in the previous section. We used bipartite noisy-OR networks, with full connectivity between layers and with the weights ()ij chosen uniformly in (0, 2jN). The number N of top-level ("disease") nodes ranged from 10 to 1000. Priors on the disease nodes were chosen uniformly in (0,1). The results are shown in Figure 1 for one and five positive findings (similar results where obtained for additional positive findings). 100 r--_ _ _ _ _~------___. -..._--_._---... --..... _--... -. 10?' -- - - - - - __ _ 10?' ....?. ...... .... . .. . ............... ... .. .... ... e w10' 1t th 10 100 #diseases 1000 10 100 1000 #diseases Figure 1: Absolute error in likelihood (averaged over many randomly generated networks) as a function of the number of disease nodes for various algorithms. The short-dashed lines are the KS upper and lower bounds (these curves overlap in the left panel), the long-dashed line is the 11 algorithm and the solid lines are MF(O), MF(2) and MF(3) (the latter two curves overlap in the right panel). The results are entirely consistent with the theoretical analysis, showing nearly exactly the expected slopes of -112, -1 and -2 on a loglog plot. 3 Moreover, the asymptotic results are 3The anomalous behavior of the KS lower bound in the second panel is due to the fact that the algorithm generally finds a vacuous lower bound of 0 in this case, which yields an error which is essentially constant as a function of the number of diseases. A. Y. Ng and M. 1. Jordan 538 also predictive of overall performance: the MF(2) and MF(3) algorithms perform best in all cases, MF(O) and JJ are roughly equivalent, and KS is the least accurate. 4.2 QMR-DT network We now present results for the QMR-DT network, in particular for the four benchmark CPC cases studied by Jaakkola and Jordan (1999). These cases all have fewer than 20 positive findings; thus it is possible to run the Heckerman (1989) "Quickscore" algorithm to obtain the true likelihood. Case 16 Case 32 '0' 10' ''' ,, '0? 10?:10 ,, , '0?' 10'r. '0 ,, 8'?? .c 8 ~t O?30 ,, ~to'" Qj ~ 1 0~ 10" ? '0- :---- to'" '0- ? 3 . , , 10? ?1 7 0 #Exactly treated findings 3 ? 5 #Exactly treated findings Figure 2: Results for CPC cases 16 and 32, for different numbers of exactly treated findings. The horizontal line is the true likelihood, the dashed line is J1's estimate, and the lower solid line is MF(3)'s estimate. Case 34 lO" ~-~--...___---,.---...,.--~------, Case 46 lO' ? .--~--...___---,.---...,.--~--__, ---10?"'F - - - - - - - - - - --------- ----:;;",.-""==;J lO'"F - - - - -_ _ _ _ _-=",...-=======j 1O~ ,0 ?o~-----!---~.- - 7,'--'-7'---:':lO:------!12 #Exacdy treated findings to...o~-____!_--~.- - 7 ' . - - - ! - ' - --::-----!12' #Exaclly treated findings Figure 3: Results for CPC cases 34 and 46. Same legend as above. In Jaakkola and Jordan (1999), a hybrid methodology was proposed in which only a portion of the findings were treated approximately ; exact methods were used to treat the remaining findings . Using this hybrid methodology, Figures 2 and 3 show the results of running JJ and MF(3) on these four cases. 4 4These experiments were run using a version of the 11 algorithm that optimizes the variational parameters just once without any findings treated exactly, and then uses these fixed values of the parameters thereafter. The order in which findings are chosen to be treated exactly is based on 11's estimates, as described in Jaakkola and Jordan (1999). Missing points in the graphs for cases 16 and Approximate Inference Algorithms for Two-Layer Bayesian Networks 539 The results show the MF algorithm yielding results that are comparable with the JJ algorithm. 5 Conclusions and extension to multilayer networks This paper has presented a class of approximate inference algorithms for graphical models of the QMR-DT type, supplied a theoretical analysis of convergence rates, verified the rates empirically, and presented promising empirical results for the difficult QMR-DT problem. Although the focus of this paper has been two-layer networks, the MF(k) family of algorithms can also be extended to multilayer networks. For example, consider a 3-layer network with nodes bi being parents of nodes d i being parents of nodes Ii. To approximate Pr[J] using (say) MF(2), we first write Pr[J] as an expectation of a function (F) of the Zi'S, and approximate this function via a second-order Taylor expansion. To calculate the expectation of the Taylor approximation, we need to calculate terms in the expansion such as E[d i ], E[didj ] and E[dn When di had no parents, these quantities were easily derived in terms of the disease prior probabilities. Now, they instead depend on the joint distribution of d i and d j , which we use our two-layer version of MF(k), applied to the first two (b i and di ) layers of the network, to approximate. It is important future work to carefully study the performance of this algorithm in the multilayer setting. Acknowledgments We wish to acknowledge the helpful advice of Tommi Jaakkola, Michael Kearns, Kevin Murphy, and Larry Saul. References [1] Barber, D., & van de Laar, P. (1999) Variational cumulant expansions for intractable distributions. Journal of Artificial Intelligence Research, 10,435-455. [2] Heckerman, D. (1989). A tractable inference algorithm for diagnosing multiple diseases. In Proceedings of the Fifth Conference on Uncertainty in Artificial Intelligence. [3] Jaakkola, T. S., & Jordan, M. 1. (1999). Variational probabilistic inference and the QMR-DT network. Journal of Artificial Intelligence Research, 10,291-322. [4] Jordan, M . 1., Ghahramani, Z., Jaakkola, T. S., & Saul, L. K. (1998). An introduction to variational methods for graphical models. In Learning in Graphical Models. Cambridge: MIT Press. [5] Kearns, M. 1., & Saul, L. K. (1998). Large deviation methods for approximate probabilistic inference, with rates of convergence. In G. F. Cooper & S. Moral (Eds.), Proceedings of the Fourteenth Conference on Uncertainty in Artificial Intelligence . San Mateo, CA: Morgan Kaufmann. [6] Murphy, K. P., Weiss, Y, & Jordan, M. 1. (1999). Loopy belief propagation for approximate inference: An empirical study. In Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence. [7] Plefka, T. (1982). Convergence condition of the TAP equation for the infinite-ranged Ising spin glass model. In 1. Phys. A: Math. Gen., 15(6). [8] Shwe, M., Middleton, B., Heckerman, D., Henrion, M., Horvitz, E., Lehmann, H., & Cooper, G. (1991). Probabilistic diagnosis using a reformulation of the INTERNIST-1/QMR knowledge base 1. The probabilistic model and inference algorithms. Methods of Information in Medicine, 30, 241-255. 34 correspond to runs where our implementation of the Quickscore algorithm encountered numerical problems.
1640 |@word version:5 briefly:1 suitably:1 simulation:2 thereby:1 solid:2 moment:1 configuration:1 contains:1 horvitz:1 current:3 written:1 numerical:1 j1:3 plot:1 intelligence:5 fewer:1 short:1 provides:1 math:1 node:19 lor:1 five:1 diagnosing:1 dn:1 viable:1 inside:1 expected:1 behavior:1 roughly:2 inspired:1 provided:1 bounded:3 matched:1 panel:3 moreover:1 evolved:1 substantially:1 minimizes:1 developed:1 unobserved:1 finding:26 berkeley:6 ti:2 concave:1 exactly:6 internist:1 control:1 zl:2 medical:2 positive:7 t1:1 local:3 treat:2 limit:1 qmr:17 establishing:1 subscript:1 approximately:4 zeroth:1 k:5 studied:1 mateo:1 factorization:1 bi:1 averaged:1 acknowledgment:1 perturbational:1 empirical:3 binaryvalued:1 close:1 lmin:2 equivalent:1 middleton:1 deterministic:1 quick:1 missing:1 simplicity:1 rule:1 justification:1 suppose:3 exact:4 us:1 approximated:1 predicts:1 ising:1 observed:4 calculate:2 ensures:1 disease:18 intuition:3 leak:1 complexity:2 carrying:1 depend:1 predictive:1 division:2 bipartite:2 easily:1 joint:5 various:1 describe:1 query:1 artificial:6 kevin:1 whose:1 say:1 statistic:1 think:1 noisy:6 rr:2 interaction:1 product:6 maximal:1 remainder:1 gen:1 archival:1 convergence:13 parent:6 incidentally:1 converges:1 tk:1 derive:1 depending:1 ij:6 odd:4 eq:1 c:2 implies:1 triangulated:1 tommi:1 closely:1 correct:1 stochastic:2 cpc:3 larry:1 extension:1 hold:3 around:1 normal:1 substituting:1 successfully:1 mit:1 always:2 modified:1 rather:2 factorizes:1 jaakkola:16 derived:1 focus:3 likelihood:9 glass:1 helpful:1 inference:20 lowerbounding:1 unlikely:1 interested:1 overall:1 uc:2 marginal:3 field:1 once:1 having:2 ng:4 sampling:2 look:1 nearly:1 future:1 randomly:1 murphy:3 replaced:1 intended:1 interest:1 message:1 deferred:1 yielding:2 accurate:2 neglecting:1 tree:2 taylor:6 desired:1 theoretical:5 formalism:2 loopy:3 deviation:4 subset:1 plefka:2 too:1 unsurprising:1 chooses:1 st:2 probabilistic:4 michael:2 connectivity:1 again:1 central:2 worse:1 expert:1 derivative:4 leading:2 pmin:1 li:2 upperbound:2 de:2 idealized:1 analyze:2 portion:1 slope:1 il:1 spin:1 kaufmann:1 who:1 yield:2 ofthe:1 correspond:1 weak:1 bayesian:5 phys:1 ed:1 associated:1 di:3 proof:5 couple:1 proved:2 recall:1 knowledge:1 variationally:1 carefully:1 factorizability:1 dt:16 methodology:2 wei:2 laar:2 symptom:1 just:1 sketch:1 horizontal:1 ei:3 assessment:1 propagation:3 effect:2 verify:3 ranged:2 true:2 hence:3 symmetric:2 i2:2 ll:1 during:1 generalized:1 outline:1 theoretic:1 fj:2 variational:22 fi:3 sigmoid:1 empirically:2 exponentially:2 jl:1 discussed:2 significant:2 refer:1 cambridge:1 imposing:1 ai:3 quickscore:2 dj:6 had:1 base:1 posterior:2 showed:1 optimizes:1 inequality:2 morgan:1 additional:1 converge:1 dashed:3 ii:3 full:2 multiple:1 reduces:1 match:1 faster:1 long:2 lin:1 prediction:2 involving:2 converging:1 anomalous:1 multilayer:3 essentially:1 expectation:4 fifteenth:1 represent:1 background:1 bringing:1 legend:1 jordan:22 marginalization:1 zi:12 architecture:1 interprets:1 qj:1 whether:1 expression:3 handled:1 moral:1 passing:1 jj:10 useful:1 generally:3 detailed:1 involve:1 ang:1 supplied:1 correctly:1 diagnosis:2 zd:1 write:1 dropping:1 key:1 four:2 thereafter:1 reformulation:1 ifj:2 prevent:1 verified:1 utilize:1 graph:7 sum:6 run:3 fourteenth:1 parameterized:1 uncertainty:3 lehmann:1 family:2 decision:1 comparable:2 entirely:2 bound:7 layer:9 fan:1 encountered:1 constraint:1 department:1 across:5 heckerman:4 em:1 appealing:1 lid:1 pr:2 equation:1 letting:1 tractable:1 fi1:1 available:1 tightest:1 operation:1 junction:2 apply:1 away:1 generic:1 jn:2 top:2 remaining:1 include:1 running:1 graphical:6 marginalized:1 jlog:2 medicine:1 ghahramani:1 quantity:6 gradient:1 ow:1 distance:1 lio:1 barber:2 difficult:4 negative:4 design:2 implementation:1 perform:1 upper:4 benchmark:3 acknowledge:1 defining:1 extended:1 discovered:1 arbitrary:1 vacuous:1 optimized:1 tap:1 established:1 usually:1 below:3 summarize:1 rf:1 belief:1 overlap:2 treated:8 eh:1 endows:1 oij:2 hybrid:2 representing:2 carried:1 prior:4 understanding:1 tangent:1 asymptotic:1 law:1 interesting:1 sufficient:1 consistent:2 pi:2 lo:4 free:1 infeasible:2 bias:1 allow:1 saul:11 taking:1 absolute:4 sparse:1 fifth:1 distributed:2 van:2 curve:2 evaluating:1 forward:1 made:1 san:1 simplified:4 approximate:12 obtains:1 unreliable:1 clique:2 search:1 promising:1 zk:2 ca:3 expanding:1 obtaining:1 expansion:11 factorizable:1 dense:1 main:1 big:1 shwe:2 child:1 advice:1 je:2 slow:1 cooper:2 wish:1 exponential:1 loglog:1 theorem:6 moralized:1 specific:1 jt:1 showing:1 intractable:1 ci:3 magnitude:3 ljn:3 mf:27 ijj:4 absorbed:1 w10:1 andrewy:1 conditional:6 goal:1 feasible:1 henrion:1 infinite:1 tay:1 uniformly:3 kearns:10 latter:1 cumulant:1
701
1,641
Correctness of belief propagation in Gaussian graphical models of arbitrary topology Yair Weiss Computer Science Division UC Berkeley, 485 Soda Hall Berkeley, CA 94720-1776 Phone: 510-642-5029 William T. Freeman Mitsubishi Electric Research Lab 201 Broadway Cambridge, MA 02139 Phone: 617-621-7527 yweiss@cs.berkeley.edu freeman @merl.com Abstract Local "belief propagation" rules of the sort proposed by Pearl [15] are guaranteed to converge to the correct posterior probabilities in singly connected graphical models. Recently, a number of researchers have empirically demonstrated good performance of "loopy belief propagation"using these same rules on graphs with loops. Perhaps the most dramatic instance is the near Shannon-limit performance of "Turbo codes", whose decoding algorithm is equivalent to loopy belief propagation. Except for the case of graphs with a single loop, there has been little theoretical understanding of the performance of loopy propagation. Here we analyze belief propagation in networks with arbitrary topologies when the nodes in the graph describe jointly Gaussian random variables. We give an analytical formula relating the true posterior probabilities with those calculated using loopy propagation. We give sufficient conditions for convergence and show that when belief propagation converges it gives the correct posterior means for all graph topologies, not just networks with a single loop. The related "max-product" belief propagation algorithm finds the maximum posterior probability estimate for singly connected networks. We show that, even for non-Gaussian probability distributions, the convergence points of the max-product algorithm in loopy networks are maxima over a particular large local neighborhood of the posterior probability. These results help clarify the empirical performance results and motivate using the powerful belief propagation algorithm in a broader class of networks. Problems involving probabilistic belief propagation arise in a wide variety of applications, including error correcting codes, speech recognition and medical diagnosis. If the graph is singly connected, there exist local message-passing schemes to calculate the posterior probability of an unobserved variable given the observed variables. Pearl [15] derived such a scheme for singly connected Bayesian networks and showed that this "belief propagation" algorithm is guaranteed to converge to the correct posterior probabilities (or "beliefs"). Several groups have recently reported excellent experimental results by running algorithms 674 Y. Weiss and W T. Freeman equivalent to Pearl's algorithm on networks with loops [8, 13, 6]. Perhaps the most dramatic instance of this performance is for "Turbo code" [2] error correcting codes. These codes have been described as "the most exciting and potentially important development in coding theory in many years" [12] and have recently been shown [10, 11] to utilize an algorithm equivalent to belief propagation in a network with loops. Progress in the analysis of loopy belief propagation has been made for the case of networks with a single loop [17, 18, 4, 1] . For these networks, it can be shown that (1) unless all the compatabilities are deterministic, loopy belief propagation will converge. (2) The difference between the loopy beliefs and the true beliefs is related to the convergence rate of the messages - the faster the convergence the more exact the approximation and (3) If the hidden nodes are binary, then the loopy beliefs and the true beliefs are both maximized by the same assignments, although the confidence in that assignment is wrong for the loopy beliefs. In this paper we analyze belief propagation in graphs of arbitrary topology, for nodes describing jointly Gaussian random variables. We give an exact formula relating the correct marginal posterior probabilities with the ones calculated using loopy belief propagation. We show that if belief propagation converges, then it will give the correct posterior means for all graph topologies, not just networks with a single loop. We show that the covariance estimates will generally be incorrect but present a relationship between the error in the covariance estimates and the convergence speed. For Gaussian or non-Gaussian variables, we show that the "max-product" algorithm, which calculates the MAP estimate in singly connected networks, only converges to points that are maxima over a particular large neighborhood of the posterior probability of loopy networks. 1 Analysis To simplify the notation, we assume the graphical model has been preprocessed into an undirected graphical model with pairwise potentials. Any graphical model can be converted into this form, and running belief propagation on the pairwise graph is equivalent to running belief propagation on the original graph [18]. We assume each node X i has a local observation Yi . In each iteration of belief propagation, each node X i sends a message to each neighboring X j that is based on the messages it received from the other neighbors, its local observation Yl and the pairwise potentials Wij(Xi , Xj) and Wii(Xi, Yi) . We assume the message-passing occurs in parallel. The idea behind the analysis is to build an unwrapped tree. The unwrapped tree is the graphical model which belief propagation is solving exactly when one applies the belief propagation rules in a loopy network [9, 20, 18]. It is constructed by maintaining the same local neighborhood structure as the loopy network but nodes are replicated so there are no loops. The potentials and the observations are replicated from the loopy graph. Figure 1 (a) shows an unwrapped tree for the diamond shaped graph in (b). By construction, the belief at the root node X- I is identical to that at node Xl in the loopy graph after four iterations of belief propagation. Each node has a shaded observed node attached to it, omitted here for clarity. Because the original network represents jointly Gaussian variables, so will the unwrapped tree. Since it is a tree, belief propagation is guaranteed to give the correct answer for the unwrapped graph. We can thus use Gaussian marginalization formulae to calculate the true mean and variances in both the original and the unwrapped networks. In this way, we calculate the accuracy of belief propagation for Gaussian networks of arbitrary topology. We assume that the joint mean is zero (the means can be added-in later). The joint distri- Correctness of Belief Propagation 675 Figure 1: Left: A Markov network with mUltiple loops. Right: The unwrapped network corresponding to this structure. bution of z =(: ) is given by P(z) = ae-!zTVz, where V = (~:: ~::) . It is straightforward to construct the inverse covariance matrix V of the joint Gaussian that describes a given Gaussian graphical model [3]. Writing out the exponent of the joint and completing the square shows that the mean I-' of x, given the observations y, is given by: (1) and the covariance matrix C~IY of x given y is: C~IY = V~-;l. We will denote by C~dY the ith row of C~IY so the marginal posterior variance of Xi given the data is (72 (i) = C~i Iy (i). We will use - for unwrapped quantities. We scan the tree in breadth first order and denote by x the vector of values in the hidden nodes of the tree when so scanned. Simlarly, we denote by y the observed nodes scanned in the same order and Vn , V~y the inverse covariance matrices. Since we are scanning in breadth first order the last nodes are the leaf nodes and we denote by L the number of leaf nodes. By the nature of unwrapping, tL(1) is the mean of the belief at node Xl after t iterations of belief propagation, where t is the number of unwrappings. Similarly 0- 2 (1) = 6~1Iy(1) is the variance of the belief at node Xl after t iterations. Because the data is replicated we can write y = Oy where O(i, j) = 1 if Yi is a replica of Yj and 0 otherwise. Since the potentials W(Xi' Yi) are replicated, we can write V~yO = OV~y. Since the W(Xi, X j) are also replicated and all non-leaf Xi have the same connectivity as the corresponding Xi, we can write V~~O = OVzz + E where E is zero in all but the last L rows. When these relationships between the loopy and unwrapped inverse covariance matrices are substituted into the loopy and unwrapped versions of equation I, one obtains the following expression, true for any iteration [19]: (2) where e is a vector that is zero everywhere but the last L components (corresponding to the leaf nodes). Our choice of the node for the root of the tree is arbitrary, so this applies to all nodes of the loopy network. This formula relates, for any node of a network with loops, the means calculated at each iteration by belief propagation with the true posterior means. Similarly when the relationship between the loopy and unwrapped inverse covariance matrices is substituted into the loopy and unwrapped definitions of C~IY we can relate the Y Weiss and W T Freeman 676 0.5 0.4 ~ 0.3 ~ .~ 0.2 n; ~ 0.1 8 "t:> ? 0 -0.1 -0.2 20 0 40 60 80 100 node Figure 2: The conditional correlation between the root node and all other nodes in the unwrapped tree of Fig. 1 after eight iterations. Potentials were chosen randomly. Nodes are presented in breadth first order so the last elements are the correlations between the root node and the leaf nodes. We show that if this correlation goes to zero, belief propagation converges and the loopy means are exact. Symbols plotted with a star denote correlations with nodes that correspond to the node Xl in the loopy graph. The sum of these correlations gives the correct variance of node Xl while loopy propagation uses only the first correlation. marginalized covariances calculated by belief propagation to the true ones [19]: -2 a (1) = a 2 (1) + CZllyel - - Czt/ye2 (3) where el is a vector that is zero everywhere but the last L components while e2 is equal to 1 for all nodes in the unwrapped tree that are replicas of Xl except for Xl. All other components of e2 are zero, Figure 2 shows Cz1l Y for the diamond network in Fig. 1. We generated random potential functions and observations and calculated the conditional correlations in the unwrapped tree. Note that the conditional correlation decreases with distance in the tree - we are scanning in breadth first order so the last L components correspond to the leaf nodes. As the number of iterations of loopy propagation is increased the size of the unwrapped tree increases and the conditional correlation between the leaf nodes and the root node decreases. From equations 2-3 it is clear that if the conditional correlation between the leaf nodes and the root nodes are zero for all sufficiently large unwrappings then (1) belief propagation converges (2) the means are exact and (3) the variances may be incorrect. In practice the conditional correlations will not actually be equal to zero for any finite unwrapping. In [19] we give a more precise statement: if the conditional correlation of the root node and the leaf nodes decreases rapidly enough then (1) belief propagation converges (2) the means are exact and (3) the variances may be incorrect. We also show sufficient conditions on the potentials III (Xi, X j) for the correlation to decrease rapidly enough: the rate at which the correlation decreases is determined by the ratio of off-diagonal and diagonal components in the quadratic fonn defining the potentials [19]. How wrong will the variances be? The tenn CZllye2 in equation 3 is simply the sum of many components of Cz11y . Figure 2 shows these components. The correct variance is the sum of all the components witHe the belief propagation variance approximates this sum with the first (and dominant) tenn. Whenever there is a positive correlation between the root node and other replicas of Xl the loopy variance is strictly less than the true variance - the loopy estimate is overconfident. 677 Correctness of Belief Propagation ~07 e iDO.6 ., ";;;05 fr ~04 SOR '"~03 0.2 0.1 20 30 40 50 60 iterations (a) (b) Figure 3: (a) 25 x 25 graphical model for simulation. The unobserved nodes (unfilled) were connected to their four nearest neighbors and to an observation node (filled). (b) The error of the estimates of loopy propagation and successive over-relaxation (SOR) as a function of iteration. Note that belief propagation converges much faster than SOR. Note that when the conditional correlation decreases rapidly to zero two things happen. First, the convergence is faster (because CZdyel approaches zero faster) . Second, the approximation error of the variances is smaller (because CZ1 / y e2 is smaller). Thus we have shown, as in the single loop case, quick convergence is correlated with good approximation. 2 Simulations We ran belief propagation on the 25 x 25 2D grid of Fig. 3 a. The joint probability was: (4) where Wij = 0 if nodes Xi, Xj are not neighbors and 0.01 otherwise and Wii was randomly selected to be 0 or 1 for all i with probability of 1 set to 0.2. The observations Yi were chosen randomly. This problem corresponds to an approximation problem from sparse data where only 20% of the points are visible. We found the exact posterior by solving equation 1. We also ran belief propagation and found that when it converged, the calculated means were identical to the true means up to machine precision. Also, as predicted by the theory, the calculated variances were too small - the belief propagation estimate was overconfident. In many applications, the solution of equation 1 by matrix inversion is intractable and iterative methods are used. Figure 3 compares the error in the means as a function of iterations for loopy propagation and successive-over-relaxation (SOR), considered one of the best relaxation methods [16]. Note that after essentially five iterations loopy propagation gives the right answer while SOR requires many more. As expected by the fast convergence, the approximation error in the variances was quite small. The median error was 0.018. For comparison the true variances ranged from 0.01 to 0.94 with a mean of 0.322. Also, the nodes for which the approximation error was worse were indeed the nodes that converged slower. Y. Weiss and W T Freeman 678 3 Discussion Independently, two other groups have recently analyzed special cases of Gaussian graphical models. Frey [7] analyzed the graphical model corresponding to factor analysis and gave conditions for the existence of a stable fixed-point. Rusmevichientong and Van Roy [14] analyzed a graphical model with the topology of turbo decoding but a Gaussian joint density. For this specific graph they gave sufficient conditions for convergence and showed that the means are exact. Our main interest in the Gaussian case is to understand the performance of belief propagation in general networks with multiple loops. We are struck by the similarity of our results for Gaussians in arbitrary networks and the results for single loops of arbitrary distributions [18]. First, in single loop networks with binary nodes, loopy belief at a node and the true belief at a node are maximized by the same assignment while the confidence in that assignment is incorrect. In Gaussian networks with multiple loops, the mean at each node is correct but the confidence around that mean may be incorrect. Second, for both singleloop and Gaussian networks, fast belief propagation convergence correlates with accurate beliefs. Third, in both Gaussians and discrete valued single loop networks, the statistical dependence between root and leaf nodes governs the convergence rate and accuracy. The two models are quite different. Mean field approximations are exact for Gaussian MRFs while they work poorly in sparsely connected discrete networks with a single loop. The results for the Gaussian and single-loop cases lead us to believe that similar results may hold for a larger class of networks. Can our analysis be extended to non-Gaussian distributions? The basic idea applies to arbitrary graphs and arbitrary potentials: belief propagation is performing exact inference on a tree that has the same local neighbor structure as the loopy graph. However, the linear algebra that we used to calculate exact expressions for the error in belief propagation at any iteration holds only for Gaussian variables. We have used a similar approach to analyze the related "max-product" belief propagation algorithm on arbitrary graphs with arbitrary distributions [5] (both discrete and continuous valued nodes). We show that if the max-product algorithm converges, the max-product assignment has greater posterior probability then any assignment in a particular large region around that assignment. While this is a weaker condition than a global maximum, it is much stronger than a simple local maximum of the posterior probability. The sum-product and max-product belief propagation algorithms are fast and parallelizable. Due to the well known hardness of probabilistic inference in graphical models, belief propagation will obviously not work for arbitrary networks and distributions. Nevertheless, a growing body of empirical evidence shows its success in many networks with loops. Our results justify applying belief propagation in certain networks with mUltiple loops. This may enable fast, approximate probabilistic inference in a range of new applications. References [1] S.M. Aji, G.B. Hom, and R.J. McEliece. On the convergence of iterative decoding on graphs with a single cycle. In Proc. 1998 ISIT, 1998. [2] c. Berrou, A. Glavieux, and P. Thitimajshima. Near Shannon limit error-correcting coding and decoding: Turbo codes. In Proc. IEEE International Communications Conference '93, 1993. [3] R. Cowell. Advanced inference in Bayesian networks. In M.1. Jordan, editor, Learning in Graphical Models . MIT Press, 1998. [4] G.D. Forney, F.R. Kschischang, and B. Marcus. Iterative decoding of tail-biting trellisses. preprint presented at 1998 Information Theory Workshop in San Diego, 1998. Correctness of Belief Propagation 679 [5] W. T. Freeman and Y. Weiss. On the fixed points of the max-product algorithm. Technical Report 99-39, MERL, 201 Broadway, Cambridge, MA 02139, 1999. [6] W.T. Freeman and E.C. Pasztor. Learning to estimate scenes from images. In M.S. Kearns, S.A. SoUa, and D.A. Cohn, editors, Adv. Neural Information Processing Systems I I. MIT Press, 1999. [7] B.J. Frey. Turbo factor analysis. In Adv. Neural Information Processing Systems 12. 2000. to appear. [8) Brendan J. Frey. Bayesian Networksfor Pattern Classification, Data Compression and Channel Coding. MIT Press, 1998. [9) R.G . Gallager. Low Density Parity Check Codes. MIT Press, 1963. [10) F. R. Kschischang and B. J. Frey. Iterative decoding of compound codes by probability propagation in graphical models. IEEE Journal on Selected Areas in Communication , 16(2):219-230, 1998. [11] R.J. McEliece, D.J .C. MackKay, and J.F. Cheng. Turbo decoding as as an instance of Pearl's 'belief propagation' algorithm. IEEE Journal on Selected Areas in Communication, 16(2): 140152,1998. [12J R.J. McEliece, E. Rodemich, and J.F. Cheng. The Turbo decision algorithm. In Proc. 33rd Allerton Conference on Communications, Control and Computing, pages 366-379, Monticello, IL, 1995. [I3J K.P. Murphy, Y. Weiss, and M.1. Jordan. Loopy belief propagation for approximate inference : an empirical study. In Proceedings of Uncertainty in AI, 1999. [14] Rusmevichientong P. and Van Roy B. An analysis of Turbo decoding with Gaussian densities. In Adv. Neural Information Processing Systems I2 . 2000. to appear. [15) Judea Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, 1988. [16J Gilbert Strang. Introduction to Applied Mathel1Ultics. Wellesley-Cambridge, 1986. [I7J Y. Weiss. Belief propagation and revision in networks with loops. Technical Report 1616, MIT AI lab, 1997. [18J Y. Weiss. Correctness of local probability propagation in graphical models with loops. Neural Computation, to appear, 2000. [19] Y. Weiss and W. T. Freeman. Loopy propagation gives the correct posterior means for Gaussians. Technical Report UCB.CSD-99-1046, Berkeley Computer Science Dept., 1999. www.cs.berkeley.edu yweiss/. [20J N. Wiberg. Codes and decoding on general graphs. PhD thesis, Department of Electrical Engineering, U. Linkoping, Sweden, 1996.
1641 |@word version:1 inversion:1 compression:1 stronger:1 simulation:2 mitsubishi:1 covariance:8 fonn:1 dramatic:2 unwrappings:2 com:1 visible:1 happen:1 tenn:2 leaf:10 selected:3 czt:1 ith:1 node:50 successive:2 allerton:1 five:1 constructed:1 incorrect:5 pairwise:3 indeed:1 hardness:1 expected:1 growing:1 freeman:8 little:1 unwrapped:16 ye2:1 distri:1 revision:1 notation:1 unobserved:2 berkeley:5 exactly:1 wrong:2 control:1 medical:1 appear:3 positive:1 engineering:1 local:9 frey:4 limit:2 shaded:1 range:1 yj:1 practice:1 aji:1 area:2 empirical:3 confidence:3 applying:1 writing:1 gilbert:1 equivalent:4 deterministic:1 demonstrated:1 map:1 quick:1 www:1 straightforward:1 go:1 independently:1 correcting:3 rule:3 construction:1 diego:1 exact:10 us:1 element:1 roy:2 recognition:1 sparsely:1 observed:3 preprint:1 electrical:1 calculate:4 region:1 connected:7 cycle:1 adv:3 decrease:6 ran:2 motivate:1 ov:1 solving:2 algebra:1 division:1 joint:6 fast:4 describe:1 neighborhood:3 whose:1 quite:2 larger:1 valued:2 plausible:1 otherwise:2 jointly:3 obviously:1 analytical:1 product:9 fr:1 biting:1 neighboring:1 loop:22 rapidly:3 poorly:1 linkoping:1 convergence:12 converges:8 help:1 nearest:1 received:1 progress:1 c:2 predicted:1 correct:10 enable:1 sor:5 yweiss:2 isit:1 strictly:1 clarify:1 hold:2 sufficiently:1 hall:1 considered:1 around:2 omitted:1 proc:3 correctness:5 wiberg:1 mit:5 gaussian:21 broader:1 derived:1 yo:1 check:1 brendan:1 inference:6 mrfs:1 el:1 hidden:2 wij:2 classification:1 exponent:1 development:1 special:1 uc:1 marginal:2 equal:2 construct:1 field:1 shaped:1 identical:2 represents:1 report:3 simplify:1 intelligent:1 randomly:3 murphy:1 william:1 interest:1 message:5 analyzed:3 behind:1 accurate:1 monticello:1 sweden:1 unless:1 tree:14 filled:1 plotted:1 theoretical:1 merl:2 instance:3 increased:1 i3j:1 assignment:7 loopy:34 too:1 reported:1 answer:2 scanning:2 ido:1 density:3 international:1 probabilistic:4 yl:1 off:1 decoding:9 iy:6 connectivity:1 thesis:1 worse:1 withe:1 potential:9 converted:1 star:1 coding:3 rusmevichientong:2 later:1 root:9 wellesley:1 lab:2 analyze:3 bution:1 sort:1 parallel:1 simlarly:1 square:1 il:1 accuracy:2 variance:15 kaufmann:1 maximized:2 correspond:2 bayesian:3 researcher:1 converged:2 parallelizable:1 whenever:1 definition:1 e2:3 judea:1 actually:1 rodemich:1 wei:9 just:2 correlation:16 mceliece:3 cohn:1 propagation:59 perhaps:2 believe:1 ranged:1 true:11 unfilled:1 i2:1 reasoning:1 image:1 recently:4 empirically:1 attached:1 tail:1 approximates:1 relating:2 cambridge:3 ai:2 rd:1 grid:1 similarly:2 stable:1 similarity:1 soua:1 dominant:1 posterior:16 showed:2 phone:2 compound:1 certain:1 binary:2 success:1 yi:5 morgan:1 greater:1 converge:3 berrou:1 relates:1 multiple:4 technical:3 faster:4 calculates:1 involving:1 basic:1 ae:1 essentially:1 iteration:13 median:1 sends:1 undirected:1 thing:1 jordan:2 near:2 iii:1 enough:2 variety:1 xj:2 marginalization:1 gave:2 topology:7 idea:2 expression:2 speech:1 passing:2 generally:1 clear:1 governs:1 singly:5 thitimajshima:1 exist:1 diagnosis:1 write:3 discrete:3 group:2 four:2 nevertheless:1 clarity:1 preprocessed:1 breadth:4 utilize:1 replica:3 graph:20 relaxation:3 sum:5 year:1 inverse:4 everywhere:2 uncertainty:1 powerful:1 soda:1 vn:1 decision:1 networksfor:1 dy:1 forney:1 completing:1 guaranteed:3 hom:1 cheng:2 quadratic:1 turbo:8 scanned:2 scene:1 speed:1 performing:1 department:1 overconfident:2 describes:1 smaller:2 equation:5 describing:1 wii:2 gaussians:3 eight:1 yair:1 slower:1 existence:1 original:3 running:3 graphical:15 maintaining:1 marginalized:1 build:1 added:1 quantity:1 occurs:1 dependence:1 diagonal:2 distance:1 marcus:1 code:9 relationship:3 ratio:1 potentially:1 relate:1 statement:1 broadway:2 diamond:2 observation:7 markov:1 pasztor:1 finite:1 defining:1 extended:1 communication:4 precise:1 arbitrary:12 struck:1 pearl:5 pattern:1 max:8 including:1 belief:61 advanced:1 scheme:2 understanding:1 strang:1 oy:1 sufficient:3 exciting:1 editor:2 unwrapping:2 row:2 last:6 parity:1 weaker:1 understand:1 wide:1 neighbor:4 sparse:1 van:2 calculated:7 made:1 replicated:5 san:1 correlate:1 approximate:2 obtains:1 global:1 xi:9 continuous:1 iterative:4 nature:1 channel:1 ca:1 kschischang:2 excellent:1 electric:1 substituted:2 main:1 csd:1 arise:1 body:1 fig:3 tl:1 precision:1 trellis:1 xl:8 third:1 formula:4 specific:1 symbol:1 evidence:1 intractable:1 workshop:1 phd:1 i7j:1 simply:1 gallager:1 applies:3 cowell:1 corresponds:1 ma:2 conditional:8 glavieux:1 cz1:1 determined:1 except:2 justify:1 kearns:1 experimental:1 shannon:2 ucb:1 scan:1 dept:1 correlated:1
702
1,642
Robust Learning of Chaotic Attractors Rembrandt Bakker* Chemical Reactor Engineering Delft Univ. of Technology r.bakker@stm.tudelft?nl Jaap C. Schouten Marc-Olivier Coppens Chemical Reactor Engineering Chemical Reactor Engineering Eindhoven Univ. of Technology Delft Univ. of Technology J.C.Schouten@tue.nl coppens@stm.tudelft?nl Floris Takens C. Lee Giles Cor M. van den Bleek Dept. Mathematics University of Groningen NEC Research Institute Princeton Nl Chemical Reactor Engineering Delft Univ. of Technology F. Takens@math.rug.nl giles@research.nj.nec.com vdbleek@stm.tudelft?nl Abstract A fundamental problem with the modeling of chaotic time series data is that minimizing short-term prediction errors does not guarantee a match between the reconstructed attractors of model and experiments. We introduce a modeling paradigm that simultaneously learns to short-tenn predict and to locate the outlines of the attractor by a new way of nonlinear principal component analysis. Closed-loop predictions are constrained to stay within these outlines, to prevent divergence from the attractor. Learning is exceptionally fast: parameter estimation for the 1000 sample laser data from the 1991 Santa Fe time series competition took less than a minute on a 166 MHz Pentium PC. 1 Introduction We focus on the following objective: given a set of experimental data and the assumption that it was produced by a deterministic chaotic system, find a set of model equations that will produce a time-series with identical chaotic characteristics, having the same chaotic attractor. The common approach consists oftwo steps: (1) identify a model that makes accurate shorttenn predictions; and (2) generate a long time-series with the model and compare the nonlinear-dynamic characteristics of this time-series with the original, measured time-series. Principe et al. [1] found that in many cases the model can make good short-tenn predictions but does not learn the chaotic attractor. The method would be greatly improved if we could minimize directly the difference between the reconstructed attractors of the model-generated and measured data, instead of minimizing prediction errors. However, we cannot reconstruct the attractor without first having a prediction model. Until now research has focused on how to optimize both step 1 and step 2. For example, it is important to optimize the prediction horizon of the model [2] and to reduce complexity as much as possible. This way it was possible to learn the attractor of the benchmark laser time series data from the 1991 Santa Fe *DelftChemTech, Chemical Reactor Engineering Lab, lulianalaan 136, 2628 BL, Delft, The Netherlands; http://www.cpt.stm.tudelft.nllcptlcre!researchlbakker/. 880 R. Bakker. J. C. Schouten. M.-Q. Coppens. F. Takens. C. L. Giles and C. M. v. d. Bleek time series competition. While training a neural network for this problem, we noticed [3] that the attractor of the model fluctuated from a good match to a complete mismatch from one iteration to another. We were able to circumvent this problem by selecting exactly that model that matches the attractor. However, after carrying out more simulations we found that what we neglected as an unfortunate phenomenon [3] is really a fundamental limitation of current approaches. An important development is the work of Principe et al. [4] who use Kohonen Self Organizing Maps (SOMs) to create a discrete representation of the state space of the system. This creates a partitioning of the input space that becomes an infrastructure for local (linear) model construction. This partitioning enables to verify if the model input is near the original data (i. e. , detect if the model is not extrapolating) without keeping the training data set with the model. We propose a different partitioning of the input space that can be used to (i) learn the outlines of the chaotic attractor by means of a new way of nonlinear Principal Component Analysis (PCA), and (ii) enforce the model never to predict outside these outlines. The nonlinear PCA algorithm is inspired by the work of Kambhatla and Leen [5] on local PCA: they partition the input space and perform local PCA in each region. Unfortunately, this introduces discontinuities between neighboring regions. We resolve them by introducing a hierarchical partitioning algorithm that uses fuzzy boundaries between the regions . This partitioning closely resembles the hierarchical mixtures of experts of Jordan and Jacobs [6]. In Sec. 2 we put forward the fundamental problem that arises when trying to learn a chaotic attractor by creating a short-term prediction model. In Sec. 3 we describe the proposed partitioning algorithm. In Sec. 4 it is outlined how this partitioning can be used to learn the outline of the attractor by defining a potential that measures the distance to the attractor. In Sec. 5 we show modeling results on a toy example, the logistic map, and on a more serious problem, the laser data from the 1991 Santa Fe time series competition. Section 6 concludes. 2 The attractor learning dilemma Imagine an experimental system with a chaotic attractor, and a time-series of noise-free measurements taken from this system. The data is used to fit the parameters of the model =FwC; ' ;_I"" ,Zt-m) whereF is a nonlinear function, wcontains its adjustable parameters and m is a positive constant. What happens if we fit the parameters w by nonlinear least squares regression? Will the model be stable, i.e. , will the closed-loop long term prediction converge to the same attractor as the one represented by the measurements? ;.1 Figure 1 shows the result of a test by Diks et al. [7] that compares the difference between the model and measured attractor. The figure shows that while the neural network is trained to predict chaotic data, the model quickly converges to the measured attractor 20 t-----;r-----r-.....,.1t_:_----.----.-----i-,----i-,----, (S=O), but once in a while, from one 15 .. ...... .. .... ~ ..... .. .. . -,, ..... ----. -. -- ,,... -.. .... . - - iteration to another, the match between , , , , .... ' , , , , the attractors is lost. I : ~ ;;;'10 ............ .. ~ .... . .. ? ? _ _ L __ ? -' ? ? _ ?? ? ? ? _ _ -' _ _ _ _ _ , ??? ? .- -_ .. _-- -..... .. , , , , ? ? ?? ? ?? ? . To understand what causes this . .... _5 _.. ---- -.-----: -_. - .? .,, , , instability, imagine that we try to fit the ~.L ?......... ~I Lo.. : parameters of a model = ii + B Zt 0'-? ?.. ? '. while the real system has a point 8000 attractor, Z= a, where Z is the state of training progress leg iterations] the system and a its attracting value. Figure 1: Diks test monitoring curve for a neural Clearly, measurements taken from this network model trained on data from an system contain no information to experimental chaotic pendulum [3]. ;.1 o Robust Learning of Chaotic Attractors 881 estimate both ii and B. If we fit the model parameters with non-robust linear least squares, B may be assigned any value and if its largest eigenvalue happens to be greater than zero, the model will be unstable! For the linear model this problem has been solved a long time ago with the introduction of singular value decomposition. There still is a need for a nonlinear counterpart of this technique, in particular since we have to work with very flexible models that are designed to fit a wide variety of nonlinear shapes, see for example the early work of Lapedes and Farber [8]. It is already common practice to control the complexity of nonlinear models by pruning or regularization. Unfortunately, these methods do not always solve the attractor learning problem, since there is a good chance that a nonlinear term explains a lot of variance in one part of the state space, while it causes instability of the attractor (without affecting the one-stepahead prediction accuracy) elsewhere. In Secs. 3 and 4 we will introduce a new method for nonlinear principal component analysis that will detect and prevent unstable behavior. 3. The split and fit algorithm The nonlinear regression procedure of this section will form the basis of the nonlinear principal component algorithm in Sec. 4. It consists of (i) a partitioning of the input space, (ii) a local linear model for each region, and (iii) fuzzy boundaries between regions to ensure global smoothness. The partitioning scheme is outlined in Procedure 1: Procedure 1: Partitioning the input space 1) Start with the entire set Z of input data 2) Determine the direction of largest variance of Z: perform a singular value decomposition of Z into the product ULVT and take the eigenvector (column of V) with the largest singular value (on the diagonal of EJ. 3) Split the data in two subsets (to be called: clusters) by creating a plane perpendicular to the direction of largest variance, through the center of gravity of Z. 4) Next, select the cluster with the largest sum squared error to be split next, and recursively apply 2-4 until a stopping criteria is met. Figures 2 and 3 show examples of the partitioning. The disadvantage of dividing regression problems into localized subproblems was pointed out by Jordan and Jacobs [6]: the spread of the data in each region will be much smaller than the spread of the data as a whole, and this will increase the variance of the model parameters. Since we always split perpendicular to the direction of maximum variance, this problem is minimized. The partitioning can be written as a binary tree, with each non-terminal node being a split and each terminal node a cluster. Procedure 2 creates fuzzy boundaries between the clusters. Procedure 2. Creating fuzzy boundaries 1) An input i enters the tree at the top of the partitioning tree. 2) The Euclidean distance to the splitting hyperplane is divided by the bandwidth f3 of the split, and passed through a sigmoidal function with range [0,1]. This results in i's share 0 in the subset on z's side of the splitting plane. The share in the other subset is I-a. 3) The previous step is carried out for all non-terminal nodes of the tree. 882 R. Bakker. J. C. Schouten, M.-Q. Coppens, F. Takens, C. L. Giles and C. M. v. d. Bleek z 4) The membership Pc of to subset (terminal node) c is computed by taking the product of all previously computed shares 0 along the path from the terminal node to the top of the tree. If we would make all parameters adjustable, that is (i) the orientation of the splitting hyperplanes, (ii) the bandwidths f3, and (iii) the local linear model parameters, the above model structure would be identical to the hierarchical mixtures of experts of Jordan and Jacobs [6]. However, we already fixed the hyperplanes and use Procedure 3 to compute the bandwidths: Procedure 3. Computing the Bandwidths 1) The bandwidths of the terminal nodes are taken to be a constant (we use 1.65, the 90% confidence limit of a normal distribution) times the variance of the subset before it was last split, in the direction of the eigenvector of that last split. 2) The other bandwidths do depend on the input z. They are computed by climbing upward in the tree. The bandwidth of node n is computed as a weighted sum between the fJs of its right and left child, by the implicit formula Pn=OL PL uRPR' in which uLand OR depend on Pn? Starting from initial guess Pn=PL if oL>O?5, or else Pn=PR' the formula is solved in a few iterations . This procedure is designed to create large overlap between neighboring regions and almost no overlap between non-neighboring regions. What remains to be fitted is the set of the local linear models. The j-th output of the split&fit model for a given input zp is computed: c Yj,p =L fl; {ii;zp +b/}. where iicand b contain the linear model parameters of subset c, C c=J and C is the number of clusters. We can determine the parameters of all local linear models in one global fit that is linear in the parameters. However, we prefer to locally optimize the parameters for two reasons: (i) it makes it possible to locally control the stability of the attractor and do the principal component analysis of Sec. 4; and (ii) the computing time for a linear regression problem with r regressors scales -O(~). If we would adopt global fitting, r would scale linearly with C and, while growing the model, the regression problem would quickly become intractable. We use the following iterative local fitting procedure instead. Procedure 4. Iterative Local Fitting 1) Initialize a J by N matrix of residuals R to zero, J being the number of outputs and N the number of data. 2) For cluster c, if an estimate for its linear model parameters already exists, forc each input vector p add flcJYv. l,p to the matrix of . residuals, otherwise add flpYj,p to R, Yj.P being the j-th element of the deSIred output vector for sample z p. 3) Least squares fit the linear model parameters of cluster c to predict the current residuals R, and subtract the (new) estimate, fl;Y;,p' from R. 4) Do 2-4 for each cluster and repeat the fitting several times (default: 3). From simulations we found that the above fast optimization method converges to the global minimum if it is repeated many times. Just as with neural network training, it is often better to use early stopping when the prediction error on an independent test set starts to increase. 883 Robust Learning o/Chaotic Attractors 4. Nonlinear Principal Component Analysis To learn a chaotic attractor from a single experimental time-series we use the method ofdelays: the state l consists of m delays taken from the time series. The embedding dimension m must be chosen large enough to ensure that it contains sufficient infonnation for faithful reconstruction of the chaotic attractor, see Takens [9]. Typically, this results in an mdimensional state space with all the measurents covering only a much lower dimensional, but non-linearly shaped, subspace. This creates the danger pointed out in Sec. 2: the stability of the model in directions perpendicular to this low dimensional subspace cannot be guaranteed. With the split & fit algorithm from Sec. 3 we can learn the non-linear shape of the low dimensional subspace, and, if the state of the system escapes from this subspace, we use the algorithm to redirect the state to the nearest point on the subspace. See Malthouse [10] for limitations of existing nonlinear peA approaches. To obtain the low dimensional subspace, we proceed according to Procedure 5. Procedure 5. Learning the Low-dimensional Subspace 1) Augment the output of the model with the m-dimensional statel: the model will learn to predict its own input. 2) In each cluster c, perfonn a singular value decomposition to create a set of m principal directions, sorted in order of decreasing explained variance. The result of this decomposition is also used in step 3 of Procedure 4. 3) Allow the local linear model of each cluster to use no more than mred of these principal directions. 4) Define a potential P to be the squared Euclidian distance between the state l and its prediction by the model. The potential P implicitly defines the lower dimensional subspace: if a state l is on the subspace, P will be zero. P will increase with the distance of l from the subspace. The model has learned to predict its own input with small error, meaning that it has tried to reduce P as much as possible at exactly those points in state space where the training data was sampled. In other words, P will be low if the input l is close to one of the original points in the training data set. From the split&fit algorithm we can analytically compute the gradient dPldl. Since the evaluation of the split&fit model involves a backward (computing the bandwidths) and forward pass (computing memberships), the gradient algorithm involves a forward and backward pass through the tree. The gradient is used to project states that are off the nonlinear subspace onto the subspace -2 -1 2 Figure 2. Projecting two-dimensional data on a onedimensional self-intersecting subspace. The colorscale represents the potential P, white indicates P>0.04 .. 884 R. Bakker, J C. Schouten, M.-O. Coppens, F. Takens, C. L. Giles and C. M. v. d. Bleek in one or a few Newton-Rhapson iterations. Figure 2 illustrates the algorithm for the problem of creating a one-dimensional representation of the number '8'. The training set consists of 136 clean samples, Xl and Fig. 2 shows how a set of 272 noisy inputs is projected by a 48 subset split&fit model onto the one-dimensional subspace. Note that the center of the '8' cannot be well represented by a one-dimensional space. We leave development of an algorithm that -1 o X1 automatically detects the optimum local subspace dimension for future research. Figure 3. Learning the attractor of the twoinput logistic map. The order of creation of the 5. Application Examples splits is indicated. The colorscale represents the potential P, white indicates P>O.05. First we show the nonlinear principal component analysis result for a toy example, the logistic map Zt+l =4z t (1-Zt). If we use a model Zt+l =Fw(zt) , where the prediction only depends on one previous output, there is no lower dimensional space to which the attractor is confined. However, if we allow the output to depend on more than a single delay, we create a possibility for unstable behavior. Figure 3 shows how well the split&fit algorithm learns the one-dimensional shape of the attractor after creating only five regions. The parabola is slightly deformed (seen from the white lines perpendicular to the attractor), but this may be solved by increasing the number of splits. Next we look at the laser data. The complex behavior of chaotic systems is caused by an interplay of destabilizing and stabilizing forces: the destabilizing forces make nearby points in state space diverge, while the stabilizing forces keep the state of the system bounded. This process, known as 'stretching and folding', results in the attractor of the system: the set of points that the state of the system will visit after all transients have died out. In the case of the laser data this behavior is clear cut: destabilizing forces make the signal grow exponentially until the increasing amplitude triggers a collapse that reinitiates the sequence. We have seen in neural network based models [3] and in this study that it is very hard for the models to cope with the sudden collapses. Without the nonlinear subspace correction of Sec. 4, most of the (a) train data 0.4 , - - - - - - - - - + - - - - - - - - - - - - - - - - - - - - , (b) 1000 time Figure 4. Laser data from the Santa Fe time series competition. The 1000 sample train data set is followed by iterated prediction of the model (a). After every prediction a correction is made to keep P (see Sec. 4) small. Plot (b) shows P before this correction. Robust Learning of Chaotic Attractors 885 models we tested grow without bounds after one or more rise and collapse sequences. That is not very surprising - the training data set contains only three examples of a collapse. Figure 4 shows how this is solved with the subspace correction: every time the model is about to grow to infinity, a high potential P is detected (depicted in Fig. 3b) and the state of the system is directed to the nearest point on the subspace as learned from the nonlinear principal component analysis. After some trial and error, we selected an embedding dimension m of 12 and a reduced dimension mred of 4. The split&fit model starts with a single dataset, and was grown until 48 subsets. At that point, the error on the 1000 sample train set was still decreasing rapidly but the error on an independent 1000 sample test set increased. We compared the reconstructed attractors of the model and measurements, using 9000 samples of closed-loop generated and 9000 samples of measured data. No significant difference between the two could be detected by the Diles test [7]. 6. Conclusions We present an algorithm that robustly models chaotic attractors. It simultaneously learns (1) to make accurate short term predictions; and (2) the outlines of the attractor. In closed-loop prediction mode, the state of the system is corrected after every prediction, to stay within these outlines. The algorithm is very fast, since the main computation is to least squares fit a set of local linear models. In our implementation the largest matrix to be stored is N by C, N being the number of data and C the number of clusters. We see many applications other than attractor learning: the split&fit algorithm can be used as a fast learning alternative to neural networks and the new form of nonlinear peA will be useful for data reduction and object recognition. We envisage to apply the technique to a wide range of applications, from the control and modeling of chaos in fluid dynamics to problems in finance and biology to fluid dynamics. Acknowledgements This work is supported by the Netherlands Foundation for Chemical Research (SON) with financial aid from the Netherlands Organization for Scientific Research (NWO). References [1] 1.e. Principe, A. Rathie, and 1.M. Kuo. "Prediction of Chaotic Time Series with Neural Networks and the Issue of Dynamic Modeling". Int. J. Bifurcation and Chaos. 2, 1992. P 989. [2] 1.M. Kuo. and 1.C. Principe. "Reconstructed Dynamics and Chaotic Signal Modeling". In Proc. IEEE Int'l Conf. Neural Networks, 5, 1994, p 3l31. [3] R Bakker, J.C. Schouten, e.L. Giles. F. Takens, e.M. van den Bleek, "Learning Chaotic Attractors by Neural Networks", submitted. [4] 1.e. Principe, L. Wang, MA Motter, "Local Dynamic Modeling with Self-Organizing Maps and Applications to Nonlinear System Identification and Control" .Proc. IEEE. 86(11). 1998. [5] N. Kambhatla, T.K. Leen. "Dimension Reduction by Local PCA", Neural Computation. 9,1997. p. 1493 [6] M.I. Jordan, RA. Jacobs. "Hierarchical Mixtures of Experts and the EM Algorithm". Neural Compution. 6. 1994. p. 181. [7] e. Diks, W.R. van Zwet. F. Takens. and 1. de Goede, "Detecting differences between delay vector distributions", PhYSical Review E. 53, 1996. p. 2169. [8] A. Lapedes. R Farber. "Nonlinear Signal Processing Using Neural Networks: Prediction and System Modelling". Los Alamos Technical Report LA-UR-87-2662. [9] F. Takens, "Detecting strange attractors in turbulence", Lecture notes in Mathematics, 898, 1981, p. 365. [10] E.C. Malthouse. "Limitations of Nonlinear PCA as performed with Generic Neural Networks. IEEE Trans. Neural Networks. 9(1). 1998. p. 165.
1642 |@word deformed:1 trial:1 simulation:2 tried:1 t_:1 jacob:4 decomposition:4 euclidian:1 recursively:1 reduction:2 initial:1 series:14 contains:2 selecting:1 lapedes:2 existing:1 current:2 com:1 surprising:1 written:1 must:1 partition:1 shape:3 enables:1 compution:1 extrapolating:1 designed:2 plot:1 tenn:2 selected:1 guess:1 plane:2 short:5 sudden:1 infrastructure:1 detecting:2 math:1 node:7 hyperplanes:2 sigmoidal:1 five:1 along:1 become:1 consists:4 fitting:4 introduce:2 ra:1 behavior:4 growing:1 ol:2 terminal:6 inspired:1 detects:1 decreasing:2 automatically:1 resolve:1 increasing:2 becomes:1 stm:4 project:1 bounded:1 what:4 bakker:6 eigenvector:2 fuzzy:4 nj:1 perfonn:1 guarantee:1 every:3 finance:1 gravity:1 exactly:2 partitioning:13 control:4 positive:1 before:2 engineering:5 local:14 died:1 limit:1 path:1 resembles:1 collapse:4 perpendicular:4 range:2 directed:1 faithful:1 yj:2 lost:1 practice:1 chaotic:21 procedure:13 danger:1 destabilizing:3 confidence:1 word:1 cannot:3 close:1 onto:2 turbulence:1 put:1 instability:2 optimize:3 www:1 deterministic:1 map:5 center:2 starting:1 focused:1 stabilizing:2 splitting:3 financial:1 stability:2 embedding:2 construction:1 imagine:2 trigger:1 olivier:1 us:1 element:1 recognition:1 parabola:1 cut:1 solved:4 enters:1 wang:1 region:9 complexity:2 colorscale:2 dynamic:6 neglected:1 trained:2 carrying:1 depend:3 creation:1 dilemma:1 creates:3 basis:1 represented:2 grown:1 laser:6 univ:4 train:3 fast:4 describe:1 detected:2 outside:1 solve:1 reconstruct:1 otherwise:1 noisy:1 envisage:1 interplay:1 sequence:2 eigenvalue:1 took:1 propose:1 reconstruction:1 product:2 kohonen:1 neighboring:3 loop:4 rapidly:1 organizing:2 competition:4 los:1 cluster:11 optimum:1 zp:2 produce:1 converges:2 leave:1 object:1 measured:5 nearest:2 progress:1 dividing:1 involves:2 met:1 direction:7 closely:1 farber:2 pea:2 transient:1 explains:1 really:1 eindhoven:1 pl:2 correction:4 normal:1 predict:6 kambhatla:2 early:2 adopt:1 estimation:1 proc:2 infonnation:1 nwo:1 largest:6 create:4 weighted:1 clearly:1 always:2 pn:4 ej:1 groningen:1 focus:1 modelling:1 indicates:2 greatly:1 pentium:1 detect:2 stopping:2 membership:2 entire:1 typically:1 upward:1 issue:1 flexible:1 orientation:1 augment:1 takens:9 development:2 constrained:1 initialize:1 bifurcation:1 once:1 never:1 having:2 f3:2 shaped:1 identical:2 represents:2 biology:1 look:1 future:1 minimized:1 report:1 serious:1 few:2 escape:1 simultaneously:2 divergence:1 reactor:5 delft:4 attractor:41 organization:1 possibility:1 evaluation:1 introduces:1 mixture:3 nl:6 pc:2 accurate:2 tree:7 euclidean:1 desired:1 fitted:1 mdimensional:1 increased:1 column:1 modeling:7 giles:6 mhz:1 disadvantage:1 introducing:1 subset:8 alamo:1 delay:3 stored:1 fundamental:3 stay:2 lee:1 off:1 diverge:1 quickly:2 intersecting:1 squared:2 conf:1 creating:5 expert:3 toy:2 potential:6 de:1 sec:11 int:2 caused:1 depends:1 performed:1 try:1 lot:1 closed:4 lab:1 pendulum:1 start:3 jaap:1 minimize:1 square:4 accuracy:1 variance:7 characteristic:2 who:1 stretching:1 identify:1 climbing:1 identification:1 iterated:1 produced:1 monitoring:1 ago:1 submitted:1 sampled:1 dataset:1 amplitude:1 improved:1 leen:2 just:1 implicit:1 l31:1 until:4 nonlinear:23 defines:1 logistic:3 mode:1 indicated:1 scientific:1 verify:1 contain:2 counterpart:1 regularization:1 assigned:1 chemical:6 analytically:1 white:3 self:3 covering:1 criterion:1 trying:1 outline:7 complete:1 meaning:1 chaos:2 common:2 physical:1 exponentially:1 onedimensional:1 measurement:4 significant:1 smoothness:1 outlined:2 mathematics:2 pointed:2 stable:1 attracting:1 add:2 own:2 binary:1 seen:2 minimum:1 rug:1 greater:1 converge:1 paradigm:1 determine:2 signal:3 ii:7 technical:1 match:4 long:3 divided:1 visit:1 prediction:20 regression:5 iteration:5 confined:1 folding:1 affecting:1 else:1 singular:4 grow:3 jordan:4 near:1 split:18 iii:2 enough:1 variety:1 fit:17 bandwidth:8 reduce:2 pca:6 passed:1 proceed:1 cause:2 cpt:1 useful:1 santa:4 clear:1 netherlands:3 locally:2 reduced:1 generate:1 http:1 discrete:1 motter:1 prevent:2 clean:1 backward:2 sum:2 oftwo:1 almost:1 strange:1 fjs:1 prefer:1 fl:2 bound:1 guaranteed:1 followed:1 infinity:1 nearby:1 according:1 smaller:1 slightly:1 son:1 em:1 ur:1 happens:2 leg:1 den:2 explained:1 pr:1 projecting:1 taken:4 equation:1 previously:1 remains:1 rathie:1 cor:1 apply:2 hierarchical:4 enforce:1 generic:1 robustly:1 alternative:1 original:3 top:2 ensure:2 unfortunate:1 newton:1 bl:1 objective:1 noticed:1 already:3 diagonal:1 gradient:3 subspace:18 distance:4 tue:1 unstable:3 reason:1 minimizing:2 unfortunately:2 fe:4 subproblems:1 rise:1 fluid:2 implementation:1 zt:6 adjustable:2 perform:2 benchmark:1 defining:1 locate:1 learned:2 discontinuity:1 trans:1 able:1 mismatch:1 tudelft:4 overlap:2 force:4 circumvent:1 residual:3 scheme:1 technology:4 redirect:1 concludes:1 carried:1 review:1 acknowledgement:1 lecture:1 limitation:3 localized:1 foundation:1 sufficient:1 share:3 lo:1 elsewhere:1 repeat:1 last:2 keeping:1 free:1 supported:1 schouten:6 side:1 allow:2 understand:1 institute:1 wide:2 taking:1 van:3 boundary:4 curve:1 default:1 dimension:5 forward:3 made:1 regressors:1 projected:1 cope:1 reconstructed:4 pruning:1 implicitly:1 keep:2 global:4 iterative:2 learn:8 robust:5 complex:1 som:1 marc:1 spread:2 main:1 linearly:2 whole:1 noise:1 child:1 repeated:1 x1:1 fig:2 aid:1 xl:1 learns:3 minute:1 formula:2 intractable:1 exists:1 nec:2 illustrates:1 horizon:1 subtract:1 depicted:1 fluctuated:1 chance:1 ma:1 sorted:1 exceptionally:1 fw:1 hard:1 corrected:1 hyperplane:1 principal:10 called:1 pas:2 kuo:2 experimental:4 la:1 select:1 principe:5 arises:1 dept:1 princeton:1 tested:1 phenomenon:1
703
1,643
Correctness of belief propagation in Gaussian graphical models of arbitrary topology Yair Weiss Computer Science Division UC Berkeley, 485 Soda Hall Berkeley, CA 94720-1776 Phone: 510-642-5029 William T. Freeman Mitsubishi Electric Research Lab 201 Broadway Cambridge, MA 02139 Phone: 617-621-7527 yweiss@cs.berkeley.edu freeman @merl.com Abstract Local "belief propagation" rules of the sort proposed by Pearl [15] are guaranteed to converge to the correct posterior probabilities in singly connected graphical models. Recently, a number of researchers have empirically demonstrated good performance of "loopy belief propagation"using these same rules on graphs with loops. Perhaps the most dramatic instance is the near Shannon-limit performance of "Turbo codes", whose decoding algorithm is equivalent to loopy belief propagation. Except for the case of graphs with a single loop, there has been little theoretical understanding of the performance of loopy propagation. Here we analyze belief propagation in networks with arbitrary topologies when the nodes in the graph describe jointly Gaussian random variables. We give an analytical formula relating the true posterior probabilities with those calculated using loopy propagation. We give sufficient conditions for convergence and show that when belief propagation converges it gives the correct posterior means for all graph topologies, not just networks with a single loop. The related "max-product" belief propagation algorithm finds the maximum posterior probability estimate for singly connected networks. We show that, even for non-Gaussian probability distributions, the convergence points of the max-product algorithm in loopy networks are maxima over a particular large local neighborhood of the posterior probability. These results help clarify the empirical performance results and motivate using the powerful belief propagation algorithm in a broader class of networks. Problems involving probabilistic belief propagation arise in a wide variety of applications, including error correcting codes, speech recognition and medical diagnosis. If the graph is singly connected, there exist local message-passing schemes to calculate the posterior probability of an unobserved variable given the observed variables. Pearl [15] derived such a scheme for singly connected Bayesian networks and showed that this "belief propagation" algorithm is guaranteed to converge to the correct posterior probabilities (or "beliefs"). Several groups have recently reported excellent experimental results by running algorithms 674 Y. Weiss and W T. Freeman equivalent to Pearl's algorithm on networks with loops [8, 13, 6]. Perhaps the most dramatic instance of this performance is for "Turbo code" [2] error correcting codes. These codes have been described as "the most exciting and potentially important development in coding theory in many years" [12] and have recently been shown [10, 11] to utilize an algorithm equivalent to belief propagation in a network with loops. Progress in the analysis of loopy belief propagation has been made for the case of networks with a single loop [17, 18, 4, 1] . For these networks, it can be shown that (1) unless all the compatabilities are deterministic, loopy belief propagation will converge. (2) The difference between the loopy beliefs and the true beliefs is related to the convergence rate of the messages - the faster the convergence the more exact the approximation and (3) If the hidden nodes are binary, then the loopy beliefs and the true beliefs are both maximized by the same assignments, although the confidence in that assignment is wrong for the loopy beliefs. In this paper we analyze belief propagation in graphs of arbitrary topology, for nodes describing jointly Gaussian random variables. We give an exact formula relating the correct marginal posterior probabilities with the ones calculated using loopy belief propagation. We show that if belief propagation converges, then it will give the correct posterior means for all graph topologies, not just networks with a single loop. We show that the covariance estimates will generally be incorrect but present a relationship between the error in the covariance estimates and the convergence speed. For Gaussian or non-Gaussian variables, we show that the "max-product" algorithm, which calculates the MAP estimate in singly connected networks, only converges to points that are maxima over a particular large neighborhood of the posterior probability of loopy networks. 1 Analysis To simplify the notation, we assume the graphical model has been preprocessed into an undirected graphical model with pairwise potentials. Any graphical model can be converted into this form, and running belief propagation on the pairwise graph is equivalent to running belief propagation on the original graph [18]. We assume each node X i has a local observation Yi . In each iteration of belief propagation, each node X i sends a message to each neighboring X j that is based on the messages it received from the other neighbors, its local observation Yl and the pairwise potentials Wij(Xi , Xj) and Wii(Xi, Yi) . We assume the message-passing occurs in parallel. The idea behind the analysis is to build an unwrapped tree. The unwrapped tree is the graphical model which belief propagation is solving exactly when one applies the belief propagation rules in a loopy network [9, 20, 18]. It is constructed by maintaining the same local neighborhood structure as the loopy network but nodes are replicated so there are no loops. The potentials and the observations are replicated from the loopy graph. Figure 1 (a) shows an unwrapped tree for the diamond shaped graph in (b). By construction, the belief at the root node X- I is identical to that at node Xl in the loopy graph after four iterations of belief propagation. Each node has a shaded observed node attached to it, omitted here for clarity. Because the original network represents jointly Gaussian variables, so will the unwrapped tree. Since it is a tree, belief propagation is guaranteed to give the correct answer for the unwrapped graph. We can thus use Gaussian marginalization formulae to calculate the true mean and variances in both the original and the unwrapped networks. In this way, we calculate the accuracy of belief propagation for Gaussian networks of arbitrary topology. We assume that the joint mean is zero (the means can be added-in later). The joint distri- Correctness of Belief Propagation 675 Figure 1: Left: A Markov network with mUltiple loops. Right: The unwrapped network corresponding to this structure. bution of z =(: ) is given by P(z) = ae-!zTVz, where V = (~:: ~::) . It is straightforward to construct the inverse covariance matrix V of the joint Gaussian that describes a given Gaussian graphical model [3]. Writing out the exponent of the joint and completing the square shows that the mean I-' of x, given the observations y, is given by: (1) and the covariance matrix C~IY of x given y is: C~IY = V~-;l. We will denote by C~dY the ith row of C~IY so the marginal posterior variance of Xi given the data is (72 (i) = C~i Iy (i). We will use - for unwrapped quantities. We scan the tree in breadth first order and denote by x the vector of values in the hidden nodes of the tree when so scanned. Simlarly, we denote by y the observed nodes scanned in the same order and Vn , V~y the inverse covariance matrices. Since we are scanning in breadth first order the last nodes are the leaf nodes and we denote by L the number of leaf nodes. By the nature of unwrapping, tL(1) is the mean of the belief at node Xl after t iterations of belief propagation, where t is the number of unwrappings. Similarly 0- 2 (1) = 6~1Iy(1) is the variance of the belief at node Xl after t iterations. Because the data is replicated we can write y = Oy where O(i, j) = 1 if Yi is a replica of Yj and 0 otherwise. Since the potentials W(Xi' Yi) are replicated, we can write V~yO = OV~y. Since the W(Xi, X j) are also replicated and all non-leaf Xi have the same connectivity as the corresponding Xi, we can write V~~O = OVzz + E where E is zero in all but the last L rows. When these relationships between the loopy and unwrapped inverse covariance matrices are substituted into the loopy and unwrapped versions of equation I, one obtains the following expression, true for any iteration [19]: (2) where e is a vector that is zero everywhere but the last L components (corresponding to the leaf nodes). Our choice of the node for the root of the tree is arbitrary, so this applies to all nodes of the loopy network. This formula relates, for any node of a network with loops, the means calculated at each iteration by belief propagation with the true posterior means. Similarly when the relationship between the loopy and unwrapped inverse covariance matrices is substituted into the loopy and unwrapped definitions of C~IY we can relate the Y Weiss and W T Freeman 676 0.5 0.4 ~ 0.3 ~ .~ 0.2 n; ~ 0.1 8 "t:> ? 0 -0.1 -0.2 20 0 40 60 80 100 node Figure 2: The conditional correlation between the root node and all other nodes in the unwrapped tree of Fig. 1 after eight iterations. Potentials were chosen randomly. Nodes are presented in breadth first order so the last elements are the correlations between the root node and the leaf nodes. We show that if this correlation goes to zero, belief propagation converges and the loopy means are exact. Symbols plotted with a star denote correlations with nodes that correspond to the node Xl in the loopy graph. The sum of these correlations gives the correct variance of node Xl while loopy propagation uses only the first correlation. marginalized covariances calculated by belief propagation to the true ones [19]: -2 a (1) = a 2 (1) + CZllyel - - Czt/ye2 (3) where el is a vector that is zero everywhere but the last L components while e2 is equal to 1 for all nodes in the unwrapped tree that are replicas of Xl except for Xl. All other components of e2 are zero, Figure 2 shows Cz1l Y for the diamond network in Fig. 1. We generated random potential functions and observations and calculated the conditional correlations in the unwrapped tree. Note that the conditional correlation decreases with distance in the tree - we are scanning in breadth first order so the last L components correspond to the leaf nodes. As the number of iterations of loopy propagation is increased the size of the unwrapped tree increases and the conditional correlation between the leaf nodes and the root node decreases. From equations 2-3 it is clear that if the conditional correlation between the leaf nodes and the root nodes are zero for all sufficiently large unwrappings then (1) belief propagation converges (2) the means are exact and (3) the variances may be incorrect. In practice the conditional correlations will not actually be equal to zero for any finite unwrapping. In [19] we give a more precise statement: if the conditional correlation of the root node and the leaf nodes decreases rapidly enough then (1) belief propagation converges (2) the means are exact and (3) the variances may be incorrect. We also show sufficient conditions on the potentials III (Xi, X j) for the correlation to decrease rapidly enough: the rate at which the correlation decreases is determined by the ratio of off-diagonal and diagonal components in the quadratic fonn defining the potentials [19]. How wrong will the variances be? The tenn CZllye2 in equation 3 is simply the sum of many components of Cz11y . Figure 2 shows these components. The correct variance is the sum of all the components witHe the belief propagation variance approximates this sum with the first (and dominant) tenn. Whenever there is a positive correlation between the root node and other replicas of Xl the loopy variance is strictly less than the true variance - the loopy estimate is overconfident. 677 Correctness of Belief Propagation ~07 e iDO.6 ., ";;;05 fr ~04 SOR '"~03 0.2 0.1 20 30 40 50 60 iterations (a) (b) Figure 3: (a) 25 x 25 graphical model for simulation. The unobserved nodes (unfilled) were connected to their four nearest neighbors and to an observation node (filled). (b) The error of the estimates of loopy propagation and successive over-relaxation (SOR) as a function of iteration. Note that belief propagation converges much faster than SOR. Note that when the conditional correlation decreases rapidly to zero two things happen. First, the convergence is faster (because CZdyel approaches zero faster) . Second, the approximation error of the variances is smaller (because CZ1 / y e2 is smaller). Thus we have shown, as in the single loop case, quick convergence is correlated with good approximation. 2 Simulations We ran belief propagation on the 25 x 25 2D grid of Fig. 3 a. The joint probability was: (4) where Wij = 0 if nodes Xi, Xj are not neighbors and 0.01 otherwise and Wii was randomly selected to be 0 or 1 for all i with probability of 1 set to 0.2. The observations Yi were chosen randomly. This problem corresponds to an approximation problem from sparse data where only 20% of the points are visible. We found the exact posterior by solving equation 1. We also ran belief propagation and found that when it converged, the calculated means were identical to the true means up to machine precision. Also, as predicted by the theory, the calculated variances were too small - the belief propagation estimate was overconfident. In many applications, the solution of equation 1 by matrix inversion is intractable and iterative methods are used. Figure 3 compares the error in the means as a function of iterations for loopy propagation and successive-over-relaxation (SOR), considered one of the best relaxation methods [16]. Note that after essentially five iterations loopy propagation gives the right answer while SOR requires many more. As expected by the fast convergence, the approximation error in the variances was quite small. The median error was 0.018. For comparison the true variances ranged from 0.01 to 0.94 with a mean of 0.322. Also, the nodes for which the approximation error was worse were indeed the nodes that converged slower. Y. Weiss and W T Freeman 678 3 Discussion Independently, two other groups have recently analyzed special cases of Gaussian graphical models. Frey [7] analyzed the graphical model corresponding to factor analysis and gave conditions for the existence of a stable fixed-point. Rusmevichientong and Van Roy [14] analyzed a graphical model with the topology of turbo decoding but a Gaussian joint density. For this specific graph they gave sufficient conditions for convergence and showed that the means are exact. Our main interest in the Gaussian case is to understand the performance of belief propagation in general networks with multiple loops. We are struck by the similarity of our results for Gaussians in arbitrary networks and the results for single loops of arbitrary distributions [18]. First, in single loop networks with binary nodes, loopy belief at a node and the true belief at a node are maximized by the same assignment while the confidence in that assignment is incorrect. In Gaussian networks with multiple loops, the mean at each node is correct but the confidence around that mean may be incorrect. Second, for both singleloop and Gaussian networks, fast belief propagation convergence correlates with accurate beliefs. Third, in both Gaussians and discrete valued single loop networks, the statistical dependence between root and leaf nodes governs the convergence rate and accuracy. The two models are quite different. Mean field approximations are exact for Gaussian MRFs while they work poorly in sparsely connected discrete networks with a single loop. The results for the Gaussian and single-loop cases lead us to believe that similar results may hold for a larger class of networks. Can our analysis be extended to non-Gaussian distributions? The basic idea applies to arbitrary graphs and arbitrary potentials: belief propagation is performing exact inference on a tree that has the same local neighbor structure as the loopy graph. However, the linear algebra that we used to calculate exact expressions for the error in belief propagation at any iteration holds only for Gaussian variables. We have used a similar approach to analyze the related "max-product" belief propagation algorithm on arbitrary graphs with arbitrary distributions [5] (both discrete and continuous valued nodes). We show that if the max-product algorithm converges, the max-product assignment has greater posterior probability then any assignment in a particular large region around that assignment. While this is a weaker condition than a global maximum, it is much stronger than a simple local maximum of the posterior probability. The sum-product and max-product belief propagation algorithms are fast and parallelizable. Due to the well known hardness of probabilistic inference in graphical models, belief propagation will obviously not work for arbitrary networks and distributions. Nevertheless, a growing body of empirical evidence shows its success in many networks with loops. Our results justify applying belief propagation in certain networks with mUltiple loops. This may enable fast, approximate probabilistic inference in a range of new applications. References [1] S.M. Aji, G.B. Hom, and R.J. McEliece. On the convergence of iterative decoding on graphs with a single cycle. In Proc. 1998 ISIT, 1998. [2] c. Berrou, A. Glavieux, and P. Thitimajshima. Near Shannon limit error-correcting coding and decoding: Turbo codes. In Proc. IEEE International Communications Conference '93, 1993. [3] R. Cowell. Advanced inference in Bayesian networks. In M.1. Jordan, editor, Learning in Graphical Models . MIT Press, 1998. [4] G.D. Forney, F.R. Kschischang, and B. Marcus. Iterative decoding of tail-biting trellisses. preprint presented at 1998 Information Theory Workshop in San Diego, 1998. Correctness of Belief Propagation 679 [5] W. T. Freeman and Y. Weiss. On the fixed points of the max-product algorithm. Technical Report 99-39, MERL, 201 Broadway, Cambridge, MA 02139, 1999. [6] W.T. Freeman and E.C. Pasztor. Learning to estimate scenes from images. In M.S. Kearns, S.A. SoUa, and D.A. Cohn, editors, Adv. Neural Information Processing Systems I I. MIT Press, 1999. [7] B.J. Frey. Turbo factor analysis. In Adv. Neural Information Processing Systems 12. 2000. to appear. [8) Brendan J. Frey. Bayesian Networksfor Pattern Classification, Data Compression and Channel Coding. MIT Press, 1998. [9) R.G . Gallager. Low Density Parity Check Codes. MIT Press, 1963. [10) F. R. Kschischang and B. J. Frey. Iterative decoding of compound codes by probability propagation in graphical models. IEEE Journal on Selected Areas in Communication , 16(2):219-230, 1998. [11] R.J. McEliece, D.J .C. MackKay, and J.F. Cheng. Turbo decoding as as an instance of Pearl's 'belief propagation' algorithm. IEEE Journal on Selected Areas in Communication, 16(2): 140152,1998. [12J R.J. McEliece, E. Rodemich, and J.F. Cheng. The Turbo decision algorithm. In Proc. 33rd Allerton Conference on Communications, Control and Computing, pages 366-379, Monticello, IL, 1995. [I3J K.P. Murphy, Y. Weiss, and M.1. Jordan. Loopy belief propagation for approximate inference : an empirical study. In Proceedings of Uncertainty in AI, 1999. [14] Rusmevichientong P. and Van Roy B. An analysis of Turbo decoding with Gaussian densities. In Adv. Neural Information Processing Systems I2 . 2000. to appear. [15) Judea Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, 1988. [16J Gilbert Strang. Introduction to Applied Mathel1Ultics. Wellesley-Cambridge, 1986. [I7J Y. Weiss. Belief propagation and revision in networks with loops. Technical Report 1616, MIT AI lab, 1997. [18J Y. Weiss. Correctness of local probability propagation in graphical models with loops. Neural Computation, to appear, 2000. [19] Y. Weiss and W. T. Freeman. Loopy propagation gives the correct posterior means for Gaussians. Technical Report UCB.CSD-99-1046, Berkeley Computer Science Dept., 1999. www.cs.berkeley.edu yweiss/. [20J N. Wiberg. Codes and decoding on general graphs. PhD thesis, Department of Electrical Engineering, U. Linkoping, Sweden, 1996. Gaussian Fields for Approximate Inference in Layered Sigmoid Belief Networks David Barber'" Stichting Neurale Netwerken Medical Physics and Biophysics Nijmegen University, The Netherlands barberdOaston.ac.uk Peter Sollich Department of Mathematics King's College, University of London London WC2R 2LS, U.K. peter.sollichOkcl.ac.uk Abstract Layered Sigmoid Belief Networks are directed graphical models in which the local conditional probabilities are parameterised by weighted sums of parental states. Learning and inference in such networks are generally intractable, and approximations need to be considered. Progress in learning these networks has been made by using variational procedures. We demonstrate, however, that variational procedures can be inappropriate for the equally important issue of inference - that is,? calculating marginals of the network. We introduce an alternative procedure, based on assuming that the weighted input to a node is approximately Gaussian distributed. Our approach goes beyond previous Gaussian field assumptions in that we take into account correlations between parents of nodes. This procedure is specialized for calculating marginals and is significantly faster and simpler than the variational procedure. 1 Introduction Layered Sigmoid Belief Networks [1] are directed graphical models [2] in which the local conditional probabilities are parameterised by weighted sums of parental states, see fig ( 1). This is a graphical representation of a distribution over a set of binary variables Si E {a, I}. Typically, one supposes that the states of the nodes at the bottom of the network are generated by states in previous layers. Whilst, in principle, there is no restriction on the number of nodes in any layer, typically, one considers structures similar to the "fan out" in fig(l) in which higher level layers provide an "explanation" for patterns generated in lower layers. Such graphical models are attractive since they correspond to layers of information processors, of potentially increasing complexity. Unfortunately, learning and inference in such networks is generally intractable, and approximations need to be considered. Progress in learning has been made by using variational procedures [3,4, 5]. However, another crucial aspect remains inference [2]. That is, given some evidence (or none), calculate the marginal of a variable, conditional on this evidence. This assumes that we have found a suitable network from some learning procedure, and now wish ?Present Address: NCRG, Aston University, Birmingham B4 7ET, U.K. 394 D. Barber and P. Sollich to query this network. Whilst the variational procedure is attractive for learning, since it generally provides a bound on the likelihood of the visible units, we demonstrate that it may not always be equally appropriate for the inference problem. A directed graphical model defines a distribution over a set of variables s = (S1 ... sn) that factorises into the local conditional distributions, n p(S1 . .. sn) = IIp(silll'i) (1) i=1 where lI'i denotes the parent nodes of node i . In a layered network, these are the nodes in the proceeding layer that feed into node i. In a sigmoid belief network the local probabilities are defined as P (s; = ll~;) = " ( ~ W;jSj + 0;) Figure 1: A Layered Sigmoid Belief Network =" (h;) (2) where the "field" at node i is defined as hi = 2:j WijSj + fh and er(h) = 1/(1 + e- h ). Wij is the strength of the connection between node i and its parent node j; if j is not a parent of i we set Wij O. Oi is a bias term that gives a parent-independent bias to the state of node i . = We are interested in inference - in particular, calculating marginals of the network for cases with and without evidential nodes. In section (2) we describe how to approximate the quantities p(Si 1) and discuss in section (2.1) why our method can improve on the standard variational mean field theory. Conditional marginals, such as p(Si = IISj = 1, Sk = 0) are considered in section (3). = 2 Gaussian Field Distributions Under the 0/1 coding for the variables Si, the mean of a variable, mi is given by the probability that it is in state 1. Using the fact from (2) that the local conditional distribution of node i is dependent on its parents only through its field hi, we have (3) where we use the notation ?(-)p to denote an average with respect to the distribution p. If there are many parents of node i, a reasonable assumption is that the distribution of the field hi will be Gaussian, p(hi ) ~ N (J,Li,er[). Under this Gaussian Field (GF) assumption, we need to work out the mean and variance, which are given by (4) j j err = ((Llh i )2) = L WijWikRjk (5) j,k where Rjk = (LlSjLlsk). We use the notation Ll (-) == (-) - ?(.) . The diagonal terms of the node covariance matrix are ~i = mi (1- mi)' In contrast to previous studies, we include off diagonal terms in the calculation of R [4] . From Gaussian Fields for Approximate Inference 395 (5) we only need to find correlations between parents i and j of a node. These are easy to calculate in the layered networks that we are considering, because neither i nor j is a descendant of the other: Rjj = p(Sj = 1, Sj = 1) = J (6) mjmj p(Si = Ilhj)p(Sj = Ilhj)p(hj, hj)dh - mimj = (0" (hd 0" (h j ) P (7) (8) (h J, h) - mjmj J Assuming that the joint distribution p( h j , h j ) is Gaussian, we again need its mean and covariance, given by ~ij = (D.hjD.hj) = L WjkWjl (D.skD.SI) = L kl WikWjlRkl (10) kl Under this scheme, we have a closed set of equations, (4,5,8,10) for the means mj and covariance matrix Rij which can be solved by forward propagation of the equations. That is, we start from nodes without parents, and then consider the next layer of nodes, repeating the procedure until a full sweep through the network has been completed. The one and two dimensional field averages, equations (3) and (8), are computed using Gaussian Quadrature. This results in an extremely fast procedure for approximating the marginals mi, requiring only a single sweep through the network. Our approach is related to that of [6] by the common motivating assumption that each node has a large number of parents. This is used in [6] to obtain actual bounds on quantities of interest such as joint marginals. Our approach does not give bounds. Its advantage, however, is that it allows fluctuations in the fields hi, which are effectively excluded in [6] by the assumed scaling of the weights Wij with the number of parents per node. 2.1 Relation to Variational Mean Field Theory In the variational approach, one fits a tractable approximating distribution Q to the SBN. Taking Q factorised, Q(s) = Dj m:' (1 - md l - 3 ? we have the bound In p (Sl ... sn) 2: L {-mj In mj - (1 - md In (1 - md} i The final term in (11) causes some difficulty even in the case in which Q is a factorised model. Formally, this is because this term does not have the same graphical structure as the tractable model Q. One way around around this difficulty is to employ a further bound, with associated variational parameters [7]. Another approach is to make the Gaussian assumption for the field hi as in section (2). Because Q is factorised, corresponding to a diagonal correlation matrix R, this gives [4] (12) 396 D. Barber and P Sollich where Pi = ~j Wijmj + Oi and (1[ = ~j w[jmj(l - mj). Note that this is a one dimensional integral of a smooth function. In contrast to [4] we therefore evaluate this quantity using Gaussian Quadrature. This has the advantage that no extra variational parameters need to be introduced. Technically, the assumption of a Gaussian field distribution means that (11) is no longer a bound. Nevertheless, in practice it is found that this has little effect on the quality of the resulting solution. In our implementation of the variational approach, we find the optimal parameters mi by maximising the above equation for each component mi separately, cycling through the nodes until the parameters mi do not change by more than 10- 1 This is repeated 5 times, and the solution with the highest bound score is chosen. Note that these equations cannot be solved by forward propagation alone since the final term contains contributions from all the nodes in the network. This is in contrast to the GF approach of section (2) . Finding appropriate parameters mi by the variational approach is therefore rather slower than using the GF method. ?. In arriving at the above equations, we have made two assumptions. The first is that the intractable distribution is well approximated by a factorised model. The second is that the field distribution is Gaussian. The first step is necessary in order to obtain a bound on the likelihood of the model (although this is slightly compromised by the Gaussian field assumption). In the GF approach we dispense with this assumption of an effectively factorised network (partially because if we are only interested in inference, a bound on the model likelihood is less relevant). The GF method may therefore prove useful for a broader class of networks than the variational approach . 2.2 Results for unconditional marginals We compared three procedures for estimating the conditional values p(Si = 1) for all the nodes in the network, namely the variational theory, as described in section (2.1), the diagonal Gaussian field theory, and the non-diagonal Gaussian field theory which includes correlation effects between parents. Results for small weight values Wij are shown in fig(2). In this case, all three methods perform reasonably well, although there is a significant improvement in using the GF methods over the variational procedure; parental correlations are not important (compare figs(2b) and (2c)) . In fig(3) the weights and biases are chosen such that the exact mean variables mi are roughly 0.5 with non-trivial correlation effects between parents. Note that the variational mean field theory now provides a poor solution, whereas the GF methods are relatively accurate. The effect of using the non-diagonal R terms is beneficial, although not dramatically so. 3 Calculating Conditional Marginals We consider now how to calculate conditional marginals, given some evidential nodes. (In contrast to [6], any set of nodes in the network, not just output nodes, can be considered evidential.) We write the evidence in the following manner E = {SCi = SCi' . . . Sc" = SC,.} = {ECl ... Ec,.} The quantities that we are interested in are conditional marginals which, from Bayes rule are related to the joint distribution by P (Si = liE) = = P (Si 1, E) (13) P (Si = 0, E) + P (Si = 1, E) That is, provided that we have a procedure for estimating joint marginals, we can obtain conditional marginals too. Without loss of generality, we therefore consider Gaussian Fields for Approximate Inference Em>ruoing1_ model 1ft 2Or--~ 397 Em>r using Ga_ Fiaid. Ooagonal ooyariance cowuiance O<Xll (a) Mean error = 0.0377 (b) Mean error = 0.0018 (c) Mean error 001 = 0.0017 Figure 2: Error in approximating p(Si = 1) for the network in fig(l), averaged over all the nodes in the network. In each of 100 trials, weights were drawn from a zero mean, unit variance Gaussian; biases were set to O. Note the different scale in (b) and (c). In (a) we use the variational procedure with a factorised Q, as in section (2.1). In (b) we use the Gaussian field equations, assuming a diagonal covariance matrix R. This procedure was repeated in (c) including correlations between parents. E+ = E U {Si = I}, which then contains n + 1 "evidential" variables. That is, the desired marginal variable is absorbed into the evidence set . For convenience, we then split the nodes into two sets, those containing the evidential or "clamped" nodes, C, and the remaining "free" nodes F . The joint evidence is then given by (14) 8F = I:p (ECllll'~l) ... p (En+llll'~"+l) p (sh 11I'jJ ...p (Sfm 11I'jJ 8F (15) where 11'; are the parents of node i, with any evidential parental nodes set to their values as specified in E+. In the sigmoid belief network if i is an evidential node otherwise (16) = p(Eklll'Z) is therefore determined by the distribution of the field hZ Li WkiS; +Ok . Examining (15), we see that the product over the "free" nodes defines a SBN in which the local probability distributions are given by those of the original network, but with any evidential parental nodes clamped to their evidence values . Therefore, (17) Consistent with our previous assumptions, we assume that the distribution of the fields h+ = (h~l'" h~"+l) is jointly Gaussian. We can then find the mean and covariance matrix for the distribution of h+ by repeating the calculation of section (2) in which evidential nodes have been clamped to their evidence values. Once this Gaussian has been determined, it can be used in (17) to determine p( E+). Gaussian averages of products of sigmoids are calculated by drawing 1000 samples from the Gaussian over which we wish to integrate 1 . Note that if there are evidential nodes lIn one and two dimensions (n = 0, 1), or n = 1, we use Gaussian Quadrature. D. Barber and P Sollich 398 Error uoing lado_ model 1M l00r-~--~----~~--~ eo Error uoing Ga_ian Field. Diegorel """arianee Em:>< uoing Ga .....n Field. Non Diagonal ""w"ianee EO 70,---~------~--------, 50 EO 50 40 30 30 20 20 10 10 o1 (a) Mean error = 0.4188 0.2 0.3 0.4 0.5 06 (b) Mean error = 0.0253 00 II.. 01 02 0.3 0.4 0.5 0.6 (c) Mean error = 0.0198 Figure 3: All weights are set to uniformly from 0 to 50. Biases are set to -0.5 of the summed parental weights plus a uniform random number from -2.5 to 2.5 . The root node is set to be 1 with probability 0.5. This has the effect of making all the nodes in the exact network roughly 0.5 in mean, with non-negligible correlations between parental nodes. 160 simulations were made. in different layers, we require the correlations between their fields h to evaluate (17) . Such 'inter-layer' correlations were not required in section (2) , and to be able to use the same calculational scheme we simply neglect them. (We leave a study of the effects of this assumption for future work .) The average in (17) then factors into groups, where each group contains evidential terms in a particular layer. The conditional marginal for node i is obtained from repeating the above procedure in which the desired marginal node is clamped to its opposite value, and then using these results in (13). The above procedure is repeated for each conditional marginal that we are interested in. Although this may seem computationally expensive, the marginal for each node is computed quickly, since the equations are solved by one forward propagation sweep only. Error uoing Gauosian Field, Diago".1 covarIanee (a) Mean error = 0.1534 (b) Mean error = 0.0931 Em:>< uoing Gau"ian Field. Non Diagonal """ariance (c) Mean error = 0.0865 Figure 4: Estimating the conditional marginal of the top node being in state 1, given that the four bottom nodes are in state 1. Weights were drawn from a zero mean Gaussian with variance 5, with biases set to -0.5 the summed parental weights plus a uniform random number from -2.5 to 2.5 . Results of 160 simulations. 3.1 Results for conditional marginals We used the same structure as in the previous experiments, as shown in fig(I). We are interested here in calculating the probability that the top node is in state 1, Gaussian Fields for Approximate Inference 399 given that the four bottom nodes are in state 1. Weights were chosen from a zero mean Gaussian with variance 5. Biases were set to negative half of the summed parent weights, plus a uniform random value from -2.5 to 2.5. Correlation effects in these networks are not as strong as in the experiments in section (2.2), although the improvement of the G F theory over the variational theory seen in fig (4) remains clear. The improvement from the off diagonal terms in R is minimal. 4 Conclusion Despite their appropriateness for learning, variational methods may not be equally suited to inference, making more tailored methods attractive. We have considered an approximation procedure that is based on assuming that the distribution of the weighted input to a node is approximately Gaussian. Correlation effects between parents of a node were taken into account to improve the Gaussian theory, although in our examples this gave only relatively modest improvements. The variational mean field theory performs poorly in networks with strong correlation effects between nodes. On the other hand, one may conjecture that the Gaussian Field approach will not generally perform catastrophically worse than the factorised variational mean field theory. One advantage of the variational theory is the presence of an objective function against which competing solutions can be compared. However, finding an optimum solution for the mean parameters mj from this function is numerically complex. Since the Gaussian Field theory is extremely fast to solve, an interesting compromise might be to prime the variational solution with the results from the Gaussian Field theory. Acknowledgments DB would like to thank Bert Kappen and Wim Wiegerinck for stimulating and helpful discussions. PS thanks the Royal Society for financial support. [1] R. Neal. Connectionist learning of Belief Networks. Artificial Intelligence, 56:71-113, 1992. [2] E. Castillo, J. M. Gutierrez, and A. S. Radi. Expert Systems and Probabilistic Network Models. Springer, 1997. [3] M. I. Jordan, Z. Gharamani, T. S. Jaakola, and L. K. Saul. An Introduction to Variational Methods for Graphical Models. In M. I. Jordan, editor, Learning in Graphical Models, pages 105-161. Kluwer, 1998. [4] L. Saul and M. I. Jordan. A mean field learning algorithm for unsupervised neural networks. In M. I. Jordan, editor, Learning in Graphical Models, 1998. [5] D. Barber and W Wiegerinck. Tractable variational structures for approximating graphical models. In M.S. Kearns, S.A. Solla, and D.A. Cohn, editors, Advances in Neural Information Processing Systems NIPS 11. MIT Press, 1999. [6] M. Kearns and 1. Saul. Inference in Multilayer Networks via Large Deviation Bounds. In Advances in Neural Information Processing Systems NIPS 11, 1999. [7] L. K. Saul, T. Jaakkola, and M. I. Jordan. Mean Field Theory for Sigmoid Belief Networks. Journal of Artificial Intelligence Research, 4:61-76, 1996.
1643 |@word trial:1 version:1 inversion:1 compression:1 stronger:1 simulation:4 mitsubishi:1 covariance:13 fonn:1 dramatic:2 catastrophically:1 kappen:1 unwrappings:2 score:1 contains:3 skd:1 err:1 com:1 si:13 visible:2 happen:1 alone:1 tenn:2 leaf:10 selected:3 czt:1 half:1 intelligence:2 ith:1 provides:2 node:101 successive:2 allerton:1 simpler:1 five:1 wijsj:1 constructed:1 incorrect:5 descendant:1 prove:1 manner:1 introduce:1 pairwise:3 inter:1 indeed:1 hardness:1 expected:1 roughly:2 nor:1 growing:1 freeman:8 little:2 actual:1 inappropriate:1 considering:1 unwrapped:16 ye2:1 distri:1 revision:1 notation:3 increasing:1 estimating:3 provided:1 whilst:2 unobserved:2 finding:2 berkeley:5 exactly:1 wrong:2 uk:2 control:1 unit:2 medical:2 appear:3 positive:1 negligible:1 engineering:1 local:15 frey:4 limit:2 despite:1 fluctuation:1 approximately:2 might:1 plus:3 shaded:1 range:1 jaakola:1 averaged:1 directed:3 acknowledgment:1 yj:1 practice:2 procedure:18 aji:1 area:2 empirical:3 significantly:1 confidence:3 gau:1 cannot:1 convenience:1 layered:6 ga:1 applying:1 writing:1 restriction:1 gilbert:1 equivalent:4 deterministic:1 demonstrated:1 map:1 quick:1 www:1 straightforward:1 go:2 independently:1 l:1 radi:1 correcting:3 rule:4 financial:1 hd:1 construction:1 diego:1 exact:12 us:1 l00r:1 element:1 roy:2 recognition:1 approximated:1 expensive:1 sparsely:1 observed:3 bottom:3 ft:1 preprint:1 electrical:1 rij:1 solved:3 calculate:7 sbn:2 region:1 connected:7 cycle:1 adv:3 solla:1 decrease:6 highest:1 ran:2 complexity:1 dispense:1 motivate:1 ov:1 solving:2 algebra:1 compromise:1 technically:1 division:1 joint:11 fast:6 describe:2 london:2 query:1 sc:2 artificial:2 neighborhood:3 whose:1 quite:2 larger:1 valued:2 plausible:1 solve:1 drawing:1 otherwise:3 jointly:4 final:2 obviously:1 advantage:3 analytical:1 product:11 fr:1 biting:1 neighboring:1 relevant:1 loop:22 rapidly:3 poorly:2 linkoping:1 convergence:12 parent:17 ecl:1 optimum:1 p:1 converges:8 leave:1 help:1 ac:2 ij:1 nearest:1 received:1 progress:3 strong:2 c:2 predicted:1 appropriateness:1 correct:10 enable:1 require:1 wc2r:1 sor:5 yweiss:2 isit:1 strictly:1 clarify:1 hold:2 sufficiently:1 hall:1 considered:6 around:4 omitted:1 fh:1 proc:3 birmingham:1 wim:1 correctness:5 gutierrez:1 wiberg:1 weighted:4 mit:6 gaussian:53 always:1 rather:1 hj:3 broader:2 jaakkola:1 derived:1 yo:1 improvement:4 check:1 likelihood:3 contrast:4 brendan:1 helpful:1 inference:19 mrfs:1 el:1 dependent:1 typically:2 hidden:2 relation:1 wij:6 interested:5 issue:1 classification:1 exponent:1 development:1 special:1 summed:3 uc:1 marginal:9 equal:2 construct:1 field:37 shaped:1 once:1 identical:2 represents:1 unsupervised:1 future:1 report:3 connectionist:1 simplify:1 intelligent:1 employ:1 randomly:3 murphy:1 william:1 interest:2 message:5 analyzed:3 sh:1 unconditional:1 behind:1 accurate:2 integral:1 monticello:1 necessary:1 sweden:1 modest:1 unless:1 tree:14 filled:1 desired:2 plotted:1 theoretical:1 minimal:1 merl:2 instance:3 increased:1 i3j:1 assignment:7 loopy:34 deviation:1 stichting:1 uniform:3 examining:1 too:2 motivating:1 reported:1 answer:2 scanning:2 supposes:1 ido:1 thanks:1 density:3 international:1 probabilistic:5 yl:1 off:3 decoding:9 physic:1 iy:6 quickly:1 connectivity:1 thesis:1 again:1 iip:1 containing:1 worse:2 expert:1 li:3 withe:1 potential:9 converted:1 account:2 factorised:7 star:1 rusmevichientong:2 coding:4 includes:1 later:1 root:10 wellesley:1 lab:2 closed:1 analyze:3 bution:1 start:1 sort:1 bayes:1 parallel:1 simlarly:1 contribution:1 jsj:1 square:1 il:1 accuracy:2 oi:2 variance:19 kaufmann:1 maximized:2 correspond:3 bayesian:3 none:1 researcher:1 processor:1 converged:2 evidential:11 parallelizable:1 whenever:1 definition:1 against:1 e2:3 associated:1 mi:9 judea:1 actually:1 rodemich:1 feed:1 higher:1 ok:1 wei:9 generality:1 just:3 parameterised:2 correlation:29 mceliece:3 until:2 hand:1 cohn:2 propagation:62 defines:2 quality:1 perhaps:2 believe:1 effect:9 ranged:1 true:11 requiring:1 excluded:1 unfilled:1 i2:1 neal:1 attractive:3 ll:2 demonstrate:2 performs:1 reasoning:1 image:1 variational:25 recently:4 sigmoid:7 common:1 specialized:1 empirically:1 attached:1 b4:1 ncrg:1 tail:1 approximates:1 relating:2 marginals:13 numerically:1 kluwer:1 significant:1 cambridge:3 ai:2 rd:1 grid:1 mathematics:1 similarly:2 dj:1 stable:1 similarity:1 soua:1 longer:1 dominant:1 posterior:16 showed:2 phone:2 prime:1 compound:1 certain:1 binary:3 success:1 yi:5 morgan:1 seen:1 greater:1 eo:3 converge:3 berrou:1 determine:1 ii:1 relates:1 multiple:4 full:1 smooth:1 technical:3 faster:5 calculation:2 dept:1 lin:1 equally:3 biophysics:1 calculates:1 involving:1 basic:1 ae:1 essentially:1 multilayer:1 iteration:13 tailored:1 whereas:1 separately:1 median:1 sends:1 crucial:1 extra:1 hz:1 undirected:1 thing:1 db:1 seem:1 jordan:7 near:2 presence:1 iii:1 enough:2 easy:1 split:1 variety:1 xj:2 marginalization:1 gave:3 fit:1 topology:7 opposite:1 competing:1 idea:2 expression:2 peter:2 speech:1 passing:2 cause:1 jj:2 dramatically:1 generally:5 useful:1 clear:2 governs:1 netherlands:1 singly:5 repeating:3 thitimajshima:1 sl:1 exist:1 per:1 diagnosis:1 write:4 discrete:3 group:4 four:4 nevertheless:2 drawn:2 clarity:1 preprocessed:1 neither:1 breadth:4 utilize:1 replica:3 graph:20 relaxation:3 year:1 sum:7 inverse:4 everywhere:2 powerful:1 uncertainty:1 soda:1 reasonable:1 vn:1 decision:1 dy:1 forney:1 networksfor:1 scaling:1 sfm:1 bound:10 hi:6 layer:10 guaranteed:3 completing:1 hom:1 cheng:2 fan:1 quadratic:1 turbo:8 strength:1 scanned:2 scene:1 aspect:1 speed:1 mimj:1 extremely:2 performing:1 relatively:2 conjecture:1 department:2 overconfident:2 poor:1 describes:1 smaller:2 sollich:4 slightly:1 beneficial:1 em:4 making:2 s1:2 taken:1 computationally:1 equation:13 remains:2 describing:1 discus:1 tractable:3 wii:2 gaussians:3 eight:1 llll:1 appropriate:2 alternative:1 yair:1 slower:2 existence:1 original:4 assumes:1 running:3 denotes:1 include:1 completed:1 graphical:25 remaining:1 maintaining:1 marginalized:1 top:2 calculating:5 neglect:1 build:1 approximating:4 society:1 sweep:3 objective:1 added:1 quantity:5 occurs:1 dependence:1 md:3 diagonal:12 cycling:1 distance:1 thank:1 sci:2 barber:5 considers:1 trivial:1 marcus:1 rjk:1 assuming:4 maximising:1 code:9 o1:1 relationship:3 ratio:1 unfortunately:1 potentially:2 relate:1 statement:1 broadway:2 nijmegen:1 negative:1 implementation:1 xll:1 diamond:2 perform:2 observation:7 markov:1 pasztor:1 ilhj:2 finite:1 defining:1 extended:1 communication:4 precise:1 bert:1 arbitrary:12 david:1 introduced:1 namely:1 required:1 struck:1 kl:2 connection:1 specified:1 pearl:5 nip:2 address:1 beyond:1 parental:8 able:1 pattern:2 max:8 including:2 explanation:1 belief:69 royal:1 suitable:1 difficulty:2 advanced:1 scheme:4 improve:2 aston:1 factorises:1 gf:7 sn:3 understanding:1 strang:1 loss:1 oy:1 interesting:1 integrate:1 sufficient:3 consistent:1 exciting:1 editor:5 principle:1 unwrapping:2 pi:1 row:2 last:6 parity:1 arriving:1 free:2 bias:7 weaker:1 understand:1 wide:1 neighbor:4 taking:1 saul:4 sparse:1 van:2 distributed:1 calculated:8 dimension:1 llh:1 forward:3 made:5 replicated:5 san:1 ec:1 correlate:1 sj:3 approximate:7 obtains:1 global:1 assumed:1 xi:9 continuous:1 iterative:4 compromised:1 sk:1 why:1 nature:1 channel:1 mj:5 ca:1 kschischang:2 reasonably:1 gharamani:1 excellent:1 complex:1 electric:1 substituted:2 main:1 csd:1 arise:1 repeated:3 quadrature:3 body:1 fig:11 tl:1 en:1 precision:1 trellis:1 wish:2 rjj:1 xl:8 lie:1 clamped:4 third:1 ian:1 formula:4 specific:1 er:2 symbol:1 evidence:8 intractable:4 workshop:1 effectively:2 phd:1 i7j:1 sigmoids:1 suited:1 simply:2 jmj:1 gallager:1 absorbed:1 partially:1 applies:3 cowell:1 springer:1 corresponds:1 dh:1 ma:2 stimulating:1 conditional:23 glavieux:1 king:1 cz1:1 change:1 determined:3 except:2 uniformly:1 justify:1 wiegerinck:2 kearns:3 castillo:1 experimental:1 shannon:2 ucb:1 formally:1 college:1 support:1 scan:1 evaluate:2 correlated:1
704
1,644
Generalized Model Selection For Unsupervised Learning In High Dimensions Shivakumar Vaithyanathan IBM Almaden Research Center 650 Harry Road San Jose, CA 95136 Shiv@almaden.ibm.com Byron Dom IBM Almaden Research Center 650 Harry Road San Jose, CA 95136 dom@almaden.ibm.com Abstract We describe a Bayesian approach to model selection in unsupervised learning that determines both the feature set and the number of clusters. We then evaluate this scheme (based on marginal likelihood) and one based on cross-validated likelihood. For the Bayesian scheme we derive a closed-form solution of the marginal likelihood by assuming appropriate forms of the likelihood function and prior. Extensive experiments compare these approaches and all results are verified by comparison against ground truth. In these experiments the Bayesian scheme using our objective function gave better results than cross-validation. 1 Introduction Recent efforts define the model selection problem as one of estimating the number of clusters[ 10, 17]. It is easy to see, particularly in applications with large number of features, that various choices of feature subsets will reveal different structures underlying the data. It is our contention that this interplay between the feature subset and the number of clusters is essential to provide appropriate views of the data.We thus define the problem of model selection in clustering as selecting both the number of clusters and the feature subset. Towards this end we propose a unified objective function whose arguments include the both the feature space and number of clusters. We then describe two approaches to model selection using this objective function. The first approach is based on a Bayesian scheme using the marginal likelihood for model selection. The second approach is based on a scheme using cross-validated likelihood. In section 3 we apply these approaches to document clustering by making assumptions about the document generation model. Further, for the Bayesian approach we derive a closed-form solution for the marginal likelihood using this document generation model. We also describe a heuristic for initial feature selection based on the distributional clustering of terms. Section 5 describes the experiments and our approach to validate the proposed models and algorithms. Section 6 reports and discusses the results of our experiments and finally section 7 provides directions for future work. 971 Model Selection for Unsupervised Learning in High Dimensions 2 Model selection in clustering Model selection approaches in clustering have primarily concentrated on determining the number of components/clusters. These attempts include Bayesian approaches [7,10], MDL approaches [15] and cross-validation techniques [17] . As noticed in [17] however, the optimal number of clusters is dependent on the feature space in which the clustering is performed. Related work has been described in [7]. 2.1 A generalized model for clustering Let D be a data-set consisting of "patterns" {d I, .. , d v }, which we assume to be represented in some feature space T with dimension M. The particular problem we address is that of clustering D into groups such that its likelihood described by a probability model p(DTIQ), is maximized, where DT indicates the representation of D in feature space T and Q is the structure of the model, which consists of the number of clusters, the partitioning of the feature set (explained below) and the assignment of patterns to clusters. This model is a weighted sum of models {P(DTIQ, ~)I~ E [Rm} where ~ is the set of all parameters associated with Q . To define our model we begin by assuming that the feature space T consists of two sets: U - useful features and N noise features. Our feature-selection problem will thus consist of partitioning T (into U and N) for a given number of clusters. Assumption 1 The feature sets represented by independent p(DTIQ,~) U and N are conditionally = P(D N I Q, ~) P(D u I Q,~) (1) where DN indicates data represented in the noise feature space and D U indicates data represented in useful feature space. Using assumption 1 and assuming that the data is independently drawn, we can rewrite equation (1) as p(DTIQ,~) = {n p(d~ I ~N). 1=1 nn k=IJED, p(dy I ~f)} (2) where V is the number of patterns in D, p(dy I ~u) is the probability of dy given the parameter vector ~f and p(d~ I ~N) is the probability of d~ given the parameter vector ~N . Note that while the explicit dependence on Q has been removed in this notation , it is implicit in the number of clusters K and the partition of T into Nand U. 2.2 Bayesian approach to model selection The objective function, represented in equation (2) is not regularized and attempts to optimize it directly may result in the set N becoming empty - resulting in overfitting. To overcome this problem we use the marginallikelihood[2]. Assumption 2 All parameter vectors are independent. n (~) =n (~N). K n n (~f) k=1 where the n( ... ) denotes a Bayesian prior distribution. The marginal likelihood, using assumption 2, can be written as P(D T I Q)= IN [UP(d~ I ~N)]n(~N)d~N. DL [!lp(d Y I ~f)]n(~f)d~f(3) S. Vaithyanathan and B. Dom 972 where SN, SV are integral limits appropriate to the particular parameter spaces. These will be omitted to simplify the notation. 3.0 Document clustering Document clustering algorithms typically start by representing the document as a "bag-of-words" in which the features can number - 10 4 to 10 5 . Ad-hoc dimensionality reduction techniques such as stop-word removal, frequency based truncations [16] and techniques such as LSI [5] are available. Once the dimensionality has been reduced, the documents are usually clustered into an arbitrary number of clusters. 3.1 Multinomial models Several models of text generation have been studied[3]. Our choice is multinomial models using term counts as the features. This choice introduces another parameter indicating the probability of the Nand U split. This is equivalent to assuming a generation model where for each document the number of noise and useful terms are determined by a probability (}s and then the terms in a document are "drawn" with a probability ((}n or ()~ ). 3.2 Marginal likelihood / stochastic complexity To apply our Bayesian objective function we begin by substituting multinomial models into (3) and simplifying to obtain P(D I Q) = (t N;,t V ) S[((}S)tN (1- (}s)t u]n((}s)d(}S . [Ii: II ({t. tIYEU}J] S[II((}k)t,.u]n((}f)d(}f ? k==1 IEDk I,UU UEV (4) ( tf J1S[II((}n)ti .? ] n((}N) d(}N [iI j=1 {tj,nlnEN} nEN where ( (.'.\) is the multinomial coefficient, ti,u is the number of occurrences of the feature term document i u in document i, tYis the total number of all useful features (terms) in (tY =L U ti,u, t~:, and ti,n are to be interpreted similar to above but for (n noise features, = k l(~~k) ! , tNis the total number of all noise features in all patterns and tVis the total number of all useful features in all patterns. To solve (4) we still need a form for the priors {n( ... )}. The Beta family is conjugate to the Binomial family [2] and we choose the Dirichlet distribution (mUltiple Beta) as the form for both n((}f) and n((}N) and the Beta distribution for n((}s). Substituting these into equation (8) and simplifying yields P (D I Q) =[ f(Ya + Yb) f(tN + Ya)f(tV + Yb) ] ? [ f(/J) II f(/Jn + t n) ] f(Ya)f(Yb) [(tV + t N + Ya + Yb) f(/J + t N) nEN f(/Jn) [ [(0') [(0'+ v) K r(O'k + IDkl)] [K f(a) [(au + tV ] D f(lD kl) ? Df(a+ Du f(a u) tU(k) (5) Model Selection for Unsupervised Learning in High Dimensions 973 where f3, and au are the hyper-parameters of the Dirichlet prior for noise and useful features respectively, f3 = f3n , a = au, U = ukand is the "gamma" function. L L UEU neN ro L k=1 Further, Yu, Yure the hyper parameters of the Beta prior for the split probability, IDkl is the number of documents in cluster k and tU(k is computed as L tf. The results iEDk reported for our evaluation will be the negative of the log of equation (5), which (following Rissanen [14]) we refer to as Stochastic Complexity (SC). In our experiments all values of the hyper-parameters pj ,aj (Jk> Ya and Ybare set equal to 1 yielding uniform priors. 3.3 Cross-Validated likelihood To compute the cross validated likelihood using multinomial models we first substitute the multinomial functional forms, using the MLE found using the training set. This results in the following equation I QP) = ,......., N ,.....,., U VIt!\ 1 ,.....,.., K ~ [(05)t" .. (1- ( 5)1,,] IT p(evf ION) . IT IT peevy I O~i)' p(q) (6) 1=1 k=IJEDk ,..., ,..., where Os, ON and O~i) are the MLE of the appropriate parameter vectors . For our implementation of MCCV, following the suggestion in [17], we have used a 50% split of the training and test set. For the vCV criterion although a value of v = 10 was suggested therein, for computational reasons we have used a value of v = 5. P(CVT --- { 3.4 Feature subset selection algorithm for document clustering As noted in section 2.1, for a feature-set of size M there are a total of 2M partitions and for large M it would be computationally intractable to search through all possible partitions to find the optimal subset. In this section we propose a heuristic method to obtain a subset of tokens that are topical (indicative of underlying topics) and can be used as features in the bag-of-words model to cluster documents. 3.4.1 Distributional clustering for feature subset selection Identifying content-bearing and topical terms, is an active research area [9]. We are less concerned with modeling the exact distributions of individual terms as we are with simply identifying groups of terms that are topical. Distributional clustering (DC), apparently first proposed by Pereira et al [13], has been used for feature selection in supervised text classification [1] and clustering images in video sequences [9]. We hypothesize that function, content-bearing and topical terms have different distributions over the documents. DC helps reduce the size of the search space for feature selection from 2M to 2 e , where C is the number of clusters produced by the DC algorithm. Following the suggestions in [9], we compute the following histogram for each token. The first bin consists of the number of documents with zero occurrences of the token, the second bin is the number of documents consisting of a single occurrence of the token and the third bin is the number of documents that contain more two or more occurrences of the term. The histograms are clustered using reLative entropy ~(. II .) as S. Vaithyanathan and B. Dom 974 a distance measure. For two terms with probability distributions PI (.) and P2(.), this is given by [4]: ,1.(Pt(t) II P2(t)) = kt '" PI(t) log PI(t) P2(t) (7) We use a k-means-style algorithm in which the histograms are normalized to sum to one and the sum in equation (7) is taken over the three bins corresponding to counts of 0,1, and ~ 2. During the assignment-to-clusters step of k-means we compute II PCk) (where pw is the normalized histogram for term wand Pq(t) is the centroid of cluster k) and the term w is assigned to the cluster for which this is minimum [13,8]. ,1.(pw 4.0 Experimental setup Our evaluation experiments compared the clustering results against human-labeled ground truth. The corpus used was the AP Reuters Newswire articles from the TREC-6 collection. A total of 8235 documents, from the routing track, existing in 25 classes were analyzed in our experiments. To simplify matters we disregarded multiple assignments and retained each document as a member of a single class. 4.1 Mutual information as an evaluation measure of clustering We verify our models by comparing our clustering results against pre-classified text. We force all clustering algorithms to produce exactly as many clusters as there are classes in the pre-classified text and we report the mutual information[ 4] (MI) between the cluster labels and pre-classified class labels 5.0 Results and discussions After tokenizing the documents and discarding terms that appeared in less than 3 documents we were left with 32450 unique terms . We experimented with several numbers of clusters for DC but report only the best (lowest SC) for lack of space. For each of these clusters we chose the best of 20 runs corresponding to different random starting clusters. Each of these sets includes one cluster that consists of high-frequency words and upon examination were found to contain primarily function words, which we eliminated from further consideration. The remaining non-function-word clusters were used as feature sets for the clustering algorithm. Only combinations of feature sets that produced good results were used for further document clustering runs. We initialized the EM algorithm using k-means algorithm - other initialization schemes are discussed in [11]. The feature vectors used in this k-means initialization were generated using the pivoted normal weighting suggested in [16]. All parameter vectors and were estimated using Laplace's Rule of Succession[2]. Table 1 shows the best results of the SC criterion, the vCV and MCCV using the feature subsets selected by the different combinations of distributional clusters. The feature subsets are coded as FSXP where X indicates the number of clusters in the distributional clustering and P indicates the cluster number(s) used as U. For SC and MI all results reported are averages over 3 runs of the k-means+EM combination with different initialization fo k-means. For clarity, the MI numbers reported are normalized such that the theoretical maximum is 1.0. We also show comparisons against no feature selection (NF) and LSI. Of eN Model Selection for Unsupervised Learning in High Dimensions 975 For LSI, the principal 165 eigenvectors were retained and k-means clustering was performed in the reduced dimensional space. While determining the number of clusters, for computational reasons we have limited our evaluation to only the feature subset that provided us with the highest MI, i.e., FS41-3 . Feature Set FS41-3 FS52 NF LSI Useful Features 6,157 386 32,450 324501165 Table SC X 107 2.66 2.8 2.96 NA 1 Comparison vCV X 107 0.61 0.3 1.25 NA Of Results MCCV X 107 1.32 0.69 2.8 NA Ml 0.61 0.51 0.58 0.57 Figure 1 1~1.. . ? , ...... .: . " '" .. . ,.. .. .?" ....,....... Log ....IQ..... L...hood ThIwI:_P'!Ch I, 1~1.. ? . Flgur.2 ~ . " . ? ... MCCV? .... gII ..... l avl .......-ood I " 5.3 Discussion The consistency between the MI and SC (Figure 1) is striking. The monotonic trend is more apparent at higher SC indicating that bad clusterings are more easily detected by SC while as the solution improves the differences are more subtle. Note that the best value of SC and Ml coincide. Given the assumptions made in deriving equation (5), this consistency and is encouraging. The interested reader is referred to [18] for more details. Figures 2 and 3 indicate that there is certainly a reasonable consistency between the cross-validated likelihood and the MI although not as striking as the SC. Note that the MI for the feature sets picked by MCCV and vCV is significantly lower than that of the best feature-set. Figures 4,5 and 6 show the plots of SC, MCCV and vCV as the number of clusters is increased. Using SC we see that FS41-3 reveals an optimal structure around 40 clusters. As with feature selection, both MCCV and vCV obtain models of lower complexity than Sc. Both show an optimum of about 30 clusters. More experiments are required before we draw final conclusions, however, the full Bayesian approach seems a practical and useful approach for model selection in document clustering. Our choice of likelihood function and priors provide a closed-form solution that is computationally tractable and provides meaningful results. 6.0 Conclusions In this paper we tackled the problem of model structure determination in clustering. The main contribution of the paper is a Bayesian objective function that treats optimal model selection as choosing both the number of clusters and the feature subset. An important aspect of our work is a formal notion that forms a basis for doing feature selection in unsupervised learning. We then evaluated two approaches for model selection: one using this objective function and the other based on cross-validation. S. Vaithyanathan and B. Dom 976 Both approaches performed reasonably well - with the Bayesian scheme outperforming the cross-validation approaches in feature selection. More experiments using different parameter settings for the cross-validation schemes and different priors for the Bayesian scheme should result in better understanding and therefore more powerful applications of these approaches . Fig.... :: I . """. ! I te. 1: '--"_ _ _ _------' t. t WCCV ? ,..,._lIogLDiIfIood Xl01 I . ....... I la~ 1-'-_ _ __ r:: _ ----1 ..... ~.. .? .. ? .. ? - + ? ?. ' , tl. " _ _ 0.-. ? ~ ?? ",._ Flgur,' ... " : 1=' ... ~ " HI " 1 I ? . . ? ? _oliO-. . ?. .I OM ~ OM ? . ? __ "0..-. References [I] Baker, D., et aI, Distributional Clustering of Words for Text Classification, SIGIR 1998. [2] Bernardo, J. M. and Smith, A. F. M., Bayesian Theory, Wiley, 1994. [3] Church, K.W. et aI, Poisson Mixtures. Natural Language Engineering. 1(12), 1995. [4] Cover, T.M . and Thomas, J.A. Elements of Information Theory. Wiley-Interscience, 1991. [5] Deerwester,S. et aI, Indexing by Latent Semantic Analysis,JASIS, 1990. [6] Dempster, A.et aI., Maximum Likelihood from Incomplete Data Via the EM Algorithm. JRSS, 39,1977. [7] Hanson,R., et aI, Bayesian Classification with Correlation and Inheritance, IJCAI,1991 . [8] Iyengar, G., Clustering images using relative entropy for efficient retrieval, VLBV, 1998. [9] Katz, S.M . , Distribution of content words and phrases in text and language modeling, NLE, 2,1996. [10] Kontkanen, P.T. et ai, Comparing Bayesian Model Class Selection Criteria by Discrete Finite Mixtures, ISIS'96 Conference, 1996. [II] Meila, M., Heckerman, D., An Experimental Comparison of Several Clustering and Initialization Methods, MSR-TR-98-06. [12] Nigam, K et aI, Learning to Classify Text from Labeled and Unlabeled Documents, AAAI, 1998. [13] Pereira, F.C.N. et ai , Distributional clustering of English words, ACL,1993. [14] Rissanen, J., Stochastic Complexity in Statistical Inquiry. World\ Scientific, 1989. [15] Rissanen, J., Ristad E., Unsupervised classification with stochastic complexity." The US/Japan Conference on the Frontiers of Statistical Modeling,1992. [16] Singhal A. et aI, Pivoted Document Length Normalization, SIGIR, 1996. [17] Smyth, P., Clustering using Monte Carlo cross-validation, KDD, 1996. [18] Vaithyanathan, S. and Dom, B. Model Selection in Unsupervised Learning with Applications to Document Clustering. IBM Research Report RJ-I 0137 (95012) Dec. 14, 1998 .
1644 |@word uev:1 msr:1 pw:2 seems:1 simplifying:2 tr:1 ld:1 reduction:1 initial:1 selecting:1 document:27 existing:1 com:2 comparing:2 written:1 partition:3 kdd:1 hypothesize:1 plot:1 selected:1 indicative:1 smith:1 provides:2 dn:1 beta:4 ood:1 consists:4 interscience:1 isi:1 encouraging:1 begin:2 estimating:1 underlying:2 notation:2 provided:1 baker:1 lowest:1 interpreted:1 unified:1 nf:2 ti:4 bernardo:1 exactly:1 ro:1 rm:1 partitioning:2 before:1 engineering:1 treat:1 limit:1 becoming:1 ap:1 chose:1 acl:1 au:3 studied:1 therein:1 initialization:4 limited:1 unique:1 practical:1 hood:1 area:1 significantly:1 word:9 road:2 pre:3 unlabeled:1 selection:27 optimize:1 equivalent:1 center:2 cvt:1 starting:1 independently:1 vit:1 sigir:2 identifying:2 rule:1 deriving:1 notion:1 laplace:1 pt:1 exact:1 smyth:1 trend:1 element:1 particularly:1 jk:1 distributional:7 labeled:2 removed:1 highest:1 dempster:1 complexity:5 dom:6 rewrite:1 ristad:1 idkl:2 upon:1 basis:1 easily:1 various:1 represented:5 describe:3 monte:1 detected:1 sc:13 hyper:3 choosing:1 whose:1 heuristic:2 apparent:1 solve:1 final:1 interplay:1 hoc:1 sequence:1 propose:2 tu:2 shiv:1 validate:1 ijcai:1 cluster:33 empty:1 optimum:1 produce:1 help:1 derive:2 iq:1 p2:3 indicate:1 uu:1 direction:1 stochastic:4 human:1 routing:1 bin:4 clustered:2 frontier:1 around:1 ground:2 normal:1 substituting:2 omitted:1 pivoted:2 bag:2 label:2 tf:2 weighted:1 iyengar:1 validated:5 likelihood:15 indicates:5 centroid:1 dependent:1 nn:1 typically:1 nand:2 interested:1 classification:4 almaden:4 mutual:2 marginal:6 equal:1 once:1 f3:2 eliminated:1 yu:1 unsupervised:8 future:1 report:4 simplify:2 primarily:2 gamma:1 individual:1 consisting:2 attempt:2 evaluation:4 certainly:1 mdl:1 introduces:1 analyzed:1 mixture:2 yielding:1 tj:1 kt:1 integral:1 incomplete:1 initialized:1 theoretical:1 increased:1 classify:1 modeling:3 cover:1 assignment:3 phrase:1 singhal:1 subset:11 uniform:1 reported:3 sv:1 na:3 aaai:1 choose:1 style:1 japan:1 harry:2 includes:1 coefficient:1 matter:1 ad:1 performed:3 view:1 picked:1 closed:3 apparently:1 doing:1 vcv:6 start:1 contribution:1 om:2 maximized:1 yield:1 succession:1 bayesian:16 produced:2 carlo:1 classified:3 inquiry:1 fo:1 against:4 ty:1 frequency:2 associated:1 mi:7 stop:1 dimensionality:2 improves:1 subtle:1 higher:1 dt:1 supervised:1 yb:4 evaluated:1 implicit:1 correlation:1 o:1 lack:1 aj:1 reveal:1 scientific:1 contain:2 normalized:3 verify:1 assigned:1 semantic:1 conditionally:1 during:1 noted:1 criterion:3 generalized:2 pck:1 tn:2 image:2 consideration:1 contention:1 multinomial:6 functional:1 qp:1 discussed:1 katz:1 refer:1 ai:9 meila:1 consistency:3 newswire:1 language:2 pq:1 recent:1 outperforming:1 minimum:1 ii:10 multiple:2 full:1 rj:1 kontkanen:1 determination:1 cross:11 retrieval:1 mle:2 coded:1 df:1 poisson:1 histogram:4 normalization:1 ion:1 dec:1 gii:1 byron:1 member:1 split:3 easy:1 concerned:1 gave:1 reduce:1 effort:1 useful:8 eigenvectors:1 concentrated:1 reduced:2 lsi:4 estimated:1 track:1 discrete:1 group:2 rissanen:3 drawn:2 clarity:1 pj:1 verified:1 sum:3 deerwester:1 wand:1 run:3 jose:2 powerful:1 striking:2 family:2 reader:1 reasonable:1 draw:1 dy:3 hi:1 tackled:1 shivakumar:1 aspect:1 argument:1 tv:3 combination:3 conjugate:1 jr:1 describes:1 heckerman:1 em:3 lp:1 making:1 explained:1 indexing:1 taken:1 computationally:2 equation:7 discus:1 count:2 tractable:1 end:1 available:1 nen:3 apply:2 appropriate:4 occurrence:4 jn:2 substitute:1 thomas:1 denotes:1 binomial:1 dirichlet:2 clustering:31 include:2 remaining:1 objective:7 noticed:1 dependence:1 distance:1 topic:1 reason:2 assuming:4 length:1 retained:2 setup:1 negative:1 implementation:1 finite:1 topical:4 dc:4 trec:1 arbitrary:1 required:1 kl:1 extensive:1 hanson:1 address:1 suggested:2 below:1 pattern:5 usually:1 appeared:1 video:1 natural:1 force:1 regularized:1 examination:1 jasis:1 representing:1 scheme:9 church:1 sn:1 text:7 prior:8 understanding:1 inheritance:1 removal:1 determining:2 relative:2 generation:4 suggestion:2 validation:6 article:1 pi:3 ibm:5 token:4 truncation:1 english:1 formal:1 overcome:1 dimension:5 world:1 avl:1 collection:1 made:1 san:2 coincide:1 ml:2 overfitting:1 active:1 reveals:1 corpus:1 search:2 latent:1 table:2 reasonably:1 ca:2 nigam:1 du:1 bearing:2 main:1 reuters:1 noise:6 fig:1 referred:1 en:1 tl:1 wiley:2 ueu:1 explicit:1 pereira:2 third:1 weighting:1 tokenizing:1 vaithyanathan:5 discarding:1 bad:1 experimented:1 dl:1 essential:1 consist:1 intractable:1 te:1 disregarded:1 entropy:2 simply:1 monotonic:1 ch:1 truth:2 determines:1 towards:1 content:3 determined:1 principal:1 f3n:1 total:5 experimental:2 ya:5 la:1 meaningful:1 indicating:2 evaluate:1
705
1,645
On input selection with reversible jump Markov chain Monte Carlo sampling Peter Sykacek Austrian Research Institute for Artificial Intelligence (OFAI) Schottengasse 3, A-10lO Vienna, Austria peter@ai. univie. ac. at Abstract In this paper we will treat input selection for a radial basis function (RBF) like classifier within a Bayesian framework. We approximate the a-posteriori distribution over both model coefficients and input subsets by samples drawn with Gibbs updates and reversible jump moves. Using some public datasets, we compare the classification accuracy of the method with a conventional ARD scheme. These datasets are also used to infer the a-posteriori probabilities of different input subsets. 1 Introduction Methods that aim to determine relevance of inputs have always interested researchers in various communities. Classical feature subset selection techniques, as reviewed in [1], use search algorithms and evaluation criteria to determine one optimal subset. Although these approaches can improve classification accuracy, they do not explore different equally probable subsets. Automatic relevance determination (ARD) is another approach which determines relevance of inputs. ARD is due to [6] who uses Bayesian techniques, where hierarchical priors penalize irrelevant inputs. Our approach is also "Bayesian": Relevance of inputs is measured by a probability distribution over all possible feature subsets. This probability measure is determined by the Bayesian evidence of the corresponding models. The general idea was already used in [7] for variable selection in linear regression mo.dels. Though our interest is different as we select inputs for a nonlinear classification model. We want an approximation of the true distribution over all different subsets. As the number of subsets grows exponentially with the total number of inputs, we can not calculate Bayesian model evidence directly. We need a method that samples efficiently across different dimensional parameter spaces. The most general method that can do this is the reversible jump Markov chain Monte Carlo sampler (reversible jump Me) recently proposed in [4]. The approach was successfully applied by [8] to determine a probability distribution in a mixture density model with variable number of kernels and in [5] to sample from the posterior of RBF regression networks with variable number of kernels. A Markov chain that switches between different input subsets is useful for two tasks: Counting how often a particular subset was visited gives us a relevance measure of the corresponding inputs; For classification, we approximate On Input Selection with Reversible Jump MCMC 639 the integral over input sets and coefficients by summation over samples from the Markov chain. The next sections will show how to implement such a reversible jump MC and apply the proposed algorithm to classification and input evaluation using some public datasets. Though the approach could not improve the MLP-ARD scheme from [6] in terms of classification accuracy, we still think that it is interesting: We can assess the importance of different feature subsets which is different than importance of single features as estimated by ARD. 2 Methods The classifier used in this paper is a RBF like model. Inference is performed within a Bayesian framework. When conditioning on one set of inputs , the posterior over model parameters is already multimodal. Therefore we resort to Markov chain Monte Carlo (MCMC) -sampling techniques to approximate the desired posterior over both model coefficients and feature subsets. In the next subsections we will propose an appropriate architecture for the classifier and a hybrid sampler for model inference. This hybrid sampler consists of two parts: We use Gibbs updates ([2]) to sample when conditioning on a particular set of inputs and reversible jump moves that carry out dimension switching updates. 2.1 The classifier I~ order to allow input relevance determination by Bayesian model selection , the classifier needs at least one coefficient that is associated with each input: Roughly speaking, the probability of each model is proportional to the likelihood of the most probable coefficients, weighted by their posterior width divided by their prior width. The first factor always increases when using more coefficients (or input features). The second will decrease the more inputs we use and together this gives a peak for the most probable model. A classifier that satisfies these constraints is the so called classification in the sampling paradigm. We model class conditional densities and together with class priors express posterior probabilities for classes. In neural network literature this approach was first proposed in [10). We use a model that allows for overlapping class conditional densities: D p(~lk) =L d=l K WkdP(~I~) , p(~) =L PkP(~lk) (1) k=l Using Pk for the J{ class priors and p(~lk) for the class conditional densities, (1) expresses posterior probabj,Jities for classes as P(kl~) = PkP(~lk)/p(~). We choose the component densities, p(~IcI> d), to be Gaussian with restricted parametrisation: Each kernel is a multivariate normal distribution with a mean and a diagonal covariance matrix. For all Gaussian kernels together, we get 2 * D * I parameters, with I denoting the current input dimension and D denoting the number of kernels. Apart from kernel coefficients, cI>d , (1) has D coefficients per class, Wkd, indicating the prior kernel allocation probabilities and J{ class priors. Model (1) allows to treat labels of patterns as missing data and use labeled as well as unlabeled data for model inference. In this case training is carried out using the likelihood of observing inputs and targets: p(T, X18) = rrr;=lrr;::=lPkPk(~nk Ifu)rr~=lp(bnI8) , (2) where T denotes labeled and X unlabeled training data. In (2) 8 k are all coefficients the k-th class conditional density depends on. We further use 8 for all model P. Sykacek 640 coefficients together, nk as number of samples belonging to class k and m as index for unlabeled samples. To make Gibbs updates possible, we further introduce two latent allocation variables. The first one, d, indicates the kernel number each sample was generated from, the second one is the unobserved class label c, introduced for unlabeled data. Typical approaches for training models like (1), e.g. [3] and [9], use the EM algorithm, which is closely related to the Gibbs sampler introduce in the next subsection. 2.2 Fixed dimension sampling In this subsection we will formulate Gibbs updates for sampling from the posterior when conditioning on a fixed set of inputs. In order to allow sampling from the full conditional, we have to choose priors over coefficients from their conjugate family: ? Each component mean, !!!d, is given a Gaussian prior: !!!d '" Nd({di). ? The inverse variance of input i and kernel d gets a Gamma prior: u;;l '" r( a, ,Bi). ? All d variances of input i have a common hyperparameter, ,Bi, that has itself a Gamma hyperprior: ,Bi ,...., r(g, hi). ? The mixing coefficients, ~, get a Dirichlet prior: ~ '" 1J (6w , ... , 6w ). ? Class priors, P, also get a Dirichlet prior: P '" 1J(6p , ... ,6p). The quantitative settings are similar to those used in [8]: Values for a are between 1 and 2, g is usually between 0.2 and 1 and hi is typically between 1/ Rr and 10/ with Ri denoting the i'th input range. The mean gets a Gaussian prior centered at the midpoint, with diagonal inverse covariance matrix ~, with "'ii = 1/ The prior counts d w and 6p are set to 1 to give the corresponding probabilities non-informative proper Dirichlet priors. Rr, Rr. e, The Gibbs sampler uses updates from the full conditional distributions in (3). For notational convenience we use ~ for the parameters that determine class conditional densities. We use m as index over unlabeled data and Cm as latent class label. The index for all data is n, dn are the latent kernel allocations and nd the number of samples allocated by the d-th component. One distribution does not occur in the prior specification. That is Mn(l, ... ) which is a multinomial-one distribution. Finally we need some counters: ml ... mK are the counts per class and mlk .. mDk count kernel allocations of class-k-patterns. The full conditional of the d-th kernel variances and the hyper parameter ,Bi contain i as index of the input dimension. There we express each u;J separately. In the expression of the d-th kernel mean, I 641 On Input Selection with Reversible Jump MCMC illd, we use .lGt to denote the entire covariance matrix. PkP(~mlfu) Mn ( 1, { I:k PkP(~mlfu)' Mn (1, {I:,WtndP(~nl~) Wt,.dP(~nl~) }) k= l..K ,d= l..D}) p(~J ..) r (9 + p(~I???) 1) (ow + mlk, ... ,ow + mDk) p(PI???) p(illdl???) 1) (op r (a + ~d, + ~ Vnld,.=d L (~n,i -llid,i)2) Da. hi (3) + ;; ud,! ) + ml, ... , op + mK) N ((nd~l + ~)-l(ndVdl~ + ~S), (ndVd 1 + ~)-l) f3i i?,. 2.3 Moving between different input subsets The core part of this sampler are reversible jump updates, where we move between different feature subsets. The probability of a feature subset will be determined by the corresponding Bayesian model evidence and by an additional prior over number of inputs. In accordance with [7J, we use the truncated Poisson prior: p(I) = 1/ ( I jax ) c ~~ , where c is a constant and Imax the total nr. of inputs. Reversible jump updates are generalizations of conventional Metropolis-Hastings updates, where moves are bijections (x, u) H (x', u'). For a thorough treatment we refer to [4J. In order to switch subsets efficiently, we will use two different types of moves. The first consist of a step where we add one input chosen at random and a matching step that removes one randomly chosen input. A second move exchanges two inputs which allows "tunneling" through low likelihood areas. Adding an input, we have to increase the dimension of all kernel means and diagonal covariances. These coefficients are drawn from their priors. In addition the move proposes new allocation probabilities in a semi deterministic way. Assuming the ordering, Wk,d ~ Wk,d+1: op Vd ~ D/2 Beta(b a , bb + 1) = Wk,D+l-d + Wk ,dOp { W~'D+l-d w~ , d = wk ,d(1 - op) (4) The matching step proposes removing a randomly chosen input. Removing corresponding kernel coefficients is again combined with a semi deterministic proposal of new allocation probabilities, which is exactly symmetric to the proposal in (4). 642 P. Sykacek Table 1: Summary of experiments Data Ionosphere Pima Wine avg(#) 4.3 4 4.4 max(#) 9 7 8 We accept births with probability: n, min( 1, lh. rt x x (~':) RBF (%,n a ) (91.5,11) (78 .9,111 (100, 01 r MLP (%,nb) 95.5,4 79.8,8 96.8,2 p(;(;/) G, J2,; g f g(.,.~ -')"-1 exp( -Ii' exp ( - 05 ;" (I'd - <d)' ) "'~-') x dm / (I + 1) x 1 bm/(Imax - I) (~, V27i) D TID exp ( -0.5Ih(J.l~ - ed)2) x 1 (Ii;)) D TID (0"~-2)a-l exp( _(3'0"~-2) ). (5) The first line in (5) are the likelihood and prior ratio. The prior ratio results from the difference in input dimension, which affects the kernel means and the prior over number of inputs. The first term of the proposal ratio is from proposing to add or remove one input. The second term is the proposal density of the additional kernel components which cancels with the corresponding term in the prior ratio. Due to symmetry of the proposal (4) and its reverse in a death move, there is no contribution from changing allocation probabilities. Death moves are accepted with probability ad = l/ab. The second type of move is an exchange move. We select a new input and one from the model inputs and propose new mean coefficients. This gives the following acceptance probability: (*,J2;) D TID exp ( -0.5"Jb(J.ld - ed)2) min( 1, lh. ratio x (~, J2;) D ITD exp ( -0.5 Ih(J.ld - ed)2) (6) cm/ I TID N(J.ldl???) ) x---:-:--'-----..,..x cm/(Imax - I) TID N(J.l~I???) . The first line of (6) are again likelihood and prior ratio. For exchange moves, the prior ratio is just the ratio from different values in the kernel means. The first term in the proposal ratio is from proposing to exchange an input. The second term is the proposal density of new kernel mean components. The last part is from proposing new allocation probabilities. 3 Experiments Although the method can be used with labeled and unlabeled data, the following experiments were performed using only labeled data. For all experiments we set a = 2 and 9 = 0.2. The first two data sets are from the VCI repositoryl. We use 1 Available at http://www.ics.uci.edu/ mlearn/MLRepository.html. On Input Selection with Reversible Jump MCMC 643 the Ionosphere data which has 33 inputs, 175 training and 176 test samples. For this experiment we use 6 kernels and set h = 0.5. The second data is the wine recognition data which provides 13 inputs, 62 training and 63 test samples. For this data, we use 3 kernels and set h = 0.28. The third experiment is performed with the Pima data provided by B. D . Ripley2. For this one we use 3 kernels and set h = 0.16. For all experiments we draw 15000 samples from the posterior over coefficients and input subsets. We discard the first 5000 samples as burn in and use the rest for predictions. Classification accuracy, is compared with an MLP classifier using R. Neals hybrid Monte Carlo sampling with ARD priors on inputs. These experiments use 25 hidden units. Table 1 contains further details: avg( #) is the average and max(#) the maximal number of inputs used by the hybrid sampler; RBF (%, na) is the classification accuracy of the hybrid sampler and the number of errors it made that were not made by the ARD-MLP; MLP(%, nb) is the same for the ARD-MLP. We compare classifiers by testing (na, nb) against the null hypothesis that this is an observation from a Binomial Bn(na +nb , 0.5) distribution. This reveals that neither difference is significant. Although we could not improve classification accuracy on these data, this does not really matter because ARD methods usually lead to high generalization accuracy and we can compete. The real benefit from using the hybrid sampler is that we can infer probabilities telling us how much different subsets contribute to an explanation of the target variables. Figure 3 shows the occurrence probabilities of feature subsets and features. Note that table 1 has also details about how many features were used in these problems. Especially the results from Ionosphere data are interesting as on average we use only 4.3 out of 33 input features. For ionosphere and wine data the Markov chain visits about 500 different input subsets within 10000 samples. For the Pima data the number is about 60 and an order of magnitude smaller. 4 Discussion In this paper we have discussed a hybrid sampler that uses Gibbs updates and reversible jump moves to approximate the a-posteriori distribution over parameters and input subsets in nonlinear classification problems. The classification accuracy of the method could compete with R . Neals MLP-ARD implementation. However the real advantage of the method is that it provides us with a relevance measure of feature subsets. This allows to infer the optimal number of inputs and how many different explanations the data provides. Acknow ledgements I want to thank several people for having used resources they provide: I have used R.Neals hybrid Markov chain sampler for the MLP experiments; The data used for the experiments were obtained form the University at Irvine repository and from B. D. Ripley. Furthermore I want to express gratitude to the anonymous reviewers for their comments and to J.F.G . de Freitas for useful discussions during the conference. This work was done in the framework of the research project GZ 607.519/2V /B/9/98 "Verbesserung der Biosignalverarbeitung durch Beruecksichtigung von Unsicherheit und Konfidenz", funded by the Austrian federal ministry of science and transport (BMWV). 2 Available at http://www.stats.ox.ac.uk P. Sykacek 644 Probabil~les 10 of input subsets-Ionosphere Probabilities of inputs - Ionosphere 30r-----~------------~_, 9.1% :[7.14) 8 .6%: [3. 4) 7.9%: [4. 7) 4.6% : [4. 15) 5 20 10 o j, I. o j 100 J 1. 200 300 400 500 10 Probabilities of input subsets-Pima 17.3% : [2. 7j 16.1%:[2. 9.8% : [2. 5) 8.6% : [2. 5. 7) 15 10 20 30 Probabilities of inputs - Pima 20r-----~--------------TO 40r---~~--------~-----' 30 20 5 10 ?0L-~~~2~0~~~~40~~~-6~0 o~--~~--~----~----~ o Probabilities of input subsets - Wine 10r---~--~--~----~--~ 9.5?,,": [1. 13) 5.2%: [7. 13) 3,4?,,": [ 13) 3.1%: [1.12.13) 5 246 8 Probabilities of inputs - Wine 3Or-------~-------------, 20 10 0U-------~------~----~ 100 200 300 400 500 o 5 10 Figure 1: Probabilities of inputs and input subsets measuring their relevance. References [1] P. A. Devijver and J. V. Kittler. Pattern Recognition. A Statistical Approach. PrenticeHall, Englewood Cliffs, NJ, 1982. [2] S. Geman and D. Geman. Stochastic relaxation, gibbs distributions and the bayesian restoration of images. IEEE Trans. Pattn. Anal. Mach. Intel., 6:721-741, 1984. [3] Z. Ghahramani, M.1. Jordan Supervised Learning from Incomplete Data via an EM Approach In Cowan J.D., et al.(eds.), Advances in Neural Information Processing Systems 6, Morgan Kaufmann, Los Altos/Palo Alto/San Francisco, pp.120-127, 1994. [4] P. J. Green. Reversible jump markov chain monte carlo computation and bayesian model determination. Biometrika, 82:711-732, 1995. [5] C. C. Holmes and B. K. Mallick. Bayesian radial basis functions of variable dimension. Neural Computation, 10:1217-1234, 1998. [6] R. M. Neal. Bayesian Learning for Neural Networks. Springer, New York, 1986. [7] D. B. Phillips and A. F. M. Smith. Bayesian model comparison via jump diffusioons. In W.R. Gilks, S. Richardson, and D.J. Spiegelhalter, editors, Markov Chain Monte Carlo in Practice, pages 215-239, London, 1996. Chapman & Hall. [8] S. Richardson and P.J. Green On Bayesian Analysis of Mixtures with an unknown number of components Journal Royal Stat. Soc. B, 59:731-792, 1997. [9] M. Stensmo, T.J . Sejnowski A Mixture Model System for Medical and Machine Diagnosis In Tesauro G., et al.(eds.), Advances in Neural Information Processing System 7, MIT Press, Cambridge/Boston/London, pp.1077-1084, 1995. [10] H. G. C. Traven A neural network approach to statistical pattern classification by "semi parametric" estimation of probability density functions IEEE Trans. Neur. Net., 2:366-377, 1991.
1645 |@word repository:1 nd:3 bn:1 covariance:4 mlk:2 ld:2 carry:1 contains:1 denoting:3 freitas:1 current:1 informative:1 remove:2 update:10 intelligence:1 smith:1 core:1 provides:3 contribute:1 wkd:1 dn:1 beta:1 consists:1 introduce:2 roughly:1 provided:1 project:1 alto:2 null:1 cm:3 proposing:3 unobserved:1 nj:1 quantitative:1 thorough:1 exactly:1 biometrika:1 classifier:8 uk:1 unit:1 medical:1 tid:5 accordance:1 treat:2 switching:1 mach:1 cliff:1 pkp:4 burn:1 lrr:1 bi:4 range:1 gilks:1 testing:1 practice:1 implement:1 area:1 matching:2 radial:2 get:5 convenience:1 unlabeled:6 selection:8 nb:4 www:2 conventional:2 deterministic:2 reviewer:1 missing:1 formulate:1 stats:1 holmes:1 imax:3 target:2 us:3 hypothesis:1 recognition:2 geman:2 labeled:4 calculate:1 kittler:1 ordering:1 decrease:1 counter:1 und:1 basis:2 multimodal:1 various:1 london:2 monte:6 sejnowski:1 artificial:1 hyper:1 birth:1 richardson:2 think:1 itself:1 advantage:1 rr:4 net:1 propose:2 maximal:1 j2:3 uci:1 mixing:1 probabil:1 los:1 ac:2 stat:1 measured:1 ard:10 op:4 soc:1 closely:1 stochastic:1 centered:1 public:2 exchange:4 generalization:2 really:1 anonymous:1 probable:3 summation:1 hall:1 ic:1 normal:1 exp:6 itd:1 mo:1 wine:5 estimation:1 label:3 visited:1 palo:1 successfully:1 weighted:1 federal:1 mit:1 always:2 gaussian:4 aim:1 notational:1 likelihood:5 indicates:1 posteriori:3 inference:3 typically:1 entire:1 accept:1 hidden:1 interested:1 classification:13 html:1 proposes:2 having:1 sampling:7 chapman:1 cancel:1 jb:1 randomly:2 gamma:2 ab:1 interest:1 mlp:8 acceptance:1 englewood:1 evaluation:2 mixture:3 nl:2 chain:9 integral:1 lh:2 incomplete:1 hyperprior:1 desired:1 mk:2 measuring:1 restoration:1 subset:26 combined:1 density:10 peak:1 together:4 prenticehall:1 parametrisation:1 na:3 again:2 von:1 choose:2 ldl:1 resort:1 de:1 wk:5 coefficient:16 matter:1 depends:1 ad:1 performed:3 observing:1 neals:3 contribution:1 ass:1 accuracy:8 variance:3 who:1 efficiently:2 kaufmann:1 bayesian:14 mc:1 carlo:6 researcher:1 mlearn:1 ed:5 against:1 pp:2 dm:1 associated:1 di:1 irvine:1 treatment:1 austria:1 subsection:3 supervised:1 done:1 though:2 ox:1 furthermore:1 just:1 sykacek:4 hastings:1 transport:1 vci:1 nonlinear:2 reversible:13 overlapping:1 grows:1 contain:1 true:1 symmetric:1 death:2 neal:1 during:1 width:2 mlrepository:1 criterion:1 image:1 recently:1 common:1 multinomial:1 conditioning:3 exponentially:1 discussed:1 refer:1 significant:1 cambridge:1 gibbs:8 ai:1 phillips:1 automatic:1 funded:1 moving:1 specification:1 add:2 posterior:8 multivariate:1 irrelevant:1 apart:1 lgt:1 reverse:1 discard:1 tesauro:1 durch:1 der:1 morgan:1 ministry:1 additional:2 determine:4 paradigm:1 ud:1 ii:3 semi:3 full:3 infer:3 determination:3 divided:1 equally:1 visit:1 prediction:1 regression:2 austrian:2 poisson:1 kernel:22 penalize:1 proposal:7 addition:1 want:3 separately:1 allocated:1 rest:1 mdk:2 comment:1 cowan:1 jordan:1 counting:1 switch:2 affect:1 architecture:1 f3i:1 idea:1 expression:1 peter:2 speaking:1 schottengasse:1 york:1 useful:2 bijections:1 http:2 estimated:1 per:2 diagnosis:1 ledgements:1 hyperparameter:1 express:4 drawn:2 changing:1 neither:1 relaxation:1 compete:2 inverse:2 family:1 draw:1 tunneling:1 hi:3 occur:1 constraint:1 ri:1 min:2 neur:1 belonging:1 conjugate:1 across:1 smaller:1 em:2 lp:1 metropolis:1 rrr:1 restricted:1 resource:1 count:3 available:2 apply:1 hierarchical:1 appropriate:1 occurrence:1 denotes:1 dirichlet:3 binomial:1 vienna:1 ghahramani:1 especially:1 classical:1 move:13 already:2 parametric:1 rt:1 diagonal:3 nr:1 ow:2 dp:1 thank:1 vd:1 me:1 assuming:1 index:4 ratio:9 pima:5 acknow:1 implementation:1 anal:1 proper:1 unknown:1 observation:1 markov:9 datasets:3 truncated:1 community:1 introduced:1 gratitude:1 kl:1 trans:2 usually:2 univie:1 pattern:4 max:2 green:2 explanation:2 royal:1 mallick:1 hybrid:8 mn:3 scheme:2 improve:3 spiegelhalter:1 lk:4 carried:1 gz:1 prior:26 literature:1 interesting:2 proportional:1 allocation:8 dop:1 x18:1 editor:1 pi:1 lo:1 summary:1 last:1 allow:2 telling:1 institute:1 midpoint:1 benefit:1 dimension:7 made:2 jump:14 ici:1 avg:2 bm:1 san:1 bb:1 approximate:4 ml:2 reveals:1 francisco:1 ripley:1 search:1 latent:3 reviewed:1 table:3 symmetry:1 da:1 pk:1 intel:1 third:1 removing:2 ionosphere:6 evidence:3 consist:1 ih:2 adding:1 importance:2 ci:1 magnitude:1 nk:2 boston:1 explore:1 springer:1 devijver:1 determines:1 satisfies:1 conditional:8 rbf:5 stensmo:1 determined:2 typical:1 sampler:11 wt:1 total:2 called:1 accepted:1 indicating:1 select:2 people:1 relevance:8 mcmc:4
706
1,646
Memory Capacity of Linear vs. Nonlinear Models of Dendritic Integration Panayiota Poirazi* Biomedical Engineering Department University of Southern California Los Angeles, CA 90089 Bartlett W. Mel* Biomedical Engineering Department University of Southern California Los Angeles, CA 90089 poirazi@sc/. usc. edu mel@lnc.usc.edu Abstract Previous biophysical modeling work showed that nonlinear interactions among nearby synapses located on active dendritic trees can provide a large boost in the memory capacity of a cell (Mel, 1992a, 1992b). The aim of our present work is to quantify this boost by estimating the capacity of (1) a neuron model with passive dendritic integration where inputs are combined linearly across the entire cell followed by a single global threshold, and (2) an active dendrite model in which a threshold is applied separately to the output of each branch, and the branch subtotals are combined linearly. We focus here on the limiting case of binary-valued synaptic weights, and derive expressions which measure model capacity by estimating the number of distinct input-output functions available to both neuron types. We show that (1) the application of a fixed nonlinearity to each dendritic compartment substantially increases the model's flexibility, (2) for a neuron of realistic size, the capacity of the nonlinear cell can exceed that of the same-sized linear cell by more than an order of magnitude, and (3) the largest capacity boost occurs for cells with a relatively large number of dendritic subunits of relatively small size. We validated the analysis by empirically measuring memory capacity with randomized two-class classification problems, where a stochastic delta rule was used to train both linear and nonlinear models. We found that large capacity boosts predicted for the nonlinear dendritic model were readily achieved in practice. -http://lnc.usc.edu P. Poirazi and B. W. Mel 158 1 Introduction Both physiological evidence and connectionist theory support the notion that in the brain, memories are stored in the pattern of learned synaptic weight values. Experiments in a variety of neuronal preparations however, inQicate that the efficacy of synaptic transmission can undergo substantial fluctuations up or down, or both, during brief trains of synaptic stimuli. Large fluctuations in synaptic efficacy on short time scales seem inconsistent with the conventional connectionist assumption of stable, high-resolution synaptic weight values. Furthermore, a recent experimental study suggests that excitatory synapses in the hippocampus-a region implicated in certain forms of explicit memory-may exist in only a few long-term stable states, where the continuous grading of synaptic strength seen in standard measures of long-term potentiation (LTP) may exist only in the average over a large population of two-state synapses with randomly staggered thresholds for learning (Petersen, Malenka, Nicoli, & Hopfield, 1998). According to conventional connectionist notions, the possibility that individual synapses hold only one or two bits of long-term state information would seem to have serious implications for the storage capacity of neural tissue. Exploration of this question is one of the main themes of this paper. In a related vein, we have found in previous biophysical modeling studies that nonlinear interactions between synapses co-activated on the same branch of an active dendritic tree could provide an alternative form of long-term storage capacity. This capacity, which is largely orthogonal to that tied up in conventional synaptic weights, is contained instead in the spatial permutation of synaptic connections onto the dendritic tree-which could in principle be modified in the course of learning or development (Mel, 1992a, 1992b). In a more abstract setting, we recently showed that a large repository of model flexibility lies in the choice as to which of a large number of possible interaction terms available in high dimension is actually included in a learning machine's discriminant function, and that the excess capacity contained in this "choice flexibility" can be quantified using straightforward counting arguments (Poirazi & Mel, 1999). 2 Two Alternative Models of Dendritic Integration In this paper, we use a similar function-counting approach to address the more biologically relevant case of a neuron with mUltiple quasi-independent dendritic compartments (fig. 1). Our primary objective has been to compare the memory capacity of a cell assuming two different modes of dendritic integration. According to the linear model, the neuron's activation level aL(x) prior to thresholding is given by a weighted sum of of its inputs over the cell as a whole. According to the nonlinear model, the k synaptic inputs to each branch are first combined linearly, a static (e.g. sigmoidal) nonlinearity is applied to each of the m branch subtotals, and the resulting branch outputs are summed to produce the cell's overall activity aN{x): (1) The expressions for aL and aN were written in similar form to emphasize that the models have an identical number of synaptic weights, differing only in the presence or absence of a fixed nonlinear function g applied to the branch subtotals. Though individual synaptic weights in both models are constrained to have a value of 1, any of the d input lines may form multiple connections on the same or different 159 Memory Capacity of Linear vs. Nonlinear Models of Dendritic Integration m 3 ? .. ' , , I . Figure 1: A cell is modeled as a set of m identical branches connected to a soma, where each branch contains k synaptic contacts driven by one of d distinct input lines. branches as a means of representing graded synaptic strengths. Similarly, an input line which forms no connection has an implicit weight of O. In light of this restriction to positive (or zero) weight values, both the linear and nonlinear models are split into two opponent channels a+ and a- dedicated to positive vs. negative coefficients, respectively. This leads to a final output for each model: yL(x) = sgn [at(x) - aL(x)] YN(X) = sgn [a;t(x) - aiV(x)] (2) where the sgn operator maps the total activation level into a class label of {-I, I}. In the following, we derive expressions for the number of distinct parameter st.ates available to the linear vs. nonlinear models, a measure which we have found to be a reliable predictor of storage capacity under certain restrictions (Poirazi & Mel, 1999). Based on these expressions, we compute the capacity boost provided by the branch nonlinearity as a function of the number of branches m, synaptic sites per branch k, and input space dimensionality d. Finally, we test the predictions of the analytical model by training both linear and nonlinear models on randomized classification problems using a stochastic delta rule, and empirically measure and compare the storage capacities of the two models. 3 Results 3.1 Counting Parameter States: Linear vs. Nonlinear Model We derived expressions for BLand B N, which estimate the total number of parameter bits available to the linear vs. nonlinear models, respectively: B N = 2log2 1) (( k+d-1) k m +m - BL = 2log 2 ( S+d-1) S (3) These expressions estimate the number of non-redundant states in each neuron type, i.e., those assignments of input lines to dendritic sites which yield distinct P Poirazi and B. W Mel 160 input-output functions YL or YN? These formulae are plotted in figure 2A with d = 100, where each curve represents a cell with a fixed number of branches (indicated by m). In each case, the capacity increases steadily as the number of synapses per branch, k, is increased. The logarithmic growth in the capacity of the linear model (evident in an asymptotic analysis of the expression for B L) is shown at the bottom of the graph (circles), from which it may be seen that the boost in capacity provided by the dendritic branch nonlinearity increases steadily with the number of synaptic sites. For a cell with 100 branches containing 100 synaptic sites each, the capacity boost relative to the linear model exceeds a factor of 20. Figure 2B shows that for a given total number of synaptic sites, in this case s = m? k = 10,000, the capacity of the nonlinear cell is maximized for a specific choice of m and k. The peak of each of the three curves (computed for different values of d) occurs for a cell containing 1,250 branches with 8 synapses each. However, the capacity is only moderately sensitive to the branch count: the capacity of a cell with 100 branches of 100 synapses each, for example, lies within a factor of two of the optimal configuration. The linear cell capacities can be found at the far right edge of the plot (m = 10,000), since a nonlinear model with one synapse per branch has a number of trainable states identical to that of a linear model. 3.2 Validating the Analytical Model To test the predictions of the analytical model, we trained both linear and nonlinear cells on randomized two-class classification problems. Training samples were drawn from a 40-dimensional spherical Gaussian distribution and were randomly assigned positive or negative labels-in some runs, training patterns were evenly divided between positive and negative labels, with similar results. Each of the 40 original input dimensions was recoded using a set of 10 I-dimensional binary, nonoverlapping receptive fields with centers spaced along each dimension such that all receptive fields would be activated equally often. This manipulation mapped the original 40-dimensional learning problem into 400 dimensions, thereby increasing the discriminability of the training samples. The relative memory capacity of linear vs. nonlinear cells was then determined empirically by comparing the number of training patterns learnable at a fixed error rate of 2%. The learning rule used for both cell types was similar to the "clusteron" learning rule described in (Mel, 1992a), and involved two mechanisms known to contribute to neural development: (1) random activity-independent synapse formation, and (2) activity-dependent synapse stabilization. In each iteration, a set of 25 synapses was chosen at random, and the "worst" synapse was identified based on the correlation over the training set of (i) the input's pre-synaptic activity, (ii) the post-synaptic activity (Le. the local nonlinear branch response for the nonlinear energy model or a constant of 1 for the linear model), and (iii) a global "delta" signal with a value of a if the cell responded correctly to the input pattern, or ?l if the cell responded incorrectly. The poorest-performing synapse on the branch was then targeted for replacement with a new synapse drawn at random from the d input lines. The probability that the replacement actually occurred was given by a Boltzmann equation based on the difference in the training set error rates before and after the replacement. A "temperature" variable was gradually lowered over the course of the simulation, which was terminated when no further improvement in error rates was seen. Results of the learning runs are shown in fig. 3 where the analytical capacity (measured in bits) was scaled to the numerical capacity (measured in training patterns Memory Capacity ofLinear vs. Nonlinear Models ofDendritic Integration B Capacity of Linear VS. Nonlinear Model A Capacity of Linear vs. Nonlinear Model for Various Geometries for Different Input Space Dimensions x 10' 8.---~----~--~----~---, d = 100 ~ ? Nonlinear Model 6 14 ~ m- 000 5 >. '0!IS .t". . . . x10' Hr m=lOOO 7 161 m 12 : d= j lOO~ , '(i.) .... 10 s = 10,000 " j , Nonlinear Model , , co ....... 4 1t3 U '"'. 2 Linear ~del l~ Linear Model 2000 4000 6000 8000 10000 ~~l Syn:p:c Sires ~ o 2000 4000 6000 8000 10000 Number of Branches (m) * Figure 2: Comparison of linear vs. nonlinear model capacity as a function of branch geometry. A. Capacity in bits for linear and several nonlinear cells with different branch counts (for d = 100). For each curve indexed by branch count m, sites per branch k increases from left to right as indicated iconically beneath the x-axis. For all cells, capacity increases with an increasing number of sites, though the capacity of the linear model grows logarithmically, leading to an increasingly large capacity boost for the size-matched nonlinear cells. B. Capacity of a nonlinear model with 10,000 sites for different values of input space dimension d. Branch count m grows along the x-axis. Cells at right edge of plot contain only one synapse per branch, and thus have a number of modifiable parameters (and hence capacity) equivalent to that of the linear model. All three curves show that there exist an optimal geometry which maximizes the capacity of the nonlinear model (in this case 1,250 branches with 8 synapses each). learned at 2% error). Two key features of the theoretical curves (dashed lines) are echoed in the empirical performance curves (solid lines), including the much larger storage capacity of the nonlinear cell model, and the specific cell geometry which maximizes the capacity boost. 4 Discussion We found using both analytical and numerical methods that in the limit of lowresolution synaptic weights, application of a fixed output nonlinearity to each compartment of a dendritic tree leads to a significant boost in capacity relative to a cell whose post-synaptic integration is linear. For example, given a cell with 10,000 synaptic contacts originating from 400 distinct input lines, the analysis predicts a 23-fold increase in capacity for the nonlinear cell, while numerical simulations using a stochastic delta rule actually achieve a I5-fold boost. Given that a linear and a nonlinear model have an identical number of synaptic contacts with uniform synaptic weight values, what accounts for the capacity boost? The principal insight gained in this work is that the attachment of a fixed nonlinearity to each branch in a neuron substantially increases its underlying "model 162 P. Poirazi and B. W. Mel Figure 3: Comparison of capacity boost predicted by analysis vs. that observed empirically when linear and nonlinear models were trained using the same - _. Analytical stochastic delta rule. Dashed (Bits/14) 70 lines: analytical curves for linNumerical ear vs. nonlinear model for a cell I \ (Training Patterns) 6 ,+ \ with 10,000 sites show capacity I, \ for varying cell geometries. Solid >. 50 \. lines: empirical performance for '13 \ Nonlinear Model same two cells at 2% error cri~ 40 ____________________\ 03 terion, using a subunit nonlinearity g(x) = x lO (similar re<..) 30 '" , sults were seen using a sigmoidal ,, ,, nonlinearity, though the param, ,. 2 eters of the optimal sigmoid depended on the cell geometry). Linear Model For both analytical and numeri, 2 cal curves, peak capacity is seen oo 10 20 30 40 50 60 70 80 90 100x10 for cell with 1,000 branches (10 Number of Branches (m) synapses per branch) .. Cap~city exceeds that of same-sIzed lmear .:Jk- model by a factor of 15 at the m ~ peak, and by more than a factor of 7 for cells ranging from about 3 to 60 synapses per branch (horizontal dotted line). 1 ~, * ---I...... flexibility" , i.e. confers upon the cell a much larger choice of distinct input-output relations from which to select during learning. This may be illustrated as follows. For the linear model, branching structure is irrelevant so that YL depends only on the number of input connections formed from each of the d input lines. All spatial permutations of a set of input connections are thus interchangeable and produce identical cell responses. This massive redundancy confines the capacity of the linear model to grow only logarithmically with an increasing number of synaptic sites (fig. 1A), an unfortunate limitation for a brain in which the formation of large numbers of synaptic contacts between neurons is routine. In contrast, the model with nonlinear subunits contains many fewer redundancies: most spatial permutations of the same set of input connections lead to non-identical values of YN, since an input x swapped from branch bi to branch b2 leads to the elimination of the k - 1 interaction terms involving x on branch bi and the creation of k -1 new interaction terms on branch b2 ? Interestingly, the particular form of the branch nonlinearity has virtually no effect on the capacity of the cell as far as the counting arguments are concerned (though it can have a profound effect on the cell's "representational bias"-see below), since the principal effect of the nonlinearity in our capacity calculations is to break the symmetry among the different branches. The issue of representational bias is a critical one, however, and must be considered when attempting to predict absolute or relative performance rates for particular classifiers confronted with specific learning problems. Thus, intrinsic differences in the geometry of linear vs. nonlinear discriminant functions mean that the param- Memory Capacity ofLinear vs. Nonlinear Models ofDendritic Integration 163 eters available to the two models may be better or worse suited to solve a given learning problem, even if the two models were equated for total parameter flexibility. While such biases are not taken into account in our analysis, they could nonetheless have a substantial effect on measured error rates-and could thus throw a performance advantage to one machine or the other. One danger is that performance differences measured empirically could be misinterpreted as arising from differences in underlying model capacity, when in fact they arise from differential suitability of the two classifiers for the learning problem at hand. To avoid this difficulty, the random classification problems we used to empirically assess memory capacity were chosen to level the playing field for the linear vs. nonlinear cells, since in a previous study we found that the coefficients on linear vs. nonlinear (quadratic) terms were about equally efficient as featUres for this task. In this way, differences in measured performance on these tasks were primarily attributable to underlying capacity differences, rather than differences in representational bias. This experimental control permitted more meaningful comparisons between our analytical and empirical tests (fig. 3). The problem of representational bias crops up in a second guise, wherein the analytical expressions for capacity in eq. 1 can significantly overestimate the actual performance of the cell. This occurs when a particular ensemble of learning problems fails to utilize all of the entropy available in the cell's parameter space-for example, by requiring the cell to visit only a small subset of its parameter states relatively often. This invalidates the maximum parameter entropy assumption made in the derivation of eq. 1, so that measured performance will tend to fall below predicted values. The actual performance of either model when confronted with an ensemble of learning problems will thus be determined by (1) the number of trainable parameters available to the neuron (as measured by eq. 1), (2) the suitability of the neuron's parameters for solving the assigned learning problems, and (3) the utilization of parameters, which relates to the entropy in the joint probability of the parameter values averaged over the ensemble of learning problems. In our comparisons here of linear and nonlinear cells, we we have calculated (1), and have attempted to control for (2) and (3). In conclusion, our results build upon the results of earlier biophysical simulations, and indicate that in the limit of a large number of low-resolution synaptic weights, nonlinear dendritic processing could nonetheless have a major impact on the storage capacity of neural tissue. References Mel, B. W. (1992a). The clusteron: Toward a simple abstraction for a complex neuron. In Moody, J., Hanson, S., & Lippmann, R. (Eds.), Advances in Neural Information Processing Systems, vol. 4, pp. 35-42: Morgan Kaufmann, San Mateo, CA. Mel, B. W. (1992b). NMDA-based pattern discrimination in a modeled cortical neuron. Neural Comp., 4, 502-516. Petersen, C. C. H., Malenka, R. C., Nicoll, R. A., & Hopfield, J. J. (1998). All-ornone potentiation and CA3-CA1 synapses. Proc. Natl. Acad. Sci. USA, 95, 4732-4737. Poirazi, P., & Mel, B. W. (1999). Choice and value flexibility jointly contribute to the capacity of a subsampled quadratic classifier. Neural Comp., in press.
1646 |@word repository:1 hippocampus:1 simulation:3 thereby:1 solid:2 configuration:1 contains:2 efficacy:2 terion:1 interestingly:1 comparing:1 activation:2 written:1 readily:1 must:1 numerical:3 realistic:1 plot:2 v:17 discrimination:1 fewer:1 short:1 contribute:2 sigmoidal:2 misinterpreted:1 along:2 differential:1 lowresolution:1 profound:1 brain:2 spherical:1 actual:2 param:2 increasing:3 provided:2 estimating:2 matched:1 underlying:3 maximizes:2 what:1 substantially:2 ca1:1 differing:1 growth:1 scaled:1 classifier:3 control:2 utilization:1 yn:3 overestimate:1 positive:4 before:1 engineering:2 local:1 limit:2 depended:1 acad:1 fluctuation:2 discriminability:1 mateo:1 quantified:1 suggests:1 co:2 bi:2 averaged:1 practice:1 danger:1 empirical:3 significantly:1 pre:1 petersen:2 onto:1 operator:1 cal:1 storage:6 restriction:2 equivalent:1 conventional:3 map:1 center:1 confers:1 straightforward:1 resolution:2 rule:6 insight:1 population:1 notion:2 limiting:1 massive:1 logarithmically:2 jk:1 located:1 predicts:1 vein:1 bottom:1 observed:1 worst:1 region:1 connected:1 substantial:2 moderately:1 trained:2 interchangeable:1 subtotal:3 solving:1 creation:1 upon:2 joint:1 hopfield:2 various:1 derivation:1 train:2 distinct:6 sc:1 formation:2 whose:1 larger:2 valued:1 solve:1 jointly:1 final:1 confronted:2 advantage:1 biophysical:3 analytical:10 interaction:5 relevant:1 beneath:1 flexibility:6 achieve:1 representational:4 los:2 transmission:1 produce:2 derive:2 oo:1 measured:7 eq:3 throw:1 predicted:3 indicate:1 quantify:1 stochastic:4 exploration:1 stabilization:1 sgn:3 elimination:1 potentiation:2 suitability:2 dendritic:16 hold:1 considered:1 echoed:1 predict:1 major:1 proc:1 label:3 sensitive:1 largest:1 city:1 weighted:1 gaussian:1 aim:1 modified:1 rather:1 avoid:1 varying:1 validated:1 focus:1 derived:1 improvement:1 staggered:1 contrast:1 dependent:1 cri:1 abstraction:1 entire:1 relation:1 originating:1 quasi:1 overall:1 among:2 classification:4 issue:1 development:2 spatial:3 integration:8 summed:1 constrained:1 field:3 identical:6 represents:1 connectionist:3 stimulus:1 serious:1 few:1 primarily:1 randomly:2 individual:2 usc:3 subsampled:1 geometry:7 replacement:3 possibility:1 light:1 activated:2 natl:1 implication:1 edge:2 orthogonal:1 tree:4 indexed:1 circle:1 plotted:1 re:1 theoretical:1 increased:1 modeling:2 earlier:1 measuring:1 assignment:1 ca3:1 subset:1 predictor:1 uniform:1 loo:1 stored:1 combined:3 st:1 peak:3 randomized:3 yl:3 moody:1 ear:1 containing:2 worse:1 leading:1 account:2 nonoverlapping:1 b2:2 coefficient:2 depends:1 break:1 ass:1 formed:1 compartment:3 responded:2 kaufmann:1 largely:1 maximized:1 yield:1 spaced:1 t3:1 ensemble:3 eters:2 comp:2 tissue:2 synapsis:13 synaptic:28 ed:1 energy:1 nonetheless:2 pp:1 steadily:2 involved:1 static:1 cap:1 dimensionality:1 nmda:1 syn:1 routine:1 actually:3 permitted:1 response:2 wherein:1 synapse:7 though:4 furthermore:1 biomedical:2 implicit:1 correlation:1 hand:1 horizontal:1 nonlinear:43 aiv:1 del:1 mode:1 indicated:2 grows:2 usa:1 effect:4 contain:1 requiring:1 hence:1 assigned:2 illustrated:1 during:2 branching:1 mel:13 evident:1 dedicated:1 temperature:1 passive:1 ranging:1 recently:1 sigmoid:1 empirically:6 occurred:1 significant:1 similarly:1 nonlinearity:10 lowered:1 stable:2 showed:2 recent:1 irrelevant:1 driven:1 manipulation:1 certain:2 binary:2 seen:5 morgan:1 redundant:1 dashed:2 ii:1 relates:1 branch:42 multiple:2 signal:1 x10:2 exceeds:2 calculation:1 long:4 divided:1 bland:1 post:2 equally:2 visit:1 impact:1 prediction:2 involving:1 crop:1 iteration:1 achieved:1 cell:44 separately:1 grow:1 swapped:1 tend:1 undergo:1 ltp:1 validating:1 virtually:1 inconsistent:1 seem:2 invalidates:1 counting:4 presence:1 exceed:1 split:1 iii:1 concerned:1 variety:1 identified:1 poirazi:8 grading:1 angeles:2 expression:8 bartlett:1 lnc:2 http:1 exist:3 dotted:1 delta:5 arising:1 per:7 correctly:1 modifiable:1 vol:1 key:1 redundancy:2 soma:1 threshold:3 drawn:2 utilize:1 graph:1 sum:1 run:2 i5:1 bit:5 poorest:1 followed:1 fold:2 quadratic:2 activity:5 strength:2 nearby:1 argument:2 malenka:2 performing:1 attempting:1 relatively:3 department:2 according:3 across:1 ate:1 increasingly:1 biologically:1 gradually:1 taken:1 equation:1 nicoll:1 count:4 mechanism:1 available:7 opponent:1 alternative:2 original:2 unfortunate:1 log2:1 build:1 graded:1 contact:4 bl:1 objective:1 question:1 occurs:3 receptive:2 primary:1 southern:2 mapped:1 sci:1 capacity:57 evenly:1 discriminant:2 toward:1 assuming:1 modeled:2 negative:3 recoded:1 boltzmann:1 neuron:12 incorrectly:1 subunit:3 connection:6 hanson:1 california:2 learned:2 ornone:1 boost:13 address:1 below:2 pattern:7 reliable:1 memory:11 including:1 critical:1 difficulty:1 hr:1 representing:1 brief:1 sults:1 axis:2 attachment:1 prior:1 asymptotic:1 relative:4 permutation:3 limitation:1 principle:1 thresholding:1 playing:1 lo:1 excitatory:1 course:2 looo:1 implicated:1 bias:5 fall:1 absolute:1 lmear:1 curve:8 dimension:6 calculated:1 cortical:1 equated:1 made:1 san:1 far:2 excess:1 emphasize:1 lippmann:1 global:2 active:3 continuous:1 channel:1 ca:3 symmetry:1 dendrite:1 complex:1 main:1 linearly:3 terminated:1 whole:1 arise:1 neuronal:1 fig:4 site:10 oflinear:2 attributable:1 guise:1 theme:1 fails:1 explicit:1 lie:2 tied:1 down:1 formula:1 specific:3 learnable:1 physiological:1 evidence:1 intrinsic:1 clusteron:2 gained:1 magnitude:1 suited:1 entropy:3 logarithmic:1 contained:2 sized:2 targeted:1 absence:1 included:1 determined:2 principal:2 total:4 experimental:2 attempted:1 meaningful:1 select:1 support:1 confines:1 preparation:1 trainable:2
707
1,647
An Information-Theoretic Framework for Understanding Saccadic Eye Movements Tai Sing Lee * Department of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 Stella X. Yu Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 tai@es.emu.edu stella@enbe.emu.edu Abstract In this paper, we propose that information maximization can provide a unified framework for understanding saccadic eye movements. In this framework, the mutual information among the cortical representations of the retinal image, the priors constructed from our long term visual experience, and a dynamic short-term internal representation constructed from recent saccades provides a map for guiding eye navigation . By directing the eyes to locations of maximum complexity in neuronal ensemble responses at each step, the automatic saccadic eye movement system greedily collects information about the external world, while modifying the neural representations in the process. This framework attempts to connect several psychological phenomena, such as pop-out and inhibition of return, to long term visual experience and short term working memory. It also provides an interesting perspective on contextual computation and formation of neural representation in the visual system. 1 Introduction When we look at a painting or a visual scene, our eyes move around rapidly and constantly to look at different parts of the scene. Are there rules and principles that govern where the eyes are going to look next at each moment? In this paper, we sketch a theoretical framework based on information maximization to reason about the organization of saccadic eye movements. ?Both authors are members of the Center for the Neural Basis of Cognition - a joint center between University of Pittsburgh and Carnegie Mellon University. Address: Rm 115, Mellon Institute, Carnegie Mellon University, Pittsburgh, PA 15213. 835 lnformation-Theoretic Framework for Understanding Saccadic Behaviors Vision is fundamentally a Bayesian inference process. Given the measurement by the retinas, the brain's memory of eye positions and its prior knowledge of the world, our brain has to make an inference about what is where in the visual scene. The retina, unlike a camera, has a peculiar design. It has a small foveal region dedicated to high-resolution analysis and a large low-resolution peripheral region for monitoring the rest of the visual field. At about 2.5 0 visual angle away from the center of the fovea, visual acuity is already reduced by a half. When we 'look' (foveate) at a certain location in the visual scene, we direct our high-resolution fovea to analyze information in that location, taking a snap shot of the scene using our retina. Figure lA-C illustrate what a retina would see at each fixation. It is immediately obvious that our retinal image is severely limited - it is clear only in the fovea and is very blurry in the surround, posing a severe constraint on the information available to our inference system. Yet, in our subjective experience, the world seems to be stable, coherent and complete in front of us. This is a paradox that have engaged philosophical and scientific debates for ages. To overcome the constraint of the retinal image, during perception, the brain actively moves the eyes around to (1) gather information to construct a mental image of the world, and (2) to make inference about the world based on this mental image. Understanding the forces that drive saccadic eye movements is important to elucidating the principles of active perception. A B C D Figure 1. A-C: retinal images in three separate fixations. D: a mental mosaic created by integrating the retinal images from these three and other three fixations. It is intuitive to think that eye movements are used to gather information. Eye movements have been suggested to provide a means for measuring the allocation of attention or the values of each kind of information in a particular context [16]. The basic assumption of our theory is that we move our eyes around to maximize our information intake from the world, for constructing the mental image and for making inference of the scene. Therefore, the system should always look for and attentively fixate at a location in the retinal image that is the most unusual or the most unexplained - and hence carries the maximum amount of information. 2 Perceptual Representation How can the brain decide which part of the retinal image is more unusual? First of all, we know the responses of VI simple cells, modeled well by the Gabor wavelet pyramid [3,7], can be used to reconstruct completely the retinal image. It is also well established that the receptive fields of these neurons developed in such a way as to provide a compact code for natural images [8,9,13,14]. The idea of compact code or sparse code, originally proposed by Barlow [2], is that early visual neurons capture the statistical correlations in natural scenes so that only a small number T. S. Lee and S. X Yu 836 of cells out of a large set will be activated to represent a particular scene at each moment. Extending this logic, we suggest that the complexity or the entropy of the neuronal ensemble response of a hypercolumn in VI is therefore closely related to the strangeness of the image features being analyzed by the machinery in that hypercolumn. A frequent event will have a more compact representation in the neuronal ensemble response. Entropy is an information measure that captures the complexity or the variability of signals. The entropy of a neuronal ensemble in a hypercolumn can therefore be used to quantify the strangeness of a particular event. A hypercolumn in the visual cortex contains roughly 200,000 neurons, dedicated to analyzing different aspects of the image in its 'visual window' . These cells are tuned to different spatial positions, orientations, spatial frequency, color disparity and other cues. There might also be a certain degree of redundancy, i.e. a number of neurons are tuned to the same feature . Thus a hypercolumn forms the fundamental computational unit for image analysis within a particular window in visual space. Each hypercolumn contains cells with receptive fields of different sizes, many significantly smaller than the aggregated 'visual window' of the hypercolumn. The entropy of a hypercolumn's ensemble response at a certain time t is the sum of entropies of all the channels, given by, H(u(R:;, t)) = - 2: 2:p(u(R:;, v, 0', B, t)) log2P(u(R:;, v, 0', B, t)) 9,<7 V where u(R:;, t) denotes the responses of all complex cell channels inside the visual window R:; of a hypercolumn at time t, computed within a 20 msec time window. u(i, 0', B, t) is the response of a VI complex cell channel of a particular scale 0' and orientation 0' at spatial location i at t. p(u(R:;, v, 0', B, t)) is the probability of cells in that channel within the visual window R:; of the hypercolumn firing v number of spikes . v can be computed as the power modulus of the corresponding simple cell channels, modeled by Gabor wavelets [see 7] . L:v p(u(R:;, v, 0', B, t)) 1. The probability p(u(R:;, v, 0', B, t)) can be computed at each moment in time because of the variations in spatial position of the receptive fields of similar cell within the hypercolumn - hence the 'same' cells in the hypercolumn are analyzing different image patches, and also because of the redundancy of cells coding similar features. = The neurons' responses in a hypercolumn are subject to contextual modulation from other hypercolumns, partly in the form of lateral inhibition from cells with similar tunings. The net observed effect is that the later part of VI neurons' response, starting at about 80 msec, exhibits differential suppression depending on the spatial extent and the nature of the surround stimulus. The more similar the surround stimulus is to the center stimuli, and the larger the spatial extent of the 'similar surround', the stronger is the suppressive effect [e.g. 6]. Simoncelli and Schwartz [15] have proposed that the steady state responses of the cells can be modeled by dividing the response of the cell (i.e. modeled by the wavelet coefficient or its power modulus) by a weighted combination ofthe responses of its spatial neighbors in order to remove the statistical dependencies between the responses of spatial neighbors. These weights are found by minimizing a predictive error between the center signal from the surround signals. In our context, this idea of predictive coding [see also 14] is captured by the concept of mutual information between the ensemble responses of the different hypercolumns as given below, I(u(R x , t); u(Ox, t - dtd) H(u(R:;, t)) - H(u(Rx , t)lu(Ox, t - dtd) 2: 2: [P(U(R:;,VR,O',B,t),u(O:;,vn,O',B,t)) <7,9 I VR,VO p(u(Rx, VR, 0', B, t), u(Ox, vn, 0', B, t)) ] og2 p(u(R:;, vR,O',B,t)),p(u(O:;,vn, O',B,t)) . Information-Theoretic Framework for Understanding Saccadic Behaviors 837 where u(Rx, t) is the ensemble response of the hypercolumn in question, and u(Ox, t) is the ensemble response of the surrounding hypercolumns. p(u(Rx , VR, (1', (), t)) is the probability that cells of a channel in the center hypercolumn assumes the response value VR and p(u(Ox, VR, (1', (), t)) the probability that cells of a similar channel in the surrounding hypercolumns assuming the response value Vn. tl is the delay by which the surround information exerts its effect on the center hypercolumn. The mutual information I can be computed from the joint probability of ensemble responses of the center and the surround. The steady state responses of the VI neurons, as a result of this contextual modulation, are said to be more correlated to perceptual pop-out than the neurons' initial responses [5,6]. The complexity of the steady state response in the early visual cortex is described by the following conditional entropy, H(u(R x , t)lu(Ox, t - dtd) = H(u(R x , t)) - I(u(R x , t); u(Ox, t - dtd). However, the computation in VI is not limited to the creation of compact representation through surround inhibition. In fact, we have suggested that VI plays an active role in scene interpretation particularly when such inference involves high resolution details [6]. Visual tasks such as the inference of contour and surface likely involve VI heavily. These computations could further modify the steady state responses of VI, and hence the control of saccadic eye movements. 3 Mental Mosaic Representation The perceptual representation provides the basic force for the brain to steer the high resolution fovea to locations of maximum uncertainty or maximum signal complexity. Foveation captures the maximum amount of available information in a location. Once a location is examined by the fovea, its information uncertainty is greatly reduced. The eyes should move on and not to return to the same spot within a certain period of time. This is called the 'inhibition of return'. How can we model this reduction of interest? We propose that the mind creates a mental mosaic of the scene in order to keep track of the information that have been gathered. By mosaic, we mean that the brain can assemble successive retinal images obtained from multiple fixations into a coherent mental picture of the scene. Figure ID provides an example of a mental mosaic created by combining information from the retinal images from 6 fixations. Whether the brain actually keeps such a mental mosaic of the scene is currently under debate. McConkie and Rayner [10] had suggested the idea of an integrative visual buffer to integrate information across multiple saccades. However, numerous experiments demonstrated we actually remember relatively little across saccades [4]. This lead to the idea that brain may not need an explicit internal representation of the world. Since the world is always out there, the brain can access whatever information it needs at the appropriate details by moving the eyes to the appropriate place at the appropriate time. The subjective feeling of a coherent and a complete world in front of us is a mere illusion [e.g. 1]. The mental mosaic represented in Figure ID might resemble McConkie and Rayner's theory superficially. But the existence of such a detailed high-resolution buffer with a large spatial support in the brain is rather biologically implausible. Rather, we think that the information corresponding to the mental mosaic is stored in an interpreted and semantic form in a mesh of Bayesian belief networks in the brain (e.g. involving PO, IT and area 46). This distributed semantic representation of T. S. Lee and S. X Yu 838 the mental mosaic, however, is capable of generating detailed (sometimes false) imagery in early visual cortex using the massive recurrent convergent feedback from the higher areas to VI. However, because of the limited support provided by VI machinery, the instantiation of mental imagery in VI has to be done sequentially one 'retinal image' frame at a time, presumably in conjunction with eye movement, even when the eyes are closed. This might explain why vivid visual dream is always accompanied by rapid eye movement in REM sleep. The mental mosaic accumulates information from the retinal images up to the last fixation and can provide prediction on what the retina will see in the current fixation. For each u(i, (T, 0) cell, there is a corresponding effective prediction signal m(i, (T, 0) fed back from the mental mosaic. This prediction signal can reduce the conditional entropy or complexity of the ensemble response in the perceptual representation by discounting the mutual information between the ensemble response to the retinal image and the mental mosaic prediction as follow, H(u(R x , t)lm(Rx, t - 6t2)) = H(u(R x , t)) - I(u(R x , t), m(Rx , t - dt2)) where 6t2 is the transmission delay from the mental mosaic back to VI. At places where the fovea has visited, the mental mosaic representation has high resolution information and m(i, (T, 0, t - 6t2) can explain u(i, (T, 0, t) fully. Hence, the mutual information is high at those hypercolumns and the conditional entropy H(u(R x , t)lm(R x , t - 6t2)) is low, with two consequences: (1) the system will not get the eyes stuck at a particular location; once the information at i is updated to the mental mosaic, the system will lose interest and move on; (2) the system will exhibit 'inhibition of return' as the information in the visited locations are fully predicted by the mental mosaic. Also, from this standpoint, the 'habituation dynamics' often observed in visual neurons when the same stimulus is presented multiple times might not be simply due to neuro-chemical fatigue, but might be understood in terms of mental mosaic being updated and then fed back to explain the perceptual representation in VI. The mental mosaic is in effect our short-term memory of the scene. It has a forgetting dynamics, and needs to be periodically updated. Otherwise, it will rapidly fade away. 4 Overall Reactive Saccadic Behaviors Now, we can combine the influence of the two predictive processes to arrive at a discounted complexity measure of the hypercolumn's ensemble response: H(u(R x , t)) -I(u(Rx , t); u(Ox, t - 6tt)) -I(u(Rx , t); m(R x , t - 6t2)) +I(u(Ox , t - 6td; m(Rx , t - 6t2)) If we can assume the long range surround priors and the mental mosaic short term memory are independent processes, we can leave out the last term, I(u(Ox, t 6td; m(Rx , t - 6t2)), of the equation. The system, after each saccade, will evaluate the new retinal scene and select the location where the perceptual representation has the maximum conditional entropy. To maximize the information gain, the system must constantly search for and make a saccade to the locations of maximum uncertainty (or complexity) computed from Information-Theoretic Framework for Understanding Saccadic Behaviors 839 the hypercolumn ensemble responses in VI at each fixation. Unless the number of saccades is severely limited, this locally greedy algorithm, coupled the inhibition of return mechanism, will likely steer the system to a relatively optimal global sampling of the world - in the sense that the average information gain per saccade is maximized, and the mental mosaic's dissonance with the world is minimized . 5 Task-dependent schema Representation However, human eye movements are not simply controlled by the generic information in a bottom-up fashion . Yarbus [16] has shown that, when staring at a face, subjects' eyes tend to go back to the same locations (eyes, mouth) over and over again. Further, he showed that when asked different questions, subjects exhibited different kinds of scan-paths when looking at the same picture. Norton and Stark [12] also showed that eye movements are not random, but often exhibit repetitive or even idiosyncratic path patterns. To capture these ideas, we propose a third representation, called task schema, to provide the necessary top-down information to bias the eye movement control. It specifies the learned or habitual scan-paths for a particular task in a particular context or assigns weights to different types of information. Given that we arenot mostly unconscious of the scan-path patterns we are making, these task-sensitive or context-sensitive habitual scan-patterns might be encoded at the levels of motor programs, and be downloaded when needed without our conscious control. These motor programs for scan-paths can be trained from reinforcement learning. For example, since the eyes and the mouths convey most of the emotional content of a facial expression, a successful interpretation of another person's emotion could provide the reward signal to reinforce the motor programs just executed or the fixations to certain facial features. These unconscious scan-path motor programs could provide the additional modulation to automatic saccadic eye movement generation. 6 Discussion In this paper, we propose that information maximization might provide a theoretical framework to understand the automatic saccadic eye movement behaviors in human. In this proposal, each hypercolumn in V 1 is considered a fundamental computational unit. The relative complexity or entropy of the neuronal ensemble response in the VI hypercolumns, discounted by the predictive effect of the surround, higher order representations and working memory, creates a force field to guide eye navigation. The framework we sketched here bridge natural scene statistics to eye movement control via the more established ideas of sparse coding and predictive coding in neural representation. Information maximization has been suggested to be a possible explanation for shaping the receptive fields in the early visual cortex according to the statistics of natural images [8,9,13,14] to create a minimum-entropy code [2,3]. As a result, a frequent event is represented efficiently with the response of a few neurons in a large set, resulting in a lower hypercolumn ensemble entropy, while unusual events provoke ensemble responses of higher complexity. We suggest that higher complexity in ensemble responses will arouse attention and draw scrutiny by the eyes, forcing the neural representation to continue adapting to the statistics of the natural scenes. The formulation here also suggests that information maximization might provide an explanation for the formation of horizontal predictive network in VI as well as higher order internal representations, consistent with the ideas of predictive coding [11, 14, 15]. Our theory hence predicts that the adaptation of the 840 T. S. Lee and S. X Yu neural representations to the statistics of natural scenes will lead to the adaptation of? saccadic eye movement behaviors. Acknowledgements The authors have been supported by a grant from the McDonnell Foundation and a NSF grant (LIS 9720350). Yu is also being supported in part by a grant to Takeo Kanade. References [1] Ballard, D. Hayhoe, M.M. Pook, P.K. & Rao, RP.N. (1997). Deictic codes for the embodiment of cognition. Behavioral and Brain Science, 20:4, December, 723-767. [2] Barlow, H.B. (1989). Unsupervised learning. Neural Computation, 1, 295-311. [3] Daugman, J.G. (1989). Entropy reduction and decorrelation in visual coding by oriented neural receptive fields. IEEE Transactions on Biomedical Engineering 36:, 107-114. [4] Irwin, D. E, 1991. Information Integration across Saccadic Eye Movements. Cognitive Psychology, 23(3):420-56. [5] Knierim, J. & Van Essen, D.C. Neural response to static texture patterns in area VI of macaque monkey. J. Neurophysiology, 67: 961-980. [6] Lee, T.S., Mumford, D., Romero R & Lamme, V.A.F. (1998). The role of primary visual cortex in higher level vision. Vision Research 38, 2429-2454. [7] Lee, T.S. (1996). Image representation using 2D Gabor wavelets. IEEE Transaction of Pattern Analysis and Machine Intelligence. 18:10, 959-971. [8] Lewicki, M. & Olshausen, B. (1998). Inferring sparse, overcomplete image codes using an efficient coding framework. In Advances in Neural Information Processing System 10, M. Jordan, M. Kearns and S. Solla (eds). MIT Press. [9] Linsker, R (1989). How to generate ordered maps by maximizing the mutual information between input and output signals. Neural Computation, 1: 402-411.0 [10] McConkie, G.W. & Rayner, K. (1976) . Identifying the span of effective stimulus in reading. Literature review and theories of reading. In H. Singer and RB. Ruddell (Eds), Theoretical models and processes of reading, 137-162. Newark, D.E.: International Reading Association. [11] Mumford, D. (1992). On the computational architecture of the neocortex II. Biological cybernetics, 66, 241-251. [12] Norton, D. and Stark, 1. (1971) Eye movements and visual perception. American, 224, 34-43. Scientific [13] Olshausen, B.A., & Field, D.J. (1996), Emergence of simple cell receptive field properties by learning a sparse code for natural images. Nature, 381: 607-609. [14] Rao R., & Ballard, D.H. (1999). Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive field effects. Nature Neuroscience, 2:1 7987. [15] Simoncelli, E.P. & Schwartz, O. (1999). Modeling surround suppression in VI neurons with a statistically-derived normalization model. In Advances in Neural Information Processing Systems 11, . M.S. Kearns, S.A. Solla, and D.A. Cohn (eds). MIT Press. [16] Yarbus, A.L. (1967). Eye movements and vision. Plenum Press.
1647 |@word mcconkie:3 neurophysiology:1 seems:1 stronger:1 integrative:1 rayner:3 shot:1 carry:1 reduction:2 moment:3 initial:1 foveal:1 contains:2 disparity:1 tuned:2 subjective:2 current:1 contextual:3 intake:1 yet:1 must:1 takeo:1 mesh:1 periodically:1 romero:1 motor:4 remove:1 half:1 cue:1 greedy:1 intelligence:1 short:4 mental:24 provides:4 location:13 successive:1 yarbus:2 constructed:2 direct:1 differential:1 fixation:9 combine:1 behavioral:1 inside:1 forgetting:1 rapid:1 roughly:1 behavior:6 brain:12 rem:1 discounted:2 td:2 little:1 window:6 provided:1 what:3 kind:2 interpreted:1 monkey:1 developed:1 unified:1 remember:1 rm:1 schwartz:2 control:4 unit:2 whatever:1 grant:3 scrutiny:1 understood:1 engineering:1 modify:1 severely:2 consequence:1 accumulates:1 analyzing:2 id:2 firing:1 modulation:3 path:6 might:8 examined:1 collect:1 suggests:1 limited:4 range:1 statistically:1 camera:1 illusion:1 spot:1 area:3 gabor:3 significantly:1 adapting:1 integrating:1 suggest:2 get:1 context:4 influence:1 map:2 demonstrated:1 center:8 maximizing:1 go:1 attention:2 starting:1 resolution:7 identifying:1 immediately:1 assigns:1 fade:1 rule:1 variation:1 updated:3 plenum:1 unconscious:2 play:1 heavily:1 massive:1 mosaic:20 pa:3 particularly:1 predicts:1 observed:2 role:2 bottom:1 capture:4 region:2 solla:2 movement:20 govern:1 complexity:11 reward:1 asked:1 dt2:1 dynamic:3 trained:1 predictive:8 creation:1 creates:2 basis:1 completely:1 po:1 joint:2 represented:2 surrounding:2 effective:2 formation:2 encoded:1 larger:1 snap:1 reconstruct:1 otherwise:1 statistic:4 think:2 emergence:1 net:1 propose:4 adaptation:2 frequent:2 combining:1 rapidly:2 intuitive:1 transmission:1 extending:1 generating:1 leave:1 illustrate:1 depending:1 recurrent:1 dividing:1 predicted:1 involves:1 resemble:1 quantify:1 closely:1 modifying:1 human:2 biological:1 around:3 considered:1 presumably:1 cognition:2 lm:2 early:4 lose:1 currently:1 unexplained:1 visited:2 sensitive:2 bridge:1 create:1 weighted:1 mit:2 always:3 rather:2 conjunction:1 derived:1 acuity:1 greatly:1 greedily:1 suppression:2 sense:1 inference:7 dependent:1 going:1 sketched:1 overall:1 among:1 orientation:2 spatial:9 integration:1 mutual:6 field:10 construct:1 once:2 emotion:1 dissonance:1 sampling:1 yu:5 look:5 unsupervised:1 linsker:1 minimized:1 t2:7 stimulus:5 fundamentally:1 few:1 retina:5 oriented:1 attempt:1 organization:1 interest:2 essen:1 elucidating:1 severe:1 navigation:2 analyzed:1 activated:1 strangeness:2 peculiar:1 capable:1 necessary:1 experience:3 facial:2 machinery:2 unless:1 overcomplete:1 theoretical:3 psychological:1 modeling:1 steer:2 rao:2 measuring:1 maximization:5 delay:2 successful:1 front:2 stored:1 connect:1 dependency:1 hypercolumns:6 person:1 fundamental:2 international:1 lee:6 imagery:2 again:1 external:1 cognitive:1 american:1 return:5 stark:2 actively:1 li:1 retinal:14 accompanied:1 coding:8 coefficient:1 provoke:1 vi:19 later:1 closed:1 analyze:1 schema:2 staring:1 efficiently:1 ensemble:17 gathered:1 ofthe:1 painting:1 maximized:1 bayesian:2 lu:2 mere:1 monitoring:1 rx:10 drive:1 cybernetics:1 explain:3 implausible:1 ed:3 norton:2 frequency:1 obvious:1 fixate:1 static:1 gain:2 knowledge:1 color:1 shaping:1 actually:2 back:4 originally:1 higher:6 follow:1 response:32 formulation:1 done:1 ox:10 just:1 biomedical:1 correlation:1 working:2 sketch:1 horizontal:1 cohn:1 scientific:2 olshausen:2 modulus:2 effect:6 concept:1 barlow:2 hence:5 discounting:1 chemical:1 semantic:2 during:1 steady:4 fatigue:1 theoretic:4 complete:2 vo:1 tt:1 dedicated:2 dtd:4 image:25 log2p:1 functional:1 association:1 interpretation:3 he:1 mellon:5 measurement:1 surround:11 automatic:3 tuning:1 had:1 moving:1 stable:1 access:1 cortex:6 surface:1 inhibition:6 habitual:2 recent:1 showed:2 perspective:1 forcing:1 certain:5 buffer:2 continue:1 captured:1 minimum:1 additional:1 aggregated:1 maximize:2 period:1 signal:8 ii:1 multiple:3 simoncelli:2 long:3 controlled:1 prediction:4 involving:1 basic:2 neuro:1 vision:4 exerts:1 repetitive:1 represent:1 sometimes:1 normalization:1 pyramid:1 robotics:1 cell:18 proposal:1 pook:1 suppressive:1 standpoint:1 extra:1 rest:1 unlike:1 exhibited:1 subject:3 tend:1 member:1 december:1 habituation:1 jordan:1 psychology:1 architecture:1 reduce:1 idea:7 whether:1 expression:1 clear:1 involve:1 detailed:2 amount:2 daugman:1 neocortex:1 locally:1 conscious:1 reduced:2 generate:1 specifies:1 nsf:1 neuroscience:1 track:1 per:1 rb:1 carnegie:4 redundancy:2 sum:1 angle:1 uncertainty:3 place:2 arrive:1 decide:1 vn:4 patch:1 vivid:1 draw:1 convergent:1 sleep:1 assemble:1 constraint:2 scene:17 aspect:1 span:1 relatively:2 department:1 according:1 peripheral:1 combination:1 mcdonnell:1 smaller:1 across:3 making:2 biologically:1 equation:1 tai:2 mechanism:1 needed:1 know:1 mind:1 singer:1 fed:2 unusual:3 available:2 away:2 appropriate:3 generic:1 blurry:1 rp:1 existence:1 denotes:1 assumes:1 top:1 emotional:1 classical:1 move:5 already:1 question:2 spike:1 mumford:2 receptive:7 saccadic:14 primary:1 said:1 exhibit:3 fovea:6 separate:1 reinforce:1 lateral:1 extent:2 reason:1 dream:1 assuming:1 code:7 modeled:4 minimizing:1 mostly:1 idiosyncratic:1 executed:1 attentively:1 debate:2 design:1 neuron:11 sing:1 variability:1 looking:1 paradox:1 directing:1 frame:1 knierim:1 hypercolumn:20 philosophical:1 coherent:3 learned:1 emu:2 pop:2 established:2 macaque:1 address:1 suggested:4 hayhoe:1 below:1 perception:3 pattern:5 reading:4 program:4 memory:5 explanation:2 belief:1 mouth:2 power:2 event:4 decorrelation:1 natural:7 force:3 eye:36 numerous:1 picture:2 created:2 stella:2 coupled:1 prior:3 understanding:6 acknowledgement:1 literature:1 review:1 relative:1 fully:2 interesting:1 generation:1 allocation:1 age:1 foundation:1 integrate:1 downloaded:1 degree:1 gather:2 consistent:1 principle:2 supported:2 last:2 bias:1 guide:1 understand:1 institute:2 neighbor:2 taking:1 face:1 sparse:4 distributed:1 van:1 overcome:1 feedback:1 cortical:1 world:11 superficially:1 contour:1 embodiment:1 author:2 stuck:1 reinforcement:1 feeling:1 transaction:2 compact:4 logic:1 keep:2 global:1 active:2 instantiation:1 sequentially:1 pittsburgh:4 search:1 why:1 kanade:1 channel:7 nature:3 ballard:2 posing:1 complex:2 constructing:1 convey:1 neuronal:5 tl:1 fashion:1 vr:7 position:3 guiding:1 msec:2 explicit:1 inferring:1 perceptual:6 third:1 wavelet:4 down:1 lamme:1 false:1 texture:1 entropy:13 simply:2 likely:2 visual:27 ordered:1 lewicki:1 saccade:7 foveate:1 constantly:2 conditional:4 content:1 foveation:1 kearns:2 called:2 engaged:1 e:1 la:1 partly:1 select:1 internal:3 support:2 scan:6 irwin:1 newark:1 reactive:1 evaluate:1 phenomenon:1 correlated:1
708
1,648
Online Independent Component Analysis With Local Learning Rate Adaptation Nicol N. Schraudolph Xavier Giannakopoulos nic<Didsia.ch xavier<Didsia.ch IDSIA, Corso Elvezia 36 6900 Lugano, Switzerland http://www.idsia.ch/ Abstract Stochastic meta-descent (SMD) is a new technique for online adaptation of local learning rates in arbitrary twice-differentiable systems. Like matrix momentum it uses full second-order information while retaining O(n) computational complexity by exploiting the efficient computation of Hessian-vector products. Here we apply SMD to independent component analysis, and employ the resulting algorithm for the blind separation of time-varying mixtures. By matching individual learning rates to the rate of change in each source signal's mixture coefficients, our technique is capable of simultaneously tracking sources that move at very different, a priori unknown speeds. 1 Introduction Independent component analysis (ICA) methods are typically run in batch mode in order to keep the stochasticity of the empirical gradient low. Often this is combined with a global learning rate annealing scheme that negotiates the tradeoff between fast convergence and good asymptotic performance. For time-varying mixtures, this must be replaced by a learning rate adaptation scheme. Adaptation of a single, global learning rate, however, facilitates the tracking only of sources whose mixing coefficients change at comparable rates [1], resp. switch all at the same time [2]. In cases where some sources move much faster than others, or switch at different times, individual weights in the unmixing matrix must adapt at different rates in order to achieve good performance. We apply stochastic meta-descent (SMD), a new online adaptation method for local learning rates [3, 4], to an extended Bell-Sejnowski ICA algorithm [5] with natural gradient [6] and kurtosis estimation [7] modifications. The resulting algorithm is capable of separating and tracking a time-varying mixture of 10 sources whose unknown mixing coefficients change at different rates. N. N. Schraudo!ph and X Giannakopou!os 790 2 The SMD Algorithm Given a sequence XQ, Xl, ... of data points, we minimize the expected value of a twice-differentiable loss function fw(x) with respect to its parameters W by stochastic gradient descent: Wt+l = Wt + Pt? 8t , where It (1) and . denotes component-wise multiplication. The local learning rates P are best adapted by exponentiated gradient descent [8, 9], so that they can cover a wide dynamic range while staying strictly positive: lnpt 1 ... npt-l - a fwt (xt) I-t alnp Pt-l . exp(1-t It .Vt) , (2) where Vt and I-t is a global meta-learning rate. This approach rests on the assumption that each element of P affects f w( x) only through the corresponding element of w. With considerable variation, (2) forms the basis of most local rate adaptation methods found in the literature. In order to avoid an expensive exponentiation [10] for each weight update, we typically use the linearization etL ~ 1 + u, valid for small luI, giving (3) where we constrain the multiplier to be at least (typically) (} = 0.1 as a safeguard against unreasonably small - or negative - values. For the meta-level gradient descent to be stable, I-t must in any case be chosen such that the multiplier for P does not stray far from unity; under these conditions we find the linear approximation (3) quite sufficient. Definition of v. The gradient trace v should accurately measure the effect that a change in local learning rate has on the corresponding weight. It is tempting to consider only the immediate effect of a change in Pt on Wt+l: declaring Wt and 8t in (1) to be independent of Pt, one then quickly arrives at ... Vt+l _ aWt+l ~l ... = u npt = ... ~ (4) Pt? Ut However, this common approach [11, 12, 13, 14, 15] fails to take into account the incremental nature of gradient descent: a change in P affects not only the current update of W, but also future ones. Some authors account for this by setting v to an exponential average of past gradients [2, 11, 16]; we found empirically that the method of Almeida et al. [15] can indeed be improved by this approach [3]. While such averaging serves to reduce the stochasticity of the product ?It-l implied by (3) and (4), the average remains one of immediate, single-step effects. It By contrast, Sutton [17, 18] models the long-term effect of P on future weight updates in a linear system by carrying the relevant partials forward through time, as is done in real-time recurrent learning [19]. This results in an iterative update rule for v, which we have extended to nonlinear systems [3, 4]. We define vas an 791 Online leA with Local Rate Adaptation exponential average of the effect of all past changes in p on the current weights: ... Vt+ 1 - = ( 1- A) ~ i=O The forgetting factor 0 into (5) gives ~ d OWt+1 . npt-i L.J 1\ 0 1 ... (5) A ~ 1 is a free parameter of the algorithm. Inserting (1) (6) where H t denotes the instantaneous Hessian of fw(i!) at time t. The approximation in (6) assumes that (Vi> 0) oPt!OPt-i = 0; this signifies a certain dependence on an appropriate choice of meta-learning rate p.. Note that there is an efficient O(n) algorithm to calculate HtVt without ever having to compute or store the matrix H t itself [20]; we shall elaborate on this technique for the case of independent component analysis below. Meta-level conditioning. The gradient descent in P at the meta-level (2) may of course suffer from ill-conditioning just like the descent in W at the main level (1); the meta-descent in fact squares the condition number when v is defined as the previous gradient, or an exponential average of past gradients. Special measures to improve conditioning are thus required to make meta-descent work in non-trivial systems. Many researchers [11, 12, 13, 14] use the sign function to radically normalize the p-update. Unfortunately such a nonlinearity does not preserve the zero-mean property that characterizes stochastic gradients in equilibrium -- in particular, it will translate any skew in the equilibrium distribution into a non-zero mean change in p. This causes convergence to non-optimal step sizes, and renders such methods unsuitable for online learning. Notably, Almeida et al. [15] avoid this pitfall by using a running estimate of the gradient's stochastic variance as their meta-normalizer. In addition to modeling the long-term effect of a change in local learning rate, our iterative gradient trace serves as a highly effective conditioner for the meta-descent: the fixpoint of (6) is given by Vt = [AHt + (I-A) diag(I/Pi)]-llt (7) - a modified Newton step, which for typical values of A (i. e., close to 1) scales with the inverse of the gradient. Consequently, we can expect the product It . Vt in (2) to be a very well-conditioned quantity. Experiments with feedforward multi-layer perceptrons [3, 4] have confirmed that SMD does not require explicit meta-level normalization, and converges faster than alternative methods. 3 Application to leA We now apply the SMD technique to independent component analysis, using the Bell-Sejnowski algorithm [5] as our base method. The goal is to find an unmixing N. N. Schraudolph and X Giannakopoulos 792 matrix W t which - up to scaling and permutation - provides a good linear estimate Vt == WtXt of the independent sources St present in a given mixture signal Xt? The mixture is generated linearly according to Xt = Atst , where At is an unknown (and unobservable) full rank matrix. We include the well-known natural gradient [6] and kurtosis estimation [7] modifications to the basic algorithm, as well as a matrix Pt of local learning rates. The resulting online update for the weight matrix W t is (8) where the gradient D t is given by Dt == 8f;;~~t) = ([Vt ? tanh(Vt)] vt - 1) W t , (9) with the sign for each component of the tanh(Vt) term depending on its current kurtosis estimate. Following Pearlmutter [20], we now define the differentiation operator RVt (g(Wt? 8g(W~r+ rVt) Ir=o u == (10) which describes the effect on 9 of a perturbation of the weights in the direction of Vt. We can use RVt to efficiently calculate the Hessian-vector product (11) where "vee" is the operator that concatenates all columns of a matrix into a single column vector. Since Rv, is a linear operator, we have Rv,(Wt ) RVt (Vt) RVt (tanh(Vd) Vt, Rv, (WtXt) = VtXt, diag( tanh' (Vt?) VtXt , (12) (13) (14) and so forth (cf. [20]). Starting from (9), we apply the RVt operator to obtain Rv,[([Vt ? tanh(Vt)] ytT - 1) Wt] Ht*Vt - + RVt([ Yt ? tanh(Vt)] vt - 1) Wt + [(1 ? diag[tanh'(Vt)]) VtXt vt + [Vt ? tanh(Vd](Vtxt)T] W t ([Vt ? tanh(Vt)] ([ Vt ? tanh(Vt)] vt - 1) Vt vt - 1) Vt (15) In conjunction with the matrix versions of our learning rate update (3) (16) and gradient trace (6) (17) this constitutes our SMD-ICA algorithm. 793 Online leA with Local Rate Adaptation 4 Experiment The algorithm was tested on an artificial problem where 10 sources follow elliptic trajectories according to Xt = (Abase + Al sin(wt) + A2 cos(wt)) St (18) where Abase is a normally distributed mixing matrix, as well as Al and A 2, whose columns represent the axes of the ellipses on which the sources travel. The velocities ware normally distributed around a mean of one revolution for every 6000 data samples. All sources are supergaussian. The ICA-SMD algorithm was implemented with only online access to the data, including on-line whitening [21]. Whenever the condition number of the estimated whitening matrix exceeded a large threshold (set to 350 here), updates (16) and (17) were disabled to prevent the algorithm from diverging. Other parameters settings were It = 0.1, >. = 0.999, and p = 0.2. Results that were not separating the 10 sources without ambiguity were discarded. Figure 1 shows the performance index from [6] (the lower the better, zero being the ideal case) along with the condition number of the mixing matrix, showing that the algorithm is robust to a temporary confusion in the separation. The ordinate represents 3000 data samples, divided into mini-batches of 10 each for efficiency. Figure 2 shows the match between an actual mixing column and its estimate, in the subspace spanned by the elliptic trajectory. The singularity occurring halfway through is not damaging performance. Globally the algorithm remains stable as long as degenerate inputs are handled correctly. 5 Conclusions Once SMD-ICA has found a separating solution, we find it possible to simultaneously track ten sources that move independently at very different, a priori unknown OOr-------~------T_------~------,_------~------~ Error index cond(A)120 ---+--- 50 40 30 Figure 1: Global view of the quality of separation N. N. Schraudolph and X Giannakopoulos 794 Or---.---.,~--,----.----~---.----.----.----r---. Estimation error - -, -2 -3 -4 -5 -, -6~--~--~~--~--~----~--~----~---L----~--~ -2.5 -2 -'.5 -0.5 o 0.5 '.5 2 2.5 Figure 2: Projection of a column from the mixing matrix. Arrows link the exact point with its estimate; the trajectory proceeds from lower right to upper left. speeds. To continue tracking over extended periods it is necessary to handle momentary singularities, through online estimation of the number of sources or some other heuristic solution. SMD's adaptation of local learning rates can then facilitate continuous, online use of ICA in rapidly changing environments. Acknowledgments This work was supported by the Swiss National Science Foundation under grants number 2000-052678.97/1 and 2100-054093.98. References [1] J. Karhunen and P. Pajunen, "Blind source separation and tracking using nonlinear PCA criterion: A least-squares approach", in Proc. IEEE Int. Conf. on Neural Networks, Houston, Texas, 1997, pp. 2147- 2152. [2] N. Murata, K.-R. Milller, A. Ziehe, and S.-i. Amari, "Adaptive on-line learning in changing environments", in Advances in Neural Information Processing Systems, M. C. Mozer, M. I. Jordan, and T . Petsche, Eds. 1997, vol. 9, pp. 599- 605, The MIT Press, Cambridge, MA. [3] N. N. Schraudolph, "Local gain adaptation in stochastic gradient descent", in Proceedings of the 9th International Conference on Artificial Neural Networks, Edinburgh, Scotland, 1999, pp. 569-574, lEE, London, ftp://ftp.idsia.ch/ pub/nic/smd.ps.gz. [4] N. N. Schraudolph, "Online learning with adaptive local step sizes", in Neural Nets - WIRN Vietri-99; Proceedings of the 11th Italian Workshop on Neural Nets, M. Marinaro and R. Tagliaferri, Eds., Vietri suI Mare, Salerno, Italy, 1999, Perspectives in Neural Computing, pp. 151-156, Springer Verlag, Berlin. Online leA with Local Rate Adaptation 795 [5] A. J. Bell and T. J. Sejnowski, "An information-maximization approach to blind separation and blind deconvolution", Neural Computation, 7(6):11291159,1995. [6] S.-i. Amari, A. Cichocki, and H. H. Yang, "A new learning algorithm for blind signal separation", in Advances in Neural Information Processing Systems, D. S. Touretzky, M. C. Mozer, and M. E. Hasselmo, Eds. 1996, vol. 8, pp. 757-763, The MIT Press, Cambridge, MA. [7] M. Girolami and C. Fyfe, "Generalised independent component analysis through unsupervised learning with emergent bussgang properties", in Proc. IEEE Int. Conf. on Neural Networks, Houston, Texas, 1997, pp. 1788-179l. [8] J. Kivinen and M. K. Warmuth, "Exponentiated gradient verSus gradient descent for linear predictors", Tech. Rep. UCSC-CRL-94-16, University of California, Santa Cruz, June 1994. [9] J. Kivinen and M. K. Warmuth, "Additive versus exponentiated gradient updates for linear prediction", in Proc. 27th Annual ACM Symposium on Theory of Computing, New York, NY, May 1995, pp. 209-218, The Association for Computing Machinery. [10] N. N. Schraudolph, "A fast, compact approximation of the exponential function", Neural Computation, 11(4):853-862, 1999. [11] R. Jacobs, "Increased rates of convergence through learning rate adaptation", Neural Networks, 1:295- 307, 1988. [12] T. Tollenaere, "SuperSAB: fast adaptive back propagation with good scaling properties", Neural Networks, 3:561-573, 1990. [13] F. M. Silva and L. B. Almeida, "Speeding up back-propagation", in Advanced Neural Computers, R. Eckmiller, Ed., Amsterdam, 1990, pp. 151-158, Elsevier. [14] M. Riedmiller and H. Braun, "A direct adaptive method for faster backpropagation learning: The RPROP algorithm", in Proc. International Conference on Neural Networks, San Francisco, CA, 1993, pp. 586-591, IEEE, New York. [15] L. B. Almeida, T. Langlois, J. D. Amaral, and A. Plakhov, "Parameter adaptation in stochastic optimization", in On-Line Learning in Neural Networks, D. Saad, Ed., Publications of the Newton Institute, chapter 6. Cambridge University Press, 1999, ftp://146.193. 2 . 131/pub/lba/papers/adsteps . ps .gz. [16] M. E. Harmon and L. C. Baird III, "Multi-player residual advantage learning with general function approximation" , Tech. Rep. WL-TR-1065, Wright Laboratory, WL/ AACF, 2241 Avionics Circle, Wright-Patterson Air Force Base, OH 45433-7308, 1996, http://vvv.leemon.com/papers/sim_tech/sim_tech.ps.gz. [17] R. S. Sutton, "Adapting bias by gradient descent: an incremental version of delta-bar-delta", in Proc. 10th National Conference on Artificial Intelligence. 1992, pp. 171-176, The MIT Press, Cambridge, MA, ftp://ftp.cs.umass.edu/ pub/anv/pub/sutton/sutton-92a.ps.gz. [18] R. S. Sutton, "Gain adaptation beats least squares?", in Proc. 7th Yale Workshop on Adaptive and Learning Systems, 1992, pp. 161-166, ftp://ftp.cs. umass.edu/pub/anv/pub/sutton/sutton-92b.ps.gz. [19] R. Williams and D. Zipser, "A learning algorithm for continually running fully recurrent neural networks", Neural Computation, 1:270-280, 1989. [20] B. A. Pearlmutter, "Fast exact multiplication by the Hessian", Neural Computation, 6(1):147-160,1994. [21] J. Karhunen, E. Oja, L. Wang, R. Vigario, and J. Joutsensalo, "A class of neural networks for independent component analysis", IEEE Trans. on Neural Networks, 8(3):486-504, 1997.
1648 |@word version:2 t_:1 jacob:1 tr:1 npt:3 uma:2 pub:6 past:3 current:3 com:1 must:3 cruz:1 additive:1 vietri:2 update:9 intelligence:1 warmuth:2 scotland:1 provides:1 along:1 ucsc:1 direct:1 symposium:1 notably:1 expected:1 indeed:1 forgetting:1 ica:6 multi:2 globally:1 pitfall:1 actual:1 etl:1 differentiation:1 every:1 braun:1 normally:2 grant:1 continually:1 positive:1 generalised:1 local:14 sutton:7 conditioner:1 ware:1 twice:2 mare:1 co:1 range:1 acknowledgment:1 rvt:7 swiss:1 backpropagation:1 riedmiller:1 empirical:1 bell:3 adapting:1 matching:1 projection:1 oor:1 close:1 salerno:1 operator:4 www:1 yt:1 williams:1 starting:1 independently:1 bussgang:1 rule:1 spanned:1 oh:1 handle:1 variation:1 resp:1 pt:6 exact:2 us:1 element:2 idsia:3 expensive:1 velocity:1 wang:1 calculate:2 mozer:2 environment:2 complexity:1 dynamic:1 carrying:1 patterson:1 efficiency:1 basis:1 emergent:1 chapter:1 leemon:1 fast:4 effective:1 london:1 sejnowski:3 artificial:3 fyfe:1 whose:3 quite:1 heuristic:1 amari:2 itself:1 online:12 sequence:1 differentiable:2 advantage:1 kurtosis:3 net:2 product:4 adaptation:14 inserting:1 relevant:1 rapidly:1 langlois:1 mixing:6 translate:1 achieve:1 degenerate:1 forth:1 normalize:1 exploiting:1 convergence:3 p:5 unmixing:2 incremental:2 converges:1 staying:1 ftp:7 depending:1 recurrent:2 implemented:1 c:2 girolami:1 switzerland:1 direction:1 stochastic:7 require:1 opt:2 singularity:2 strictly:1 around:1 wright:2 exp:1 equilibrium:2 a2:1 estimation:4 proc:6 travel:1 tanh:10 hasselmo:1 wl:2 supergaussian:1 mit:3 modified:1 avoid:2 varying:3 publication:1 conjunction:1 ax:1 june:1 rank:1 tech:2 contrast:1 normalizer:1 elsevier:1 typically:3 italian:1 unobservable:1 abase:2 ill:1 priori:2 retaining:1 special:1 once:1 having:1 represents:1 unsupervised:1 constitutes:1 future:2 others:1 employ:1 oja:1 simultaneously:2 preserve:1 national:2 individual:2 replaced:1 highly:1 mixture:6 arrives:1 capable:2 partial:1 necessary:1 machinery:1 harmon:1 circle:1 increased:1 column:5 modeling:1 cover:1 maximization:1 signifies:1 predictor:1 combined:1 st:2 international:2 lee:1 safeguard:1 quickly:1 ambiguity:1 conf:2 account:2 coefficient:3 int:2 baird:1 vi:1 blind:5 view:1 characterizes:1 pajunen:1 minimize:1 square:3 ir:1 air:1 variance:1 efficiently:1 murata:1 accurately:1 trajectory:3 confirmed:1 researcher:1 llt:1 touretzky:1 whenever:1 ed:5 definition:1 rprop:1 against:1 corso:1 pp:11 gain:2 ut:1 back:2 exceeded:1 dt:1 follow:1 lba:1 improved:1 done:1 just:1 nonlinear:2 o:1 propagation:2 mode:1 quality:1 disabled:1 facilitate:1 effect:7 multiplier:2 xavier:2 laboratory:1 sin:1 criterion:1 confusion:1 pearlmutter:2 silva:1 wise:1 instantaneous:1 common:1 empirically:1 conditioning:3 avionics:1 association:1 cambridge:4 stochasticity:2 nonlinearity:1 stable:2 access:1 whitening:2 owt:1 base:2 perspective:1 italy:1 store:1 verlag:1 certain:1 meta:12 rep:2 continue:1 vt:31 vtxt:4 houston:2 period:1 tempting:1 signal:3 rv:4 full:2 faster:3 adapt:1 match:1 schraudolph:6 long:3 divided:1 ellipsis:1 va:1 prediction:1 basic:1 plakhov:1 vigario:1 normalization:1 represent:1 lea:4 addition:1 annealing:1 source:13 saad:1 rest:1 facilitates:1 vvv:1 jordan:1 zipser:1 yang:1 ideal:1 feedforward:1 iii:1 switch:2 affect:2 reduce:1 tradeoff:1 texas:2 handled:1 pca:1 suffer:1 render:1 hessian:4 cause:1 york:2 santa:1 ten:1 ph:1 fixpoint:1 nic:2 http:2 sign:2 estimated:1 delta:2 correctly:1 track:1 shall:1 vol:2 eckmiller:1 threshold:1 changing:2 prevent:1 ht:1 halfway:1 run:1 inverse:1 exponentiation:1 separation:6 scaling:2 comparable:1 layer:1 smd:11 yale:1 annual:1 adapted:1 constrain:1 aht:1 speed:2 according:2 fwt:1 describes:1 unity:1 modification:2 remains:2 awt:1 skew:1 serf:2 apply:4 appropriate:1 elliptic:2 petsche:1 batch:2 alternative:1 supersab:1 assumes:1 denotes:2 unreasonably:1 running:2 include:1 cf:1 anv:2 newton:2 unsuitable:1 giving:1 implied:1 move:3 quantity:1 dependence:1 gradient:23 subspace:1 link:1 separating:3 berlin:1 vd:2 trivial:1 index:2 mini:1 vee:1 unfortunately:1 trace:3 negative:1 unknown:4 upper:1 discarded:1 descent:14 beat:1 immediate:2 extended:3 ever:1 perturbation:1 arbitrary:1 ordinate:1 required:1 california:1 temporary:1 ytt:1 trans:1 bar:1 proceeds:1 below:1 including:1 natural:2 force:1 kivinen:2 residual:1 advanced:1 scheme:2 improve:1 gz:5 cichocki:1 speeding:1 xq:1 literature:1 multiplication:2 nicol:1 asymptotic:1 loss:1 expect:1 permutation:1 fully:1 declaring:1 versus:2 foundation:1 sufficient:1 pi:1 negotiates:1 course:1 supported:1 free:1 bias:1 exponentiated:3 institute:1 wide:1 distributed:2 edinburgh:1 giannakopoulos:3 valid:1 author:1 forward:1 adaptive:5 san:1 far:1 amaral:1 compact:1 keep:1 global:4 francisco:1 continuous:1 iterative:2 nature:1 concatenates:1 robust:1 ca:1 joutsensalo:1 diag:3 main:1 wtxt:2 linearly:1 arrow:1 elaborate:1 ny:1 fails:1 momentum:1 stray:1 explicit:1 momentary:1 exponential:4 lugano:1 xl:1 sui:1 xt:4 revolution:1 showing:1 deconvolution:1 workshop:2 elvezia:1 linearization:1 conditioned:1 occurring:1 karhunen:2 amsterdam:1 tracking:5 springer:1 ch:4 radically:1 acm:1 ma:3 goal:1 consequently:1 crl:1 considerable:1 change:9 fw:2 typical:1 lui:1 wt:10 averaging:1 diverging:1 tollenaere:1 player:1 cond:1 perceptrons:1 ziehe:1 damaging:1 almeida:4 tested:1
709
1,649
Local probability propagation for factor analysis Brendan J. Frey Computer Science, University of Waterloo, Waterloo, Ontario, Canada Abstract Ever since Pearl's probability propagation algorithm in graphs with cycles was shown to produce excellent results for error-correcting decoding a few years ago, we have been curious about whether local probability propagation could be used successfully for machine learning. One of the simplest adaptive models is the factor analyzer, which is a two-layer network that models bottom layer sensory inputs as a linear combination of top layer factors plus independent Gaussian sensor noise. We show that local probability propagation in the factor analyzer network usually takes just a few iterations to perform accurate inference, even in networks with 320 sensors and 80 factors. We derive an expression for the algorithm's fixed point and show that this fixed point matches the exact solution in a variety of networks, even when the fixed point is unstable. We also show that this method can be used successfully to perform inference for approximate EM and we give results on an online face recognition task. 1 Factor analysis A simple way to encode input patterns is to suppose that each input can be wellapproximated by a linear combination of component vectors, where the amplitudes of the vectors are modulated to match the input. For a given training set, the most appropriate set of component vectors will depend on how we expect the modulation levels to behave and how we measure the distance between the input and its approximation. These effects can be captured by a generative probabilit~ model that specifies a distribution p(z) over modulation levels z = (Zl, ... ,ZK) and a distribution p(xlz) over sensors x = (Xl, ... ,XN)T given the modulation levels. Principal component analysis, independent component analysis and factor analysis can be viewed as maximum likelihood learning in a model of this type, where we assume that over the training set, the appropriate modulation levels are independent and the overall distortion is given by the sum of the individual sensor distortions. In factor analysis, the modulation levels are called factors and the distributions have the following form: p(Zk) = N(Zk; 0,1), p(z) = nf=lP(Zk) = N(z; 0, I), p(xnl z ) = N(xn; E~=l AnkZk, 'l/Jn), p(xlz) = n:=IP(xnlz) = N(x; Az, 'It). (1) The parameters of this model are the factor loading matrix A, with elements Ank, and the diagonal sensor noise covariance matrix 'It, with diagonal elements 'l/Jn. A belief network for the factor analyzer is shown in Fig. 1a. The likelihood is p(x) = 1 N(z; 0, I)N(x; Az, 'It)dz = N(x; 0, AA T + 'It), (2) 443 Local Probability Propagation for Factor Analysis (b) - ...., -... - 'J " ... 'r .... 1t. , E '" "' I:~ ~. '. Figure 1: (a) A belief network for factor analysis. (b) High-dimensional data (N = 560). and online factor analysis consists of adapting A and q, to increase the likelihood of the current input, such as a vector of pixels from an image in Fig. lb. Probabilistic inference - computing or estimating p{zlx) - is needed to do dimensionality reduction and to fill in the unobserved factors for online EM-type learning. In this paper, we focus on methods that infer independent factors. p(zlx) is Gaussian and it turns out that the posterior means and variances of the factors are E[zlx] = (ATq,-l A + 1)-1 AT q,-lx, diag(COV(zlx)) = diag(AT q,-l A + 1)-1). (3) Given A and q" computing these values exactly takes O(K2 N) computations, mainly because of the time needed to compute AT q,-l A. Since there are only K N connections in the network, exact inference takes at least O{K) bottom-up/top down iterations. Of course, if the same network is going to be applied more than K times for inference (e.g., for batch EM), then the matrices in (3) can be computed once and reused. However, this is not directly applicable in online learning and in biological models. One way to circumvent computing the matrices is to keep a separate recognition network, which approximates E[zlx] with Rx (Dayan et al., 1995). The optimal recognition network, R = (A Tq,-l A+I)-l ATq,-l, can be approximated by jointly estimating the generative network and the recognition network using online wakesleep learning (Hinton et al., 1995). 2 Probability propagation in the factor analyzer network Recent results on error-correcting coding show that in some cases Pearl's probability propagation algorithm, which does exact probabilistic inference in graphs that are trees, gives excellent performance even if the network contains so many cycles that its minimal cut set is exponential (Frey and MacKay, 1998; Frey, 1998; MacKay, 1999). In fact, the probability propagation algorithm for decoding lowdensity parity-check codes (MacKay, 1999) and turbocodes (Berrou and Glavieux, 1996) is widely considered to be a major breakthrough in the information theory community. When the network contains cycles, the local computations give rise to an iterative algorithm, which hopefully converges to a good answer. Little is known about the convergence properties of the algorithm. Networks containing a single cycle have been successfully analyzed by Weiss (1999) and Smyth et al. (1997), but results for networks containing many cycles are much less revealing. The probability messages produced by probability propagation in the factor analyzer network of Fig. 1a are Gaussians. Each iteration of propagation consists of passing a mean and a variance along each edge in a bottom-up pass, followed by passing a mean and a variance along each edge in a top-down pass. At any instant, the 444 B.J. Frey bottom-up means and variances can be combined to form estimates of the means and variances of the modulation levels given the input. Initially, the variance and mean sent from the kth top layer unit to the nth sensor is set to vk~ = 1 and 7]i~ = 0. The bottom-up pass begins by computing a noise level and an error signal at each sensor using the top-down variances and means from the previous iteration: s~) = 'l/Jn + 2:{:=1 A;kVk~-I) , e~) = Xn - 2: {:=1Ank7]i~-l). (4) These are used to compute bottom-up variances and means as follows: ",(i) = s(i)/A2 _ v(i-l) lI(i) = e(i)/A k + 7](i-l) (5) 'l'nk n nk kn' r'nk n n kn' The bottom-up variances and means are then combined to form the current estimates of the modulation variances and means: (i) Vk N (i) A(i) _ = 1/(1 + 2:n=1 1/?nk)' Zk - (i)"",N Vk (i)/",(i) L..Jn=lJ.tnk 'l'nk' (6) The top-down pass proceeds by computing top-down variances and means as follows: vk~ = l/(l/vii ) - l/?~l), 7]i~ = vk~(.iki) /vii ) - J.t~V?~l)? (7) Notice that the variance updates are independent of the mean updates, whereas the mean updates depend on the variance updates. 2.1 Performance of local probability propagation. We created a total of 200,000 factor analysis networks with 20 different sizes ranging from K = 5, N = 10 to K = 80, N = 320 and for each size of network we measured the inference error as a function of the number of iterations of propagation. Each of the 10,000 networks of a given size was produced by drawing the AnkS from standard normal distributions and then drawing each sensor variance 'l/Jn from an exponential distribution with mean 2:{:=1 A;k' (A similar procedure was used by Neal and Dayan (1997).) For each random network, a pattern was simulated from the network and probability propagation was applied using the simulated pattern as input. We measured the error between the estimate z(i) and the correct value E[zlx] by computing the difference between their coding costs under the exact posterior distribution and then normalizing by K to get an average number of nats per top layer unit. Fig. 2a shows the inference error on a logarithmic scale versus the number of iterations (maximum of 20) in the 20 different network sizes. In all cases, the median error is reduced below .01 nats within 6 iterations. The rate of convergence of the error improves for larger N, as indicated by a general trend for the error curves to drop when N is increased. In contrast, the rate of convergence of the error appears to worsen for larger K, as shown by a general slight trend for the error curves to rise when K is increased. For K ~ N/8, 0.1% of the networks actually diverge. To better understand the divergent cases, we studied the means and variances for all of the divergent networks. In all cases, the variances converge within a few iterations whereas the means oscillate and diverge. For K = 5, N = 10, 54 of the 10,000 networks diverged and 5 of these are shown in Fig. 2b. This observation suggests that in general the dynamics are determined by the dynamics of the mean updates. 2.2 Fixed points and a condition for global convergence. When the variance updates converge, the dynamics of probability propagation in factor analysis networks become linear. This allows us to derive the fixed point of propagation in closed form and write an eigenvalue condition for global convergence. 445 Local Probability Propagation for Factor Analysis (a) K = 5 K ~',:~ ~ 01~ = 10 K=20 K=40 K=80 g"Xl~ 10 ... II ~ 1:~ 11' 0 ~ ~ .01 ~ ',: u 2: 0, ~ ',: ~ ~ o 0' ,OO~ ~' O~ ~ 01 0 10 20 Figure 2: (a) Performance of probability propagation . Median inference error (bold curve) on a logarithmic scale as a function of the number of iterations for different sizes of network parameterized by K and N. The two curves adjacent to the bold curve show the range within which 98% of the errors lie. 99 .9% of the errors were below the fourth, topmost curve. (b) The error, bottom-up variances and top-down means as a function of the number of iterations (maximum of 20) for 5 divergent networks of size K = 5, N = 10. To analyze the system of mean updates, we define the following length K N vec. - (i) _ ( (i) (i) (i) (i ) (i))T - (i) _ ? tors 0 f means an d t he mput . TJ - 1711,1721"'" 17Kl' 1712' " . , 17KN , P, (i) (i) (i ) (i) (i) )T ( )T ( J-tll,J-t12 ' ''' ,J-tlK , J-t21'''' , J-tNK , X= Xl,Xl, .. ? ,Xl,X2, .. ? , X2 , XN, .. ? ,XN , where each Xn is repeated K times in the last vector. The network parameters are represented using K N x K N diagonal matrices, A and q,. The diagonal of A is A11, ... , AIK , A21, ... , ANK, and the diagonal of q, is '1/111, '1/121, ... , '1/INI, where 1 is the K x K identity matrix. The converged bottom-up variances are represented using a diagonal matrix ~ with diagonal ?11, ... , ?IK , ?21, .. . , ?NK. The summation operations in the propagation formulas are represented by a K N x K N matrix I: z that sums over means sent down from the top layer and a K N x K N matrix I: x that sums over means sent up from the sensory input: 1 1 ) :Ex = ' (~i 1 1 (8) 1 These are N x N matrices of K x K blocks, where 1 is the K x K block of ones and 1 is the K x K identity matrix. Using the above representations, the bottom-up pass is given by ji, (i) = A-I X _ A- I (:E z - I)Af7(i-l), (9) and the top-down pass is given by f7( i) = (I + diag(:Ex~ -1 :Ex) _ ~ -1) -1 (I: x _ I)~ -1 ji,( i ) . Substituting (10) into (9), we get the linear update for ji,(i) = A-I X _ A-I (:E z _ (10) ji,: I)A(I + diag(:Exci -l:E x) _ c) -1) -1 (:Ex _ I)ci -1 ji, (i -l). (11) B.J. Frey 446 B[]Bga~Q~g[] 1.24 1.07 1.49 1.13 1.03 1.02 1.09 1.01 1.11 1.06 Figure 3: The error (log scale) versus number of iterations (log scale. max. of 1000) in 10 of the divergent networks with K = 5. N = 10. The means were initialized to the fixed point solutions and machine round-off errors cause divergence from the fixed points. whose errors are shown by horizontal lines. The fixed point of this dynamic system, when it exists, is ji,* = ~ (A~ + (tz - I)A(I + diag(I:xc) -lt x) - ~ -1) -\t x - I)) -1 x. (12) A fixed point exists if the determinant of the expression in large braces in (12) is nonzero. We have found a simplified expression for this determinant in terms of the determinants of smaller, K x K matrices. Reinterpreting the dynamics in (11) as dynamics for Aji,(i), the stability of a fixed point is determined by the largest eigenvalue of the update matrix, (I: z - I)A (I + - - -1 - - -1 -1 - - -1 - -1 diag(Exc}) Ex)-c}) ) (Ex-I)c}) A . If the modulus ofthe largest eigenvalue is less than 1, the fixed point is stable. Since the system is linear, if a stable fixed point exists, the system will be globally convergent to this point. Of the 200,000 networks we explored, about 99.9% of the networks converged. For 10 of the divergent networks with K = 5, N = 10, we used 1000 iterations of probability propagation to compute the steady state variances. Then, we computed the modulus of the largest eigenvalue of the system and we computed the fixed point. After initializing the bottom-up means to the fixed point values, we performed 1000 iterations to see if numerical errors due to machine precision would cause divergence from the fixed point. Fig. 3 shows the error versus number of iterations (on logarithmic scales) for each network, the error of the fixed point, and the modulus of the largest eigenvalue. In some cases, the network diverges from the fixed point and reaches a dynamic equilibrium that has a lower average error than the fixed point. 3 Online factor analysis To perform maximum likelihood factor analysis in an online fashion, each parameter should be modified to slightly increase the log-probability of the current sensory input,logp(x). However, since the factors are hidden, they must be probabilistically "filled in" using inference before an incremental learning step is performed. If the estimated mean and variance of the kth factor are Zk and Vk, then it turns out (e.g., Neal and Dayan, 1997) the parameters can be updated as follows: Ank 'IjIn where 1} f- f- Ank + l}[Zk(Xn - Ef=1 AnjZj) - (l-l})'ljln + l}[(xn - VkAnk]/'ljln, Ef=1 AnjZj)2 + Ef=1 VkA~j], (13) is a learning rate. Online learning consists of performing some number of iterations of probability propagation for the current input (e.g., 4 iterations) and then modifying the parameters before processing the next input. 3.1 Results on simulated data. We produced 95 training sets of 200 cases each, with input sizes ranging from 20 sensors to 320 sensors. For each of 19 sizes of factor analyzer, we randomly selected 5 sets of parameters as described above and generated a training set. The factor analyzer sizes were K E {5, 10,20,40, 80}, Local Probability Propagation for Factor Analysis 447 Figure 4: (a) Achievable errors after the same number of epochs of learning using 4 iterations versus 1 iteration. The horizontal axis gives the log-probability error (log scale) for learning with 1 iteration and the vertical axis gives the error after the same number of epochs for learning with 4 iterations. (b) The achievable errors for learning using 4 iterations of propagation versus wake-sleep learning using 4 iterations. > K. For each factor analyzer and simulated data set, we estimated the optimal log-probability of the data using 100 iterations of EM. N E {20, 40, 80,160, 320}, N For learning, the size of the model to be trained was set equal to the size of the model that was used to generate the data. To avoid the issue of how to schedule learning rates, we searched for achievable learning curves, regardless of whether or not a simple schedule for the learning rate exists. So, for a given method and randomly initialized parameters, we performed one separate epoch of learning using each of the learning rates, 1,0.5, ... ,0.5 20 and picked the learning rate that most improved the log-probability. Each successive learning rate was determined by comparing the performance using the old learning rate and one 0.75 times smaller. We are mainly interested in comparing the achievable curves for different methods and how the differences scale with K and N. For two methods with the same K and N trained on the same data, we plot the log-probability error (optimal logprobability minus log-probability under the learned model) of one method against the log-probability error of the other method. Fig. 4a shows the achievable errors using 4 iterations versus using 1 iteration. Usually, using 4 iterations produces networks with lower errors than those learned using 1 iteration. The difference is most significant for networks with large K, where in Sec. 2.1 we found that the convergence of the inference error was slower. Fig. 4b shows the achievable errors for learning using 4 iterations of probability propagation versus wake-sleep learning using 4 iterations. Generally, probability propagation achieves much smaller errors than wake-sleep learning, although for small K wake-sleep performs better very close to the optimum log-probability. The most significant difference between the methods occurs for large K, where aside from local optima probability propagation achieves nearly optimal log-probabilities while the log-probabilities for wake-sleep learning are still close to their values at the start of learning. 4 Online face recognition Fig. 1b shows examples from a set of 30,000 20 x 28 greyscale face images of 18 different people. In contrast to other data sets used to test face recognition methods, these faces include wide variation in expression and pose. To make classification more difficult, we normalized the images for each person so that each pixel has B.J. Frey 448 the same mean and variance. We used probability propagation and a recognition network in a factor analyzer to reduce the dimensionality of the data online from 560 dimensions to 40 dimensions. For probability propagation, we rather arbitrarily chose a learning rate of 0.0001, but for wake-sleep learning we tried learning rates ranging from 0.1 down to 0.0001. A multilayer perceptron with one hidden layer of 160 tanh units and one output layer of 18 softmax units was simultaneously being trained using gradient descent to predict face identity from the mean factors. The learning rate for the multilayer perceptron was set to 0.05 and this value was used for both methods. For each image, a prediction was made before the parameters were modified. Fig. 5 shows online error curves obtained by filtering the losses. The curve for probability propagation is generally below the curves for wake-sleep learning. The figure also shows the error curves for two forms of online nearest neighbors, where only the most recent W cases are used to make a prediction. The form of nearest neighbors that performs the worst has W set so that the storage requirements are the same as for the factor analysis / multilayer perceptron method. The better form of nearest neighbors has W set so that the number of computations is the same as for the factor analysis / multilayer perceptron method. 5 Summary ~ j "' " \ ' i .. ',"--""~ \<',::,::'--, ...,... "" '. ~, '. Number of pattern presentations Figure 5: Online error curves for probability propagation (solid), wake-sleep learning (dashed), nearest neighbors (dot-dashed) and guessing (dotted). It turns out that iterative probability propagation can be fruitful when used for learning in a graphical model with cycles, even when the model is densely connected. Although we are more interested in extending this work to more complex models where exact inference takes exponential time, studying iterative probability propagation in the factor analyzer allowed us to compare our results with exact inference and allowed us to derive the fixed point of the algorithm. We are currently applying iterative propagation in multiple cause networks for vision problems. References C. Berrou and A. Glavieux 1996. Near optimum error correcting coding and decoding: Turbo-codes. IEEE TI-ans. on Communications, 44, 1261-1271. P. Dayan, G. E. Hinton, R. M. Neal and R. S. Zemel 1995. The Helmholtz machine. Neural Computation 1, 889-904. B. J. Frey and D. J. C. MacKay 1998. A revolution: Belief propagation in graphs with cycles. In M. Jordan, M. Kearns and S. Solla (eds), Advances in Neural Information Processing Systems 10, Denver, 1997. B. J. Frey 1998. Graphical Models for Machine Learning and Digital Communication. MIT Press, Cambridge MA. See http://wvv.cs.utoronto.ca/-frey . G. E. Hinton, P. Dayan, B. J. Frey and R. M. Neal 1995. The wake-sleep algorithm for unsupervised neural networks. Science 268, 1158-1161. D. J. C. MacKay 1999. Information Theory, Inference and Learning Algorithms. Book in preparation, currently available at http://wol.ra.phy.cam.ac . uk/mackay. R. M. Neal and P. Dayan 1997. Factor analysis using delta-rule wake-sleep learning. Neural Computation 9, 1781-1804. P. Smyth, R. J . McEliece, M. Xu, S. Aji and G. Horn 1997. Probability propagation in graphs with cycles. Presented at the workshop on Inference and Learning in Graphical Models, Vail, Colorado. Y. Weiss 1998. Correctness of local probability propagation in graphical models. To appear in Neural Computation.
1649 |@word determinant:3 achievable:6 loading:1 reused:1 iki:1 tried:1 covariance:1 minus:1 solid:1 wellapproximated:1 reduction:1 phy:1 contains:2 current:4 comparing:2 must:1 numerical:1 drop:1 plot:1 update:9 aside:1 generative:2 selected:1 lx:1 successive:1 along:2 become:1 ik:1 consists:3 reinterpreting:1 ra:1 globally:1 little:1 begin:1 estimating:2 unobserved:1 nf:1 ti:1 exactly:1 k2:1 uk:1 zl:1 unit:4 appear:1 before:3 local:10 frey:10 modulation:7 plus:1 chose:1 studied:1 suggests:1 range:1 horn:1 block:2 procedure:1 aji:2 probabilit:1 adapting:1 revealing:1 get:2 close:2 storage:1 applying:1 fruitful:1 xlz:2 dz:1 regardless:1 correcting:3 rule:1 wvv:1 fill:1 stability:1 variation:1 updated:1 suppose:1 colorado:1 aik:1 exact:6 smyth:2 element:2 trend:2 recognition:7 approximated:1 helmholtz:1 cut:1 bottom:11 initializing:1 worst:1 t12:1 cycle:8 connected:1 solla:1 topmost:1 nats:2 turbocodes:1 cam:1 dynamic:7 trained:3 depend:2 represented:3 zemel:1 whose:1 widely:1 larger:2 distortion:2 drawing:2 cov:1 jointly:1 ip:1 online:13 eigenvalue:5 t21:1 ontario:1 az:2 convergence:6 optimum:3 diverges:1 requirement:1 produce:2 a11:1 incremental:1 converges:1 extending:1 derive:3 oo:1 ac:1 pose:1 measured:2 nearest:4 c:1 tlk:1 correct:1 modifying:1 wol:1 biological:1 summation:1 considered:1 normal:1 equilibrium:1 diverged:1 predict:1 substituting:1 major:1 tor:1 achieves:2 a2:1 f7:1 applicable:1 tanh:1 currently:2 waterloo:2 largest:4 correctness:1 successfully:3 mit:1 sensor:10 gaussian:2 modified:2 rather:1 avoid:1 tll:1 probabilistically:1 encode:1 focus:1 vk:6 likelihood:4 mainly:2 check:1 contrast:2 brendan:1 inference:15 dayan:6 lj:1 initially:1 hidden:2 going:1 interested:2 pixel:2 overall:1 issue:1 classification:1 breakthrough:1 mackay:6 softmax:1 equal:1 once:1 unsupervised:1 nearly:1 few:3 randomly:2 simultaneously:1 divergence:2 densely:1 individual:1 tq:1 message:1 analyzed:1 kvk:1 tj:1 accurate:1 edge:2 tree:1 filled:1 old:1 initialized:2 minimal:1 increased:2 logp:1 cost:1 kn:3 answer:1 combined:2 person:1 probabilistic:2 off:1 decoding:3 diverge:2 containing:2 tz:1 book:1 li:1 coding:3 bold:2 sec:1 performed:3 picked:1 closed:1 analyze:1 start:1 worsen:1 lowdensity:1 variance:22 ofthe:1 produced:3 rx:1 ago:1 converged:2 reach:1 ed:1 against:1 dimensionality:2 improves:1 schedule:2 amplitude:1 actually:1 appears:1 wei:2 improved:1 just:1 mceliece:1 horizontal:2 hopefully:1 propagation:35 indicated:1 modulus:3 effect:1 normalized:1 nonzero:1 neal:5 adjacent:1 round:1 steady:1 ini:1 vail:1 performs:2 image:4 ranging:3 ef:3 ji:6 denver:1 tnk:2 slight:1 approximates:1 he:1 significant:2 cambridge:1 vec:1 analyzer:10 dot:1 stable:2 posterior:2 recent:2 arbitrarily:1 captured:1 converge:2 berrou:2 signal:1 ii:1 dashed:2 multiple:1 infer:1 match:2 prediction:2 multilayer:4 vision:1 iteration:29 whereas:2 ank:4 wake:10 median:2 brace:1 sent:3 jordan:1 curious:1 near:1 variety:1 reduce:1 whether:2 expression:4 passing:2 oscillate:1 cause:3 logprobability:1 generally:2 simplest:1 reduced:1 generate:1 specifies:1 http:2 notice:1 dotted:1 estimated:2 delta:1 per:1 write:1 graph:4 year:1 sum:3 parameterized:1 fourth:1 atq:2 layer:8 followed:1 convergent:1 sleep:10 turbo:1 x2:2 performing:1 combination:2 smaller:3 slightly:1 em:4 lp:1 turn:3 needed:2 studying:1 available:1 gaussians:1 operation:1 appropriate:2 batch:1 slower:1 jn:5 top:11 include:1 graphical:4 instant:1 xc:1 occurs:1 diagonal:7 guessing:1 gradient:1 kth:2 distance:1 separate:2 simulated:4 exc:1 unstable:1 code:2 length:1 difficult:1 greyscale:1 rise:2 perform:3 vertical:1 observation:1 descent:1 behave:1 hinton:3 ever:1 communication:2 lb:1 community:1 canada:1 kl:1 connection:1 learned:2 zlx:6 pearl:2 proceeds:1 usually:2 pattern:4 below:3 max:1 belief:3 circumvent:1 nth:1 axis:2 created:1 epoch:3 wakesleep:1 expect:1 loss:1 filtering:1 versus:7 digital:1 xnl:1 course:1 summary:1 parity:1 last:1 understand:1 perceptron:4 wide:1 neighbor:4 face:6 curve:13 dimension:2 xn:8 sensory:3 made:1 adaptive:1 simplified:1 approximate:1 keep:1 global:2 iterative:4 zk:7 ca:1 excellent:2 complex:1 diag:6 noise:3 repeated:1 allowed:2 xu:1 fig:10 fashion:1 precision:1 a21:1 exponential:3 xl:5 lie:1 down:9 formula:1 utoronto:1 revolution:1 explored:1 divergent:5 normalizing:1 exists:4 workshop:1 ci:1 nk:6 vii:2 exci:1 logarithmic:3 lt:1 aa:1 ma:1 glavieux:2 viewed:1 identity:3 presentation:1 determined:3 principal:1 kearns:1 called:1 total:1 pas:6 searched:1 mput:1 people:1 modulated:1 preparation:1 ex:6
710
165
537 A MASSIVELY PARALLEL SELF-TUNING CONTEXT-FREE PARSER! Eugene Santos Jr. Department of Computer Science Brown University Box 1910, Providence, RI 02912 eSj@cs.brown.edu ABSTRACT The Parsing and Learning System(PALS) is a massively parallel self-tuning context-free parser. It is capable of parsing sentences of unbounded length mainly due to its parse-tree representation scheme. The system is capable of improving its parsing performance through the presentation of training examples. INTRODUCTION Recent PDP research[Rumelhart et al .? 1986; Feldman and Ballard, 1982; Lippmann, 1987] involving natural language processtng[Fanty, 1988; Selman, 1985; Waltz and Pollack, 1985] have unrealistically restricted sentences to a fixed length. A solution to this problem was presented in the system CONPARSE[Charniak and Santos. 1987]. A parse-tree representation scheme was utilized which allowed for processing sentences of any length. Although successful as a parser. it's achitecture was strictly hand-constructed with no learning of any form. Also. standard learning schemes were not appUcable since it differed from all the popular architectures, in particular. connectionist ones. In this paper. we present the Parsing and Learning System(PALS) which attempts to integrate a learning scheme into CONPARSE. It basically allows CONPARSE to modify and improve its parsing capability. IThis research was supported in part by the Office of Naval Research under contract NOOOI4-79-C-0592, the National Science Foundation under contracts IST-8416034 and IST-8515005, and by the Defense Advanced Research Projects Agency under ARPA Order No. 4786. 538 Santos REPRESENTATION OF PARSE TREE A parse-tree Is represented by a matrix where the bottom row consists of the leaves of the tree In left-to-right order and the entries In each column above each leaf correspond to the nodes In the path from leaf to root. For example, looking at the simple parse-tree for the sentence "noun verb noun", the column entries for verb would be verb, vp. and S. (see Figure 1) (As In previous work, PALS takes part-of-speech as input, not words.) S S NP noun ,..S """'IIIl VP VP NP noun ... verb..oil Figure 1. Parse tree as represented by a collection of columns in the matrix. In addition to the columns of nontermlnals. we introduce the binder entries as a means of easily determining whether two Identical nonterminals in adjacent columns represent the same nonterminal in a parse tree (see Figure 2). , S NP noun -S VP verb / / S ~VP NP noun s ~ NP VP noun NP I verb '" noL Figure 2. Parse tree as represented by a collection of columns in the matrix plus binders. To distributively represent the matrix. each entry denotes a collection of labeled computational units. The value of the entry is taken to be the label of the unit with the largest value. A Massively Parallel Self-Tuning Context-Free Parser A nontenninal entry has units which are labeled with the nontenninals of a language plus a special label ''blank''. When the "blank" unit is largest, this indicates that the entry plays no part in representing the current parse tree. binder entry has units which are labeled from 1 to the number of rows in the matrix. A unit labeled k then denotes the binding of the nontenninal entry on its immediate left to the nontenninal entry in the kth row on its right. To indicate that no binding exists, we use a special unit label "e" called an edge. A In general, it is easiest to view an entry as a vector of real numbers where each vector component denotes some symbol. (For more infonnation see [Charntak and Santos, 1987].) In the current implementation of PALS, entry unit values range from 0 to 1. The ideal value for entry units is thus 1 for the largest entry unit and 0 for all remaining entry units. We essentially have "1" being "yes" and "0" being no. LANGUAGE RULES In order to determine the values of the computational units mentioned in the previous section, we apply a set of language rules. Each compuatatlonal unit will be detenntned by some subset of these rules. Each language rule is represented by a single node, called a rule node. A rule node takes its input from several computational units and outputs to a Single computational unit. The output of each rule node is also modified by a non-negative value called a rule-weight. This weight represents the applicability of a language rule to the language we are attempting to parse (see PARSING). In the current implementation of PALS, rule-weight values range from 0 to 1 being similar to probabilities. Basically, a rule node attempts to express some rule of grammar. As with CONPARSE, PALS breaks context-free grammars into several subrules. For example, as part of S --> NP VP, PALS would have a rule stating that an NP entry would like to have an S immediately above it in the same column. Our rule for this grammar rule will then take as input the entry's computational unit labeled NP and output to the unit labeled S in the entry immediately above(see Figure 3). 539 540 Santos Entry iJ Rule-Node Entry i-l,j Figure 3. A rule node for S above NP. a more complex example, if an entry is a NP, the NP does not continue right. I.e .. has an edge, and above is an S that continues to the right. then below the second S is a VP. As In general. to determine a unit's value, we take all the rule nodes and combine their influences. This will be much clearer when we discuss parsing in the next section. PARSING Since we are dealing with a PDP-type architecture, the size of our matrix is fixed ahead of time. However. the way we use the matrix representation scheme allows us to handle sentences of unbounded length as we shall see. The system parses a sentence by taking the first word and placing it in the lower rightmost entry; it then attempts to construct the column above the word by using its rule nodes. Mter this processing. the system shifts the first word and its column left and inserts the second word. Now both words are processed simultaneously. This shifting and processing continues until the last word is shifted through the matrix (see Figure 4). Since sentence lengths may exceed the size of the matrix. we are only processing a portion at a time. creating partial parse-trees. The complete parse-tree is the combination of these partial ones. A Massively Parallel Self-Tuning Context-Free Parser 8 NP noun notDl verb t,f8 8 8 1- 1-8 NP VP noun verb noun - ~8 / t,fVP NP VP / NP n01Dl verb notDl Figure 4. Parsing of noun verb noun. Basically, the system builds the tree in a bottom-up fashion. However, it can also build left-right, right-left, and top-down since columns may be of differing height. In general, columns on the left in the matrix will be more complete and hence possibly higher than those on the right. LEARNING The goal of PALS is to learn how to parse a given language. Given a system consisting of a matrix with a set of language rules. we learn parsing by determining how to apply each language rule. In general, when a language rule is inconsistent with the language we are learning to parse, its corresponding rule-weight drops to zero, essentially disconnecting the rule. When a language rule is consistent, its ruleweight then approaches one. In PALS, we learn how to parse a sentence by using training examples. The teacher/trainer gives to the system the complete parse tree of the sentence to be learned. Because of the restrictions imposed by our matrix, we may be unable to fully represent the complete parse tree given by the teacher. To learn how to parse the sentence, we can only utilize a portion of the complete parse tree at anyone time. Given a complete parse tree. the system simply breaks it up into manageable chunks we call snapshots. Snapshots are matrices which contain a portion of the complete parse tree. Given this sequence of snapshots, we present them to the system in a fashion Similar to parsing. The only difference is that we clamped the 541 542 Santos snapshot to the system matrix while it fires its rule nodes. From this, we can easily detenntne whether a rule node has incorrectly fired or not by seeing if it fired consistently with given snapshot. We punish and reward accordingly. As the system is trained more and more, our rule-weights contain more and more information. We would like the rule-weights of those rules used frequently during training to not change as much as those not frequently used. This serves to stabilize our knowledge. It also prevents the system from being totally corrupted when presented with an incorrect training example. As in traditional methods, we find the new rule-weights by minimizing some function which gives us our desired learning. The function which captures this learning method is Lt,j {CfJ ( Clf.,j - ~i,j )2 + [ ~,j ~i,j + ( 1 - ~,j ) ( 1 - ~iJ ) I 2 ri,j2} where i are the unit labels for some matrix entry. j are the language rules associated with units i, (l1.j are the old rule-weights. ~i,j are the new ruleweights, ci,J is the knowledge preservation coefficient which is a function of the frequency that language rule j for entry unit i has been fired during learning. ri.j is the unmodified rule output using snapshot as input, and Si,j is the measure of the correctness of language rule j for unit 1. RESULTS In the current implementation of PALS, we utilize a 7x6 matrix and an average of fifty language rules per entry unit to parse English. Obviously, our set of language rules will determine what we can and cannot learn. Currently, the system is able to learn and parse a modest subset of the English language. It can parse sentences with moderate sentence embedding and phrase attachment from the following grammar: SM S NP --> S per --> NPVP --> (det) (adj)* noun (PP)* (WHCL) (INFPI) NP PP WHCL S/NP S/NP --> INFP2 --> prep NP --> that S/NP -->VP --> NP VP/NP A Massively Parallel Self-Tuning Context-Free Parser INFPl INFI VP VP VP VP VP VP/NP INFP2 INF2 --> (NP) INFI --> to (adv) VP/NP --> (aux) tIVerb NP (PP)* --> (aux) intrverb --> (aux) copverb NP --> (aux) copverb PP --> (aux) copverb adj --> (aux) trverb (PP)* --> (NP) INF2 --> to (adv) VP We have found that sentences only require approximately two trainings. We have also found that by adding more consistent language rules, the system improved by actually generating parse trees which were more "certain" than previously generated ones. In other words, the values of the entry units in the final parse tree were much closer to the ideal. When we added inconsistent language rules, the system degraded. However, with slightly more training, the system was back to normal. It actually had to first eliminate the inconsistent rules before being able to apply the consistent ones. Finally, we attempted to train the system with incorrect training examples after being trained with correct ones. We found that even though the system degraded slightly, previous learning was not completely lost. This was basically due to the stability employed during learning. CONCLUSIONS We have presented a system capable of parsing and learning how to parse. The system parses by creating a sequence of partial parse trees and then combining them into a parse tree. It also places no limit on sentence length. Given a system consisting of a matrix and an aSSOCiated set of language rules, we attempt to learn how to parse the language described by the complete parse tree supplied by a teacher/trainer. The same set of language rules may also be able to learn a different language. Depending on the diversity of the language rules, it may also learn both simultaneously, te., parse both languages. (A simple example is two languages with distinct terminals and nontenntnals.) The system learns by being presented complete parse-trees and adding to its knowledge by modifying its rule-weights. 543 544 Santos The system requires a small number of trainings per training example. Also, incorrect training examples do not totally corrupt the system. PROBLEMS AND FUTURE RESEARCH Eventhough the PALS system has managed to integrate learning. there are still some problems. First, as in the CONPARSE system, we can only handle moderately embedded sentences. Second, the system is very positional. Something that is learned in one portion of the matrix is not generalized to other portions. There is no rule aquisition in the PALS system. Currently, all rules are assumed to be built-in to the system. PALS's ability to suppress incorrect rules would entail rule learning if the possible set of language rules was very tightly constrained so that, in effect. all rules could be tried. Some linguists have suggested quite limited schemes but if any would work with PALS is not known. REFERENCES Rumelhart, D. et ~., ParaUel Distributed Processing: Explorations in the Microstructures oj Cognition. Volume 1, The MIT Press (1986). Charniak, E. and Santos, E., "A connectionist context-free parser which is not context-free, but then it is not really connectionist either," The Ninth Annual Coriference oj the Cognitive Science Society pp. 70-77 (1987). Fanty, M., "Learning in Structured Connectionist Networks." TR 252. University of Rochester Computer Science Department (1988). Selman. Bart, "Rule-based processing in a connectionist system for natural language understanding." Technical Report CSRI-168. Computer Systems Research Institute. University of Toronto (1985). Waltz. D. and Pollack. J., "MaSsively parallel parsing: a strongly interactive model of natural language interpretation," Cognitive Science 9 pp. 51-74 (1985). Feldman. J.A. and Ballard, D.H., "Connectionist models and their properties," Cognitive Science 6 pp. 205-254 (1982). Lippmann. R, "An introduction to computing with neural nets," IEEE ASSP Magazine pp. 4-22 (April 1987).
165 |@word manageable:1 tried:1 tr:1 charniak:2 esj:1 rightmost:1 blank:2 current:4 adj:2 si:1 parsing:13 drop:1 bart:1 leaf:3 accordingly:1 node:12 toronto:1 unbounded:2 height:1 constructed:1 incorrect:4 consists:1 combine:1 introduce:1 frequently:2 terminal:1 totally:2 project:1 easiest:1 santos:8 what:1 differing:1 interactive:1 unit:24 before:1 modify:1 limit:1 punish:1 path:1 approximately:1 plus:2 binder:3 limited:1 nol:1 range:2 lost:1 cfj:1 word:8 seeing:1 cannot:1 context:8 influence:1 disconnecting:1 restriction:1 imposed:1 immediately:2 rule:52 embedding:1 handle:2 stability:1 play:1 parser:7 magazine:1 rumelhart:2 utilized:1 continues:2 labeled:6 bottom:2 capture:1 adv:2 mentioned:1 agency:1 moderately:1 reward:1 trained:2 completely:1 easily:2 represented:4 train:1 distinct:1 quite:1 grammar:4 ability:1 final:1 obviously:1 sequence:2 net:1 fanty:2 j2:1 combining:1 fired:3 generating:1 depending:1 stating:1 clearer:1 nonterminal:1 ij:2 c:1 indicate:1 correct:1 modifying:1 exploration:1 require:1 really:1 aquisition:1 strictly:1 insert:1 distributively:1 normal:1 cognition:1 label:4 currently:2 infonnation:1 largest:3 correctness:1 mit:1 modified:1 office:1 naval:1 consistently:1 indicates:1 mainly:1 eliminate:1 trainer:2 noun:13 special:2 constrained:1 iiil:1 construct:1 identical:1 represents:1 placing:1 future:1 connectionist:6 np:30 report:1 simultaneously:2 national:1 tightly:1 consisting:2 fire:1 attempt:4 inf2:2 waltz:2 edge:2 capable:3 partial:3 closer:1 modest:1 tree:24 old:1 desired:1 pollack:2 arpa:1 column:11 unmodified:1 phrase:1 applicability:1 entry:26 subset:2 successful:1 nonterminals:1 pal:14 providence:1 teacher:3 corrupted:1 chunk:1 contract:2 possibly:1 cognitive:3 creating:2 diversity:1 stabilize:1 coefficient:1 root:1 view:1 break:2 portion:5 parallel:6 capability:1 rochester:1 degraded:2 correspond:1 yes:1 vp:20 basically:4 frequency:1 pp:9 associated:2 popular:1 noooi4:1 knowledge:3 actually:2 back:1 higher:1 x6:1 improved:1 april:1 box:1 though:1 strongly:1 eventhough:1 until:1 hand:1 parse:33 microstructures:1 oil:1 effect:1 brown:2 contain:2 managed:1 hence:1 adjacent:1 during:3 self:5 generalized:1 complete:9 l1:1 volume:1 interpretation:1 feldman:2 tuning:5 nontenninal:3 language:31 had:1 entail:1 something:1 recent:1 moderate:1 massively:6 certain:1 continue:1 employed:1 determine:3 preservation:1 technical:1 involving:1 essentially:2 represent:3 addition:1 unrealistically:1 fifty:1 inconsistent:3 call:1 ideal:2 exceed:1 architecture:2 det:1 shift:1 whether:2 defense:1 speech:1 linguist:1 processed:1 supplied:1 shifted:1 per:3 shall:1 express:1 ist:2 f8:1 utilize:2 place:1 annual:1 ahead:1 ri:3 anyone:1 attempting:1 department:2 structured:1 combination:1 jr:1 slightly:2 restricted:1 taken:1 previously:1 discus:1 serf:1 apply:3 denotes:3 remaining:1 top:1 build:2 society:1 added:1 traditional:1 kth:1 unable:1 clf:1 length:6 minimizing:1 negative:1 suppress:1 implementation:3 snapshot:6 sm:1 incorrectly:1 immediate:1 looking:1 assp:1 pdp:2 ninth:1 verb:10 sentence:15 learned:2 able:3 suggested:1 below:1 built:1 oj:2 shifting:1 natural:3 advanced:1 representing:1 scheme:6 improve:1 attachment:1 eugene:1 understanding:1 determining:2 embedded:1 fully:1 par:2 foundation:1 integrate:2 consistent:3 corrupt:1 row:3 supported:1 last:1 free:8 english:2 institute:1 taking:1 distributed:1 selman:2 collection:3 lippmann:2 dealing:1 mter:1 assumed:1 ballard:2 learn:9 improving:1 complex:1 prep:1 allowed:1 differed:1 fashion:2 clamped:1 learns:1 down:1 symbol:1 exists:1 adding:2 ci:1 te:1 lt:1 simply:1 positional:1 prevents:1 binding:2 goal:1 presentation:1 change:1 ithis:1 called:3 attempted:1 aux:6
711
1,650
A MCMC approach to Hierarchical Mixture Modelling Christopher K. I. Williams Institute for Adaptive and Neural Computation Division of Informatics, University of Edinburgh 5 Forrest Hill, Edinburgh EHI 2QL, Scotland, UK ckiw@dai.ed.ac.uk http://anc.ed.ac.uk Abstract There are many hierarchical clustering algorithms available, but these lack a firm statistical basis. Here we set up a hierarchical probabilistic mixture model, where data is generated in a hierarchical tree-structured manner. Markov chain Monte Carlo (MCMC) methods are demonstrated which can be used to sample from the posterior distribution over trees containing variable numbers of hidden units. 1 Introduction Over the past decade or two mixture models have become a popular approach to clustering or competitive learning problems. They have the advantage of having a well-defined objective function and fit in with the general trend of viewing neural network problems in a statistical framework. However, one disadvantage is that they produce a "flat" cluster structure rather than the hierarchical tree structure that is returned by some clustering algorithms such as the agglomerative single-link method (see e.g. [12]). In this paper I formulate a hierarchical mixture model, which retains the advantages of the statistical framework, but also features a tree-structured hierarchy. The basic idea is illustrated in Figure 1(a). At the root of the tree (level l) we have a single centre (marked with a x). This is the mean of a Gaussian with large variance (represented by the large circle). A random number of centres (in this case 3) are sampled from the level 1 Gaussian, to produce 3 new centres (marked with o's). The variance associated with the level 2 Gaussians is smaller. A number of level 3 units are produced and associated with the level 2 Gaussians. The centre of each level 3 unit (marked with a +) is sampled from its parent Gaussian. This hierarchical process could be continued indefinitely, but in this example we generate data from the level 3 Gaussians, as shown by the dots in Figure lea). A three-level version of this model would be a standard mixture model with a Gaussian prior on where the centres are located. In the four-level model the third level centres are clumped together around the second level means, and it is this that distinguishes the model from a flat mixture model. Another view of the generative process is given in Figure l(b), where the tree structure denotes which nodes are children of particular parents. Note also that this is a directed acyclic graph, with the arrows denoting dependence of the position of the child on that of the parent. A MCMC Approach to Hierarchical Mixture Modelling 681 In section 2 we describe the theory of probabilistic hierarchical clustering and give a discussion of related work. Experimental results are described in section 3. (a) (b) Figure 1: The basic idea of the hierarchical mixture model. (a) x denotes the root of the tree, the second level centres are denoted by o's and the third level centres by +'s. Data is generated from the third level centres by sampling random points from Gaussians whose means are the third level centres. (b) The corresponding tree structure. 2 Theory We describe in turn (i) the prior over trees, (ii) the calculation of the likelihood given a data vector, (iii) Markov chain Monte Carlo (MCMC) methods for the inference of the tree structure given data and (iv) related work. 2.1 Prior over trees We describe first the prior over the number of units in each layer, and then the prior on connections between layers. Consider a L layer hierarchical model. The root node is in levell, there are n2 nodes in level 2, and so on down to nL nodes on level L. These n's are collected together in the vector n. We use a Markovian model for P(n), so that P(n) = P(ndP(n2In1) ... P(nLlnL-1) with P(n1) = 8(n1,I). Currently these are taken to be Poisson distributions offset by 1, so that P(ni+llni) rv PO(Aini) + 1, where Ai is a parameter associated with level i. The offset is used so that there must always be at least one unit in any layer. Given n, we next consider how the tree is formed. The tree structure describes which node in the ith layer is the parent of each node in the (i + 1)th layer, for i = 1, ... , L - 1. Each unit has an indicator vector which stores the index of the parent to which it is attached. We collect all these indicator vectors together into a matrix, denoted Z(n). The probability of a node in layer (i + 1) connecting to any node in layer i is taken to be I/ni. Thus L-1 P(n, Z(n)) = P(n)P(Z(n)ln) = P(n) IT (I/ni)n + i 1? i=l We now describe the generation of a random tree given nand Z(n). For simplicity we describe the generation of points in I-d below, although everything can be extended to arbitrary dimension very easily. The mean f-Ll of the level 1 Gaussian is at the origin 1. The I It is easy to relax this assumption so that /-L 1 has a prior Gaussian distribution, or is located at some point other than the origin. 682 C. K. l. Williams level 2 means It], j = 1, ... ,n2 are generated from N (It 1 , af), where ar is the variance associated with the level ] node. Similarly, the position of each level 3 node is generated from its level 2 parent as a displacement from the position of the level 2 parent. This This process continues displacement is a Gaussian RV with zero mean and variance on down to the visible variables. In order for this model to be useful, we require that ar > > ... > i.e. that the variability introduced at successive levels declines monotonically (cj scaling of wavelet coefficients). ai. ai 2.2 aI-I' Calculation of the likelihood The data that we observe are the positions of the points in the final layer; this is denoted x. To calculate the likelihood of x under this model, we need to integrate out the locations of the means of the hidden variables in levels 2 through to L - 1. This can be done explicitly, however, we can shorten this calculation by realizing that given Z(n), the generative distribution for the observables x is Gaussian N(O, C). The covariance matrix C can be calculated as follows. Consider two leaf nodes indexed by k and i. The Gaussian RVs that generated the position of these two leaves can be denoted l _ 1 Xk - Wk + 2 Wk + ... + (L-l) Wk , XI = WI1 + WI2 + ... + WI(L-l) . To calculate the covariance between Xk and Xl, we simply calculate (XkXI). This depends crucially on how many of the w's are shared between nodes k and l (cj path analysis). For example, if Wk i- wl, i.e. the nodes lie in different branches of the tree at levell, their covariance is zero. If k = l, the variance is just the sum of the variances of each RV in the tree. In between, the covariance of Xk and XI can be determined by finding at what level in the tree their common parent occurs. Under these assumptions, the log likelihood L of x given Z (n) is 1 nL L=-"21 x T C -1 x-"2logICI-Tlog21r. (1) In fact this calculation can be speeded up by taking account of the tree structure (see e.g. [8]). Note also that the posterior means (and variances) of the hidden variables can be calculated based on the covariances between the hidden and visible nodes. Again, this calculation can be carried out more efficiently; see Pearl [11] (section 7.2) for details. 2.3 Inference for nand Z (n) Given n we have the problem of trying to infer the connectivity structure Z given the observations x. Of course what we are interested in is the posterior distribution over Z, i.e. P(Zlx, n). One approach is to use a Markov chain Monte Carlo (MCMC) method to sample from this posterior distribution. A straightforward way to do this is to use the Metropolis algorithm, where we propose changes in the structure by changing the parent of a single node at a time. Note the similarities of this algorithm to the work of Williams and Adams [14] on Dynamic Trees COTs); the main differences are Ci) that disconnections are not allowed, i.e. we maintain a single tree (rather than a forest), and (ii) that the variables in the DT image models are discrete rather than Gaussian. We also need to consider moves that change n. This can be effected with a split/merge move. In the split direction, consider a node with a parent and several children. Split this node and randomly assign the children to the two split nodes. Each of the split nodes keeps the same parent. The probability of accepting this move under the Metropolis-Hastings scheme is a . (1 = mm P(n', Z(n')lx)Q(n', Z(n'); n, Z(n))) 'P(n, Z(n)lx)Q(n, Z(n); n', Z(n')) , A MCMC Approach to Hierarchical Mixture Modelling 683 where Q(n', Z(n'); n, Z(n)) is the proposal probability of configuration (n', Z(n')) given configuration (n, Z (n)). This scheme is based on the work on MCMC model composition (MC 3 ) by Madigan and York [9], and on Green 's work on reversible jump MCMC [5]. Another move that changes n is to remove "dangling" nodes, i.e. nodes which have no children. This occurs when all the nodes in a given layer "decide" not to use one or more nodes in the layer above. An alternative to sampling from the posterior is to use approximate inference, such as mean-field methods. These are currently being investigated for DT models [1]. 2.4 Related work There are a very large number of papers on hierarchical clustering; in this work we have focussed on expressing hierarchical clustering in terms of probabilistic models. For example Ambros-Ingerson et at [2] and Mozer [10] developed models where the idea is to cluster data at a coarse level, subtract out mean and cluster the residuals (recursively). This paper can be seen as a probabilistic interpretation of this idea. The reconstruction of phylogenetic trees from biological sequence (DNA or protein) information gives rise to the problem of inferring a binary tree from the data. Durbin et al [3] (chapter 8) show how a probabilistic formulation of the problem can be developed, and the link to agglomerative hierarchical clustering algorithms as approximations to the full probabilistic method (see ?8.6 in [3]). Much of the biological sequence work uses discrete variables, which diverges somewhat from the focus of the current work. However work by Edwards (1970) [4] concerns a branching Brownian-motion process, which has some similarities to the model described above. Important differences are that Edwards' model is in continuous time, and the the variances of the particles are derived from a Wiener process (and so have variance proportional to the lifetime of the particle). This is in contrast to the decreasing sequence of variances at a given number of levels assumed in the above model. One important difference between the model discussed in this paper and the phylogenetic tree model is that points in higher levels of the phylogenetic tree are taken to be individuals at an earlier time in evolutionary history, which is not the interpretation we require here. An very different notion of hierarchy in mixture models can be found in the work on the AutoClass system [6]. They describe a model involving class hierarchy and inheritance, but their trees specify over which dimensions sharing of parameters occurs (e.g. means and covariance matrices for Gaussians). In contrast, the model in this paper creates a hierarchy over examples labelled 1, ... ,n rather than dimensions. Xu and Pearl [15] discuss the inference of a tree-structured belief network based on knowledge of the covariances of the leaf nodes. This algorithm cannot be applied directly in our case as the covariances are not known, although we note that if multiple runs from a given tree structure were available the covariances might be approximated using sample estimates. Other ideas concerning hierarchical clustering are discussed in [13] and [7]. 3 Experiments We describe two sets of experiments to explore these ideas. 3.1 Searching over Z with n fixed 100 4-level random trees were generated from the prior, using values of >'1 = 1.5, >'2 = 2, 3, and 10, (Ji 1, (J~ 0.01. These trees had between 4 and 79 leaf >'3 = (JI = = = C. K I. Williams 684 nodes, with an average of 30. For each tree n was kept the same as in the generative tree, and sampling was carried out over Z starting from a random initial configuration. A given node proposes changing its parent, and this proposal is accepted or rejected with the usual Metropolis probability. In one sweep, each node in levels 3 and 4 makes such a move. (Level 2 nodes only have one parent so there is no point in such a move there.) To obtain a representative sample of P(Z(n)ln, x), we should run the chain for as long as possible. However, we can also use the chain to find configurations with high posterior probability, and in this case running for longer only increases the chances of finding a better configuration. In our experiments the sampler was run for 100 sweeps. As P(Z(n)ln) is uniform for fixed n, the posterior is simply proportional to the likelihood term. It would also be possible to run simulated annealing with the same move set to search explicitly for the maximum a posteriori (MAP) configuration. The results are that for 76 of the 100 cases the tree with the highest posterior probability (HPP) configuration had higher posterior probability than the generative tree, for 20 cases the same tree was found and in 4 cases the HPP solution was inferior to the generative tree. The fact that in almost all cases the sampler found a configuration as good or better than the generative one in a relatively small number of sweeps is very encouraging. In Figure 2 the generative (left column) and HPP trees for fixed n (middle column) are plotted for two examples. In panel (b) note the "dangling" node in level 2, which means that the level 3 nodes to the left end up in a inferior configuration to (a). By contrast, in panel (e) the sampler has found a better (less tangled) configuration than the generative model (d). _'1150841 (a) (b) (c) (d) (e) (f) Figure 2: (a) and (d) show the generative trees for two examples. The corresponding HPP trees for fixed n are plotted in (b) and (e) and those for variable n in (c) and (f). The number in each panel is the log posterior probability of the configuration. The nodes in levels 2 and 3 are shown located at their posterior means. Apparent non-tree structures are caused by two nodes being plotted almost on top of each other. 3.2 Searching over both nand Z Given some data x we will not usually know appropriate numbers of hidden units. This motivates searching over both Z and n, which can be achieved using the split/merge moves discussed in section 2.3. In the experiments below the initial numbers of units in levels 2 and 3 (denoted n2 and A MCMC Approach to Hierarchical Mixture Modelling 685 113) were set using the simple-minded formulae 113 = rdim(x)/A31112 = r113/A21. A proper inferential calculation for 71,2 and 71,3 can be carried out, but it requires the solution of a non-linear optimization problem. Given 112 and 113, the initial connection configuration was chosen randomly. The search method used was to propose a split/merge move (with probability 0.5:0.5) in level 2, then to sample the level 2 to level 3 connections, and then to propose a split-merge move in level 3, and then update the level 3 to level 4 connections. This comprised a single sweep, and as above 100 sweeps were used. Experiments were conducted on the same trees used in section 3.1. In this case the results were that for 50 out of the 100 cases, the HPP configuration had higher posterior probability than the generative tree, for 11 cases the same tree was found and in 39 cases the HPP solution was inferior to the generative tree. Overall these results are less good than the ones in section 3.1, but it should be remembered that the search space is now much larger, and so it would be expected that one would need to search longer. Comparing the results from fixed n against those with variable n shows that in 42 out of 100 cases the variable n method gave a higher posterior probability. in 45 cases it was lower and in 13 cases the same trees were found. The rightmost column of Figure 2 shows the HPP configurations when sampling with variable n on the two examples discussed above. In panel (c) the solution found is not very dissimilar to that in panel (b), although the overall probability is lower. In Cf), the solution found uses just one level 2 centre rather than two, and obtains a higher posterior probability than the configurations in (e) and Cd). 4 Discussion The results above indicate that the proposed model behaves sensibly, and that reasonable solutions can be found with relatively short amounts of search. The method has been demonstrated on univariate data, but extending it to multivariate Gaussian data for which each dimension is independent given the tree structure is very easy as the likelihood calculation is independent on each dimension. There are many other directions is which the model can be developed. Firstly, the model as presented has uniform mixing proportions, so that children are equally likely to connect to each potential parent. This can be generalized so that there is a non-uniform vector of connection probabilities in each layer. Also, given a tree structure and independent Dirichlet priors over these probability vectors, these parameters can be integrated out analytically. Secondly, the model can be made to generate iid data by regarding the penultimate layer as mixture centres; in this case the term P(nLI71,L-l) would be ignored when computing the probability of the tree. Thirdly, it would be possible to add the variance variables to the MCMC scheme, e.g. using the Metropolis algorithm, after defining a suitable prior on the The constraint that all variances in the same level are sequence of variances equal could also be relaxed by allowing them to depend on hyperparameters set at every level. Fourthly, there may be improved MCMC schemes that can be devised. For example, in the current implementation the posterior means of the candidate units are not taken into account when proposing merge moves Ccj [5]). Fifthly, for the multivariate Gaussian version we can consider a tree-structured factor analysis model, so that higher levels in the tree need not have the same dimensionality as the data vectors. ai, ... ,ai-I' One can also consider a version where each dimension is a multinomial rather than a continuous variable. In this case one might consider a model where a multinomial parameter vector (}l in the tree is generated from its parent by (}l = 'Y(}l-l + (1- 'Y)r where 'Y E [0,1] and r is a random vector of probabilities. An alternative model could be to build a tree 686 C. K. 1. Williams structured prior on the a parameters of the Dirichlet prior for the multinomial distribution. Acknowledgments This work is partially supported through EPSRC grant GRIL 78161 Probabilistic Models for Sequences. I thank the Gatsby Computational Neuroscience Unit (UCL) for organizing the "Mixtures Day" in March 1999 and supporting my attendance, and Peter Green, Phil Dawid and Peter Dayan for helpful discussions at the meeting. I also thank Amos Storkey for helpful discussions and Magnus Rattray for (accidentally!) pointing me towards the chapters on phylogenetic trees in [3]. References [1] N. J. Adams, A. Storkey, Z. Ghahramani, and C. K. 1. Williams. MFDTs: Mean Field Dynamic Trees. Submitted to ICPR 2000,1999. [2] J. Ambros-Ingerson, R. Granger, and G. Lynch. Simulation of Paleocortex Performs Hierarchical Clustering. Science , 247:1344-1348,1990. [3] R. Durbin, S. Eddy, A. Krogh, and G. Mitchison. Biological Sequence Analysis. Cambridge University Press, Cambridge, UK, 1998. [4] A. w. F. Edwards. Estimation of the Branch Points of a Branching Diffusion Process. Journal of the Royal Statistical Society B, 32(2): 155-174, 1970. [5] P. J. Green. Reversible Jump Markov chain Monte Carlo computation and Bayesian model determination. Biometrika, 82(4):711-732, 1995. [6] R. Hanson, J. Stutz, and P. Cheeseman. Bayesian Classification with Correlation and Inheritance. In IlCAI-91: Proceedings of the Twelfth International Joint Conference on Artificial Intelligence, 1991 . Sydney, Australia. [7] T. Hofmann and J. M. Buhmann. Hierarchical Pairwise Data Clustering by MeanField Annealing. In F. Fogelman-Soulie and P. Gallinari, editors, Proc. ICANN 95. EC2 et Cie, 1995. [8] M. R. Luettgen and A. S. Wi11sky. Likelihood Calculation for a Class of Multiscale Stochastic Models, with Application to Texture Discrimination. IEEE Trans. Image Processing, 4(2): 194-207, 1995. [9] D. Madigan and J. York. Bayesian Graphical Models for Discrete Data. International Statistical Review, 63:215-232,1995. [10] M. C. Mozer. Discovering Discrete Distributed Representations with Iterated Competitive Learning. In R. P. Lippmann, J. E. Moody, and D. S. Touretzky, editors, Advances in Neural Information Processing Systems 3. Morgan Kaufmann, 1991. [11] J. Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, San Mateo, CA, 1988. [12] B. Ripley. Pattern Recognition and Neural Networks. Cambridge University Press, Cambridge, UK, 1996. [13] N. Vasconcelos and A. Lippmann. Learning Mixture Hierarchies. In M. S. Kearns, S. A. Solla, and D. A. Cohn, editors, Advances in Neural Information Processing Systems 11, pages 606-612. MIT Press, 1999. [14] C. K. 1. Williams and N. J. Adams. DTs: Dynamic Trees. In M. J. Kearns, S. A. Solla, and D. A. Cohn, editors, Advances in Neural Information Processing Systems 11 . MIT Press, 1999. [15] L. Xu and J. Pearl. Structuring Causal Tree Models with Continuous Variables. In L. N. Kanal, T. S. Levitt, and J. F. Lemmer, editors, Uncertainty in Artificial Intelligence 3. Elsevier, 1989.
1650 |@word middle:1 version:3 proportion:1 twelfth:1 simulation:1 crucially:1 covariance:9 recursively:1 initial:3 configuration:15 denoting:1 rightmost:1 past:1 current:2 comparing:1 hpp:7 must:1 visible:2 hofmann:1 remove:1 update:1 discrimination:1 generative:11 leaf:4 intelligence:2 discovering:1 ccj:1 xk:3 scotland:1 ith:1 realizing:1 short:1 indefinitely:1 accepting:1 coarse:1 node:32 location:1 successive:1 lx:2 firstly:1 phylogenetic:4 become:1 manner:1 pairwise:1 expected:1 cot:1 paleocortex:1 decreasing:1 encouraging:1 panel:5 what:2 developed:3 proposing:1 finding:2 every:1 sensibly:1 biometrika:1 uk:5 gallinari:1 unit:10 grant:1 path:1 merge:5 might:2 mateo:1 collect:1 speeded:1 directed:1 acknowledgment:1 displacement:2 inferential:1 madigan:2 protein:1 cannot:1 tangled:1 map:1 demonstrated:2 phil:1 williams:7 straightforward:1 starting:1 formulate:1 simplicity:1 shorten:1 continued:1 searching:3 notion:1 hierarchy:5 us:2 origin:2 trend:1 dawid:1 approximated:1 storkey:2 located:3 continues:1 recognition:1 epsrc:1 calculate:3 solla:2 highest:1 mozer:2 dynamic:3 depend:1 creates:1 division:1 basis:1 observables:1 po:1 easily:1 joint:1 represented:1 chapter:2 describe:7 monte:4 artificial:2 firm:1 whose:1 apparent:1 larger:1 plausible:1 relax:1 final:1 advantage:2 sequence:6 ucl:1 propose:3 reconstruction:1 rdim:1 organizing:1 mixing:1 parent:15 cluster:3 diverges:1 extending:1 produce:2 adam:3 ac:2 krogh:1 sydney:1 edward:3 indicate:1 direction:2 stochastic:1 australia:1 viewing:1 everything:1 require:2 assign:1 biological:3 secondly:1 fourthly:1 mm:1 around:1 magnus:1 pointing:1 wi1:1 estimation:1 proc:1 currently:2 wl:1 minded:1 amos:1 mit:2 gaussian:12 always:1 lynch:1 rather:6 structuring:1 derived:1 focus:1 modelling:4 likelihood:7 contrast:3 posteriori:1 inference:5 helpful:2 dayan:1 elsevier:1 integrated:1 nand:3 hidden:5 interested:1 fogelman:1 overall:2 classification:1 denoted:5 proposes:1 field:2 equal:1 having:1 vasconcelos:1 sampling:4 attendance:1 intelligent:1 distinguishes:1 randomly:2 individual:1 n1:2 maintain:1 mixture:14 nl:2 chain:6 stutz:1 tree:55 iv:1 indexed:1 circle:1 plotted:3 causal:1 column:3 earlier:1 markovian:1 ar:2 disadvantage:1 retains:1 uniform:3 comprised:1 conducted:1 connect:1 my:1 ehi:1 international:2 ec2:1 probabilistic:8 informatics:1 together:3 connecting:1 moody:1 connectivity:1 again:1 autoclass:1 luettgen:1 containing:1 ambros:2 account:2 potential:1 wk:4 coefficient:1 explicitly:2 caused:1 depends:1 root:3 view:1 competitive:2 effected:1 formed:1 ni:3 wiener:1 variance:13 kaufmann:2 efficiently:1 bayesian:3 iterated:1 produced:1 iid:1 mc:1 carlo:4 history:1 submitted:1 touretzky:1 sharing:1 ed:2 against:1 associated:4 sampled:2 popular:1 knowledge:1 dimensionality:1 cj:2 eddy:1 higher:6 dt:2 day:1 specify:1 improved:1 formulation:1 done:1 lifetime:1 just:2 rejected:1 correlation:1 hastings:1 christopher:1 cohn:2 multiscale:1 reversible:2 lack:1 analytically:1 illustrated:1 ll:1 branching:2 inferior:3 generalized:1 trying:1 hill:1 performs:1 motion:1 reasoning:1 image:2 common:1 behaves:1 multinomial:3 ji:2 attached:1 thirdly:1 discussed:4 interpretation:2 expressing:1 composition:1 cambridge:4 ai:6 similarly:1 particle:2 clumped:1 centre:12 had:3 dot:1 similarity:2 longer:2 add:1 posterior:15 brownian:1 multivariate:2 store:1 binary:1 remembered:1 meeting:1 seen:1 morgan:2 dai:1 somewhat:1 relaxed:1 monotonically:1 ii:2 rv:4 branch:2 full:1 multiple:1 infer:1 levell:2 determination:1 calculation:8 af:1 long:1 devised:1 concerning:1 equally:1 involving:1 basic:2 poisson:1 cie:1 achieved:1 lea:1 proposal:2 annealing:2 iii:1 easy:2 split:8 fit:1 gave:1 idea:6 decline:1 regarding:1 lemmer:1 peter:2 returned:1 york:2 ignored:1 useful:1 amount:1 dna:1 http:1 generate:2 ilcai:1 dangling:2 neuroscience:1 rattray:1 discrete:4 four:1 changing:2 diffusion:1 kept:1 graph:1 sum:1 run:4 uncertainty:1 almost:2 reasonable:1 decide:1 forrest:1 scaling:1 layer:13 durbin:2 constraint:1 flat:2 relatively:2 structured:5 icpr:1 march:1 smaller:1 describes:1 ingerson:2 wi:1 metropolis:4 taken:4 ln:3 turn:1 discus:1 granger:1 know:1 end:1 available:2 gaussians:5 ckiw:1 hierarchical:19 observe:1 appropriate:1 alternative:2 denotes:2 clustering:10 running:1 top:1 cf:1 dirichlet:2 graphical:1 ghahramani:1 build:1 society:1 sweep:5 objective:1 move:11 occurs:3 dependence:1 usual:1 evolutionary:1 link:2 thank:2 simulated:1 penultimate:1 me:1 agglomerative:2 collected:1 index:1 ql:1 rise:1 implementation:1 motivates:1 proper:1 allowing:1 observation:1 markov:4 supporting:1 defining:1 extended:1 variability:1 arbitrary:1 introduced:1 connection:5 hanson:1 dts:1 zlx:1 pearl:4 trans:1 below:2 usually:1 pattern:1 wi2:1 green:3 royal:1 belief:1 suitable:1 meanfield:1 indicator:2 cheeseman:1 residual:1 buhmann:1 scheme:4 carried:3 prior:11 review:1 inheritance:2 generation:2 proportional:2 acyclic:1 integrate:1 editor:5 cd:1 course:1 supported:1 accidentally:1 disconnection:1 institute:1 taking:1 focussed:1 edinburgh:2 distributed:1 soulie:1 dimension:6 calculated:2 made:1 adaptive:1 jump:2 san:1 approximate:1 obtains:1 lippmann:2 keep:1 assumed:1 xi:2 ndp:1 mitchison:1 ripley:1 continuous:3 search:5 decade:1 ca:1 kanal:1 forest:1 anc:1 investigated:1 icann:1 main:1 arrow:1 hyperparameters:1 n2:3 child:6 allowed:1 xu:2 levitt:1 representative:1 gatsby:1 position:5 inferring:1 a21:1 xl:1 lie:1 candidate:1 third:4 wavelet:1 down:2 formula:1 offset:2 concern:1 ci:1 texture:1 subtract:1 simply:2 explore:1 univariate:1 likely:1 partially:1 chance:1 marked:3 towards:1 labelled:1 shared:1 change:3 determined:1 sampler:3 kearns:2 accepted:1 experimental:1 dissimilar:1 mcmc:11
712
1,651
Agglomerative Information Bottleneck Noam Slonim Naftali Tishby* Institute of Computer Science and Center for Neural Computation The Hebrew University Jerusalem, 91904 Israel email: {noamm.tishby}(Qcs.huji.ac.il Abstract We introduce a novel distributional clustering algorithm that maximizes the mutual information per cluster between data and given categories. This algorithm can be considered as a bottom up hard version of the recently introduced "Information Bottleneck Method". The algorithm is compared with the top-down soft version of the information bottleneck method and a relationship between the hard and soft results is established. We demonstrate the algorithm on the 20 Newsgroups data set. For a subset of two newsgroups we achieve compression by 3 orders of magnitudes loosing only 10% of the original mutual information. 1 Introduction The problem of self-organization of the members of a set X based on the similarity of the conditional distributions of the members of another set, Y, {p(Ylx)}, was first introduced in [8] and was termed "distributional clustering" . This question was recently shown in [9] to be a special case of a much more fundamental problem: What are the features of the variable X that are relevant for the prediction of another, relevance, variable Y? This general problem was shown to have a natural in~ormation theoretic formulation: Find a compressed representation of the variable X, denoted X, such that the mutual information between X and Y, I (X; Y), is as high as possible, under a constraint on the mutual information between X and X, I (X; X). Surprisingly, this variational problem yields an exact self-consistent equations for the conditional distributions p(ylx), p(xlx), and p(x). This constrained information optimization problem was called in [9] The Information Bottleneck Method. The original approach to the solution of the resulting equations, used already in [8], was based on an analogy with the "deterministic annealing" approach to clustering (see [7]). This is a top-down hierarchical algorithm that starts from a single cluster and undergoes a cascade of cluster splits which are determined stochastically (as phase transitions) into a "soft" (fuzzy) tree of clusters. In this paper we propose an alternative approach to the information bottleneck N Slonim and N Tishby 618 problem, based on a greedy bottom-up merging. It has several advantages over the top-down method. It is fully deterministic, yielding (initially) "hard clusters", for any desired number of clusters. It gives higher mutual information per-cluster than the deterministic annealing algorithm and it can be considered as the hard (zero temperature) limit of deterministic annealing, for any prescribed number of clusters. Furthermore, using the bottleneck self-consistent equations one can "soften" the resulting hard clusters and recover the deterministic annealing solutions without the need to identify the cluster splits, which is rather tricky. The main disadvantage of this method is computational, since it starts from the limit of a cluster per each member of the set X. 1.1 The information bottleneck method The mutual information between the random variables X and Y is the symmetric functional of their joint distribution, I(X;Y) = L p(x,y) log ( ~~~'~\) L = p(x)p(ylx) log (P(Y(lx))) . PY (1) The objective of the information bottleneck method is to extract a compact representation of the variable X, denoted here by X, with minimal loss of mutual information to another, relevance, variable Y. More specifically, we want to find a (possibly stochastic) map, p(xlx), that minimizes the (lossy) coding length of X via X, I(Xi X), under a constraint on the mutual information to the relevance variable I(Xi Y). In other words, we want to find an efficient representation of the variable X, X, such that the predictions of Y from X through X will be as close as possible to the direct prediction of Y from X. xEX,yEY P PY xEX,yEY As shown in [9], by introducing a positive Lagrange multiplier 13 to enforce the mutual information constraint, the problem amounts to minimization of the Lagrangian: ?[P(xlx)] = I(Xi X) - f3I(Xi Y) , (2) with respect to p(xlx), subject to the Markov condition X -t X -t Y and normalization. This minimization yields directly the following self-consistent equations for the map p(xlx), as well as for p(Ylx) and p(x): p(xlx) = i(~~;) exp (-f3D KL [P(Ylx)lIp(Ylx)]) { p(Ylx) = 2:xp(Ylx)p(xlx)~ p(x) = 2: x p(xlx)p(x) (3) where Z(f3, x) is a normalization function. The functional DKL[Pllq] == 2: y p(y) log ~f~~ is the Kulback-Liebler divergence [3J , which emerges here from the variational principle. These equations can be solved by iterations that are proved to converge for any finite value of 13 (see [9]). The Lagrange multiplier 13 has the natural interpretation of inverse temperature, which suggests deterministic annealing [7] to explore the hierarchy of solutions in X, an approach taken already in [8J . The variational principle, Eq.(2), determines also the shape of the annealing process, since by changing 13 the mutual informations Ix == I(X; X) and Iy == I(Y; X) vary such that My = 13- 1 (4) Mx . Agglomerative Information Bottleneck 619 Thus the optimal curve, which is analogous to the rate distortion function in information theory [3], follows a strictly concave curve in the (Ix,Iy) plane, called the information plane. Deterministic annealing, at fixed number of clusters, follows such a concave curve as well, but this curve is suboptimal beyond a certain critical value of f3. Another interpretation of the bottleneck principle comes from the relation between the mutual information and Bayes classification error. This error is bounded above and below (see [6]) by an important information theoretic measure of the class conditional distributions p(XIYi), called the Jensen-Shannon divergence. This measure plays an important role in our context. The Jensen-Shannon divergence of M class distributions, Pi(X), each with a prior 1 ~ i ~ M, is defined as, [6,4]. 7ri, M M JS7r (Pl,P2, ???,PM] == H[L 7riPi(X)] - L 7ri H [Pi(X)] , i=l (5) i=l where H[P(x)] is Shannon's entropy, H[P(x)] = - Ex p(x) logp(x). The convexity of the entropy and Jensen inequality guarantees the non-negativity of the JSdivergence. 1.2 The hard clustering limit For any finite cardinality of the representation IXI == m the limit f3 -+ 00 of the Eqs.(3) induces a hard partition of X into m disjoint subsets. In this limit each member x E X belongs only to the subset x E X for which p(Ylx) has the smallest DKL[P(ylx)lIp(ylx)] and the probabilistic map p(xlx) obtains the limit values 0 and 1 only. In this paper we focus on a bottom up agglomerative algorithm for generating "good" hard partitions of X. We denote an m-partition of X, i.e. X with cardinality m, also by Zm = {Zl,Z2, ... ,Zm}, in which casep(x) =p(Zi). We say that Zm is an optimal m-partition (not necessarily unique) of X if for every other m-partition of X, Z:n, I(Zm; Y) ~ I(Z:n; Y). Starting from the trivial N-partition, with N = lXI, we seek a sequence of merges into coarser and coarser partitions that are as close as possible to optimal. It is easy to verify that in the f3 -+ 00 limit Eqs.(3) for the m-partition distributions are simplified as follows. Let x == Z = {XI,X2, ... ,xl z l} , Xi E X denote a specific component (Le. cluster) of the partition Zm, then { 1 p(zlx) = I 0 if x E Z . Vx E X oth erWlse p(ylz) = plz) El~l P(Xi, y) Vy p(z) = El~l P(Xi) E Y (6) Using these distributions one can easily evaluate the mutual information between Zm and Y, I(Zm; Y), and between Zm and X, I(Zm; X), using Eq.(l). Once any hard partition, or hard clustering, is obtained one can apply "reverse annealing" and "soften" the clusters by decreasing f3 in the self-consistent equations, Eqs.( 3). Using this procedure we in fact recover the stochastic map, p(xlx), from the hard partition without the need to identify the cluster splits. We demonstrate this reverse deterministic annealing procedure in the last section. 620 1.3 N. Slonim and N. Tishby Relation to other work A similar agglomerative procedure, without the information theoretic framework and analysis, was recently used in [1] for text categorization on the 20 newsgroup corpus. Another approach that stems from the distributional clustering algorithm was given in [5] for clustering dyadic data. An earlier application of mutual information for semantic clustering of words was used in [2]. 2 The agglomerative information bottleneck algorithm The algorithm starts with the trivial partition into N = IXI clusters or components, with each component contains exactly one element of X. At each step we merge several components of the current partition into a single new component in a way that locally minimizes the loss of mutual information leX; Y) = l(Zm; Y). Let Zm be the current m-partition of X and Zm denote the new m-partition of X after the merge of several components of Zm. Obviously, m < m. Let {Zl, Z2, ... , zd ~ Zm denote the set of components to be merged, and Zk E Zm the new component that is generated by the merge, so m = m - k + 1. To evaluate the reduction in the mutual information l(Zm; Y) due to this merge one needs the distributions that define the new m-partition, which are determined as follows. For every Z E Zm, Z f:. Zk, its probability distributions (p(z),p(ylz),p{zlx? remains equal to its distributions in Zm. For the new component, Zk E Zm, we define, p(Zk) = L~=l P(Zi) { p(yIZk) = P(~Ic) ~~=l P(Zi, y) \ly E Y . (7) (zlx) = 1f x ~i for some ~ ~ ~ k \Ix X p 0 otherw1se E {1 E 1 It is easy to verify that Zm is indeed a valid m-partition with proper probability distributions. Using the same notations, for every merge we define the additional quantities: ? The merge prior distribution: defined by ilk == (71"1,71"2, ... , 7I"k), where is the prior probability of Zi in the merged subset, i.e. 7I"i == :t::). 7I"i ? The Y -information decrease: the decrease in the mutual information l(X; Y) due to a single merge, cHy(Zl' ""Zk) == l(Zm; Y) - l(Zm; Y) ? The X-information decrease: the decrease in the mutual information l(X, X) due to a single merge, cHx (Z1, Z2, ... , Zk) == l(Zm; X) - l(Zm; X) Our algorithm is a greedy procedure, where in each step we perform "the best possible merge" , i.e. merge the components {Z1, ... , zd of the current m-partition which minimize cHy(Z1, ... , Zk). Since cHy(Zl, ... , Zk) can only increase with k (corollary 2), for a greedy procedure it is enough to check only the possible merging of pairs of components of the current m-partition. Another advantage of merging only pairs is that in this way we go through all the possible cardinalities of Z = X, from N to 1. For a given m-partition Zm = {Z1,Z2, ... ,Zm} there are m(~-1) possible pairs to merge. To find "the best possible merge" one must evaluate the reduction of information cHy(Zi, Zj) = l(Zm; Y) - l(Zm-1; Y) for every pair in Zm , which is O(m . WI) operations for every pair. However, using proposition 1 we know that cHy(Zi, Zj) = (P(Zi) + p(Zj? . JSn 2 (P(YIZi),p(Ylzj?, so the reduction in the mutual 621 Agglomerative Information Bottleneck information due to the merge of Zi and Zj can be evaluated directly (looking only at this pair) in O(IYI) operations, a reduction of a factor of m in time complexity (for every merge). Input: Empirical probability matrix p(x,y), N = Output: IX\, M = IYI Zm : m-partition of X into m clusters, for every 1 ::; m ::; N Initialization: ? Construct Z == X - For i = 1.. .N * Zi = {x;} * P(Zi) = p(Xi) * p(YIZi) = p(YIXi) for every y E Y * p(zlxj) = 1 if j = i and 0 otherwise - Z={Zl, ... ,ZN} ? for every i, j = 1.. .N, i < j, calculate di,j = (p(Zi) +p(Zj))' JSn2[P(ylzi),p(ylzj)] (every di,j points to the corresponding couple in Z) Loop: ? For t = 1... (N - 1) - Find {a,.B} = argmini ,j {di,j } (if there are several minima choose arbitrarily between them) - Merge {z"" Zj3} => z : * p(z) = p(z",) + p(Zj3) * p(ylz) = r>li) (p(z""y) +p(zj3,y)) for every y E Y * p(zlx) = 1 if x E z'" U Zj3 and 0 otherwise, for every x E X - Update Z = {Z - {z"" z,q}} U{z} (Z is now a new (N - t)-partition of X with N - t clusters) - Update di ,j costs and pointers w.r.t. z (only for couples contained z'" or Zj3). ? End For Figure 1: Pseudo-code of the algorithm. 3 Discussion The algorithm is non-parametric, it is a simple greedy procedure, that depends only on the input empirical joint distribution of X and Y. The output of the algorithm is the hierarchy of all m-partitions Zm of X for m = N, (N - 1), ... ,2,1. Moreover, unlike most other clustering heuristics, it has a built in measure of efficiency even for sub-optimal solutions, namely, the mutual information I(Zm; Y) which bounds the Bayes classification error. The quality measure of the obtained Zm partition is the fraction of the mutual information between X and Y that Zm captures. This is given by the curve II Z,;,-: vs. m = 1Zm I. We found that empirically this curve was concave. If this is always true the decrease in the mutual information at every step, given by 8(m) == I(Z7n;Y)(-:~7n-l;Y) can only increase with decreasing m. Therefore, if at some point 8(m) becomes relatively high it is an indication that we have reached a value of m with "meaningful" partition or clusters. Further N. Slonim and N. Tishby 622 merging results in substantial loss of information and thus significant reduction in the performance of the clusters as features. However, since the computational cost of the final (low m) part of the procedure is very low we can just as well complete the merging to a single cluster. 0 .8 1 >" :;:~ ? 06 - :::"0.5 >- >" ~ ~O . 4 0.2 ... NG100 -.6.+ I- o~'----;-;;;;;;------ "3000:=-- -= 121 0 .2 0 .4 0 .6 I(Z;X) / H(X) 0.8 %~--~ '0~--~2~0----~ 30~---4~ 0----~ ~ ~ IZI Figure 2: On the left figure the results of the agglomerative algorithm are shown in the "information plane", normalized I(Z ;Y) vs. normalized I(Z ; X) for the NGlOOO dataset. It is compared to the soft version of the information bottleneck via "reverse annealing" for IZI = 2,5, 10, 20, 100 (the smooth curves on the left). For IZI = 20 , 100 the annealing curve is connected to the starting point by a dotted line. In this plane the hard algorithm is clearly inferior to the soft one. On the right-hand side: I(Zm, Y) of the agglomerative algorithm is plotted vs. the cardinality of the partition m for three subsets of the newsgroup dataset . To compare the performance over the different data cardinalities we normalize I(Zm ;Y) by the value of I(Zso ;Y) , thus forcing all three curves to start (and end) at the same points. The predictive information on the newsgroup for NGlOOO and NGIOO is very similar, while for the dichotomy dataset, 2ng, a much better prediction is possible at the same IZI, as can be expected for dichotomies. The inset presents the full curve of the normalized I(Z ;Y) vs. IZI for NGIOO data for comparison. In this plane the hard partitions are superior to the soft ones. 4 Application To evaluate the ideas and the algorithm we apply it to several subsets of the 20Newsgroups dataset, collected by Ken Lang using 20, 000 articles evenly distributed among 20 UseNet discussion groups (see [1]). We replaced every digit by a single character and by another to mark non-alphanumeric characters. Following this preprocessing, the first dataset contained the 530 strings that appeared more then 1000 times in the data. This dataset is referred as NG1000 . Similarly, all the strings that appeared more then 100 times constitutes the NG100 dataset and it contains 5148 different strings. To evaluate also a dichotomy data we used a corpus consisting of only two discussion groups out of the 20Newsgroups with similar topics: alt. atheism and talk. religion. misc. Using the same pre-processing, and removing strings that occur less then 10 times, the resulting "lexicon" contained 5765 different strings. We refer to this dataset as 2ng. We plot the results of our algorithm on these three data sets in two different planes. First , the normalized information ;g~~~ vs. the size of partition of X (number of clusters) , IZI. The greedy procedure directly tries to maximize J(Z ; Y) for a given IZI, as can be seen by the strong concavity of these curves (figure 2, right). Indeed the procedure is able to maintain a high percentage of the relevant mutual information of the original data, while reducing the dimensionality of the "features" , Agglomerative Information Bottleneck 623 IZI, by several orders of magnitude. On the right hand-side of figure 2 we present a comparison between the efficiency of the procedure for the three datasets. The two-class data, consisting of 5765 different strings, is compressed by two orders of magnitude, into 50 clusters, almost without loosing any of the mutual information about the news groups (the decrease in I(Xi Y) is about 0.1%). Compression by three orders of magnitude, into 6 clusters, maintains about 90% of the original mutual information. Similar results, even though less striking, are obtained when Y contain all 20 newsgroups. The NG100 dataset was compressed from 5148 strings to 515 clusters, keeping 86% of the mutual information, and into 50 clusters keeping about 70% of the information. About the same compression efficiency was obtained for the NG1000 dataset . The relationship between the soft and hard clustering is demonstrated in the Information plane, i.e., the normalized mutual information values, :ti;; ~~ vs. Ik(~)' In this plane, the soft procedure is optimal since it is a direct maximization of I(Z; Y) while constraining I(Zi X). While the hard partition is suboptimal in this plane, as confirmed empirically, it provides an excellent starting point for reverse annealing. In figure 2 we present the results of the agglomerative procedure for NG1000 in the information plane, together with the reverse annealing for different values of IZI. As predicted by the theory, the annealing curves merge at various critical values of f3 into the globally optimal curve, which correspond to the "rate distortion function" for the information bottleneck problem. With the reverse annealing ("heating") procedure there is no need to identify the cluster splits as required in the original annealing ("cooling") procedure. As can be seen, the "phase diagram" is much better recovered by this procedure, suggesting a combination of agglomerative clustering and reverse annealing as the ultimate algorithm for this problem. References [1] L. D. Baker and A. K. McCallum. Distributional Clustering of Words for Text Classification In ACM SIGIR 98, 1998. [2] P. F . Brown, P.V. deSouza, R .L. Mercer, V.J. DellaPietra, and J.C. Lai. Class-based n-gram models of natural language. In Computational Linguistics, 18( 4}:467-479, 1992. [3] T. M. Cover and J. A. Thomas. Elements of Information Theory. John Wiley & Sons, New York, 1991. [4] R. EI-Yaniv, S. Fine, and N. Tishby. Agnostic classification of Markovian sequences. In Advances in Neural Information Processing (NIPS'97) , 1998. [5] T. Hofmann, J . Puzicha, and M. Jordan. Learning from dyadic data. In Advances in Neural Information Processing (NIPS'98), 1999. [6] J. Lin. Divergence Measures Based on the Shannon Entropy. IEEE Transactions on Information theory, 37(1}:145- 151, 1991. [7] K. Rose. Deterministic Annealing for Clustering, Compression, Classification, Regression, and Related Optimization Problems. Proceedings of the IEEE, 86(11}:2210- 2239, 1998. [8] F.C. Pereira, N. Tishby, and L. Lee. Distributional clustering of English words. In 30th Annual Meeting of the Association for Computational Linguistics, Columbus, Ohio, pages 183-190, 1993. [9] N. Tishby, W. Bialek, and F . C. Pereira. The information bottleneck method: Extracting relevant information from concurrent data. Yet unpublished manuscript, NEC Research Institute TR, 1998.
1651 |@word version:3 compression:4 seek:1 tr:1 reduction:5 contains:2 current:4 z2:4 recovered:1 xiyi:1 lang:1 yet:1 must:1 john:1 alphanumeric:1 partition:29 hofmann:1 shape:1 plot:1 xex:2 update:2 v:6 greedy:5 noamm:1 plane:10 mccallum:1 pointer:1 provides:1 lexicon:1 lx:1 direct:2 ik:1 introduce:1 expected:1 indeed:2 globally:1 decreasing:2 cardinality:5 becomes:1 bounded:1 notation:1 maximizes:1 moreover:1 baker:1 agnostic:1 israel:1 what:1 minimizes:2 string:7 fuzzy:1 guarantee:1 pseudo:1 every:14 ti:1 concave:3 exactly:1 tricky:1 zl:5 ly:1 positive:1 slonim:4 limit:7 usenet:1 merge:16 initialization:1 suggests:1 unique:1 digit:1 procedure:15 empirical:2 cascade:1 word:4 pre:1 close:2 context:1 py:2 map:4 deterministic:9 center:1 lagrangian:1 demonstrated:1 jerusalem:1 go:1 starting:3 sigir:1 analogous:1 hierarchy:2 play:1 exact:1 ixi:2 element:2 cooling:1 distributional:5 coarser:2 bottom:3 role:1 solved:1 capture:1 calculate:1 ormation:1 connected:1 news:1 decrease:6 substantial:1 rose:1 convexity:1 complexity:1 predictive:1 efficiency:3 easily:1 joint:2 various:1 talk:1 plz:1 dichotomy:3 heuristic:1 distortion:2 say:1 otherwise:2 compressed:3 final:1 obviously:1 advantage:2 sequence:2 indication:1 propose:1 zm:37 relevant:3 loop:1 achieve:1 normalize:1 kulback:1 cluster:27 yaniv:1 generating:1 categorization:1 ac:1 eq:5 strong:1 p2:1 predicted:1 come:1 merged:2 stochastic:2 vx:1 proposition:1 strictly:1 pl:1 considered:2 ic:1 exp:1 vary:1 smallest:1 concurrent:1 minimization:2 clearly:1 always:1 rather:1 corollary:1 focus:1 check:1 el:2 initially:1 relation:2 classification:5 among:1 denoted:2 constrained:1 special:1 mutual:26 equal:1 once:1 f3:6 construct:1 ng:2 constitutes:1 divergence:4 replaced:1 phase:2 consisting:2 maintain:1 organization:1 yielding:1 oth:1 tree:1 desired:1 plotted:1 minimal:1 soft:8 earlier:1 markovian:1 cover:1 disadvantage:1 logp:1 zn:1 maximization:1 soften:2 cost:2 introducing:1 subset:6 tishby:8 my:1 fundamental:1 huji:1 probabilistic:1 lee:1 together:1 iy:2 choose:1 possibly:1 stochastically:1 li:1 suggesting:1 coding:1 depends:1 try:1 reached:1 start:4 bayes:2 recover:2 maintains:1 xlx:10 minimize:1 il:1 yield:2 identify:3 correspond:1 confirmed:1 liebler:1 email:1 di:4 couple:2 proved:1 dataset:10 emerges:1 dimensionality:1 manuscript:1 higher:1 izi:9 formulation:1 evaluated:1 though:1 furthermore:1 just:1 hand:2 ei:1 undergoes:1 quality:1 columbus:1 lossy:1 verify:2 multiplier:2 true:1 normalized:5 contain:1 brown:1 symmetric:1 semantic:1 misc:1 self:5 inferior:1 naftali:1 theoretic:3 demonstrate:2 complete:1 temperature:2 variational:3 novel:1 recently:3 ohio:1 superior:1 functional:2 empirically:2 association:1 interpretation:2 significant:1 refer:1 pm:1 similarly:1 language:1 similarity:1 iyi:2 belongs:1 reverse:7 termed:1 forcing:1 certain:1 inequality:1 arbitrarily:1 meeting:1 seen:2 minimum:1 additional:1 converge:1 maximize:1 ii:1 full:1 stem:1 smooth:1 lin:1 qcs:1 lai:1 dkl:2 prediction:4 regression:1 iteration:1 normalization:2 want:2 fine:1 annealing:18 diagram:1 unlike:1 subject:1 member:4 jordan:1 extracting:1 constraining:1 split:4 easy:2 enough:1 newsgroups:5 zi:12 suboptimal:2 f3i:1 idea:1 bottleneck:16 ultimate:1 york:1 ylx:11 amount:1 locally:1 induces:1 category:1 ken:1 percentage:1 vy:1 zj:5 dotted:1 disjoint:1 per:3 zd:2 group:3 changing:1 fraction:1 ylz:3 inverse:1 striking:1 almost:1 bound:1 ilk:1 annual:1 occur:1 constraint:3 ri:2 x2:1 chx:1 prescribed:1 relatively:1 combination:1 son:1 character:2 wi:1 taken:1 equation:6 remains:1 know:1 end:2 operation:2 apply:2 hierarchical:1 enforce:1 pllq:1 alternative:1 lxi:1 original:5 thomas:1 top:3 clustering:14 linguistics:2 objective:1 question:1 already:2 lex:1 quantity:1 parametric:1 bialek:1 mx:1 evenly:1 topic:1 agglomerative:11 collected:1 trivial:2 length:1 code:1 relationship:2 hebrew:1 noam:1 proper:1 perform:1 markov:1 datasets:1 finite:2 looking:1 introduced:2 pair:6 yey:2 kl:1 namely:1 z1:4 required:1 unpublished:1 merges:1 established:1 zlx:4 nip:2 beyond:1 able:1 below:1 appeared:2 built:1 critical:2 natural:3 negativity:1 extract:1 text:2 prior:3 f3d:1 fully:1 loss:3 analogy:1 consistent:4 xp:1 article:1 principle:3 mercer:1 pi:2 surprisingly:1 last:1 keeping:2 english:1 side:2 institute:2 distributed:1 curve:13 transition:1 valid:1 gram:1 concavity:1 preprocessing:1 simplified:1 transaction:1 compact:1 obtains:1 desouza:1 corpus:2 xi:10 lip:2 zk:8 erwlse:1 yixi:1 excellent:1 necessarily:1 main:1 heating:1 dyadic:2 atheism:1 referred:1 wiley:1 sub:1 jsn:1 pereira:2 xl:1 ix:4 yizi:2 down:3 removing:1 specific:1 inset:1 jensen:3 alt:1 merging:5 magnitude:4 nec:1 entropy:3 explore:1 lagrange:2 contained:3 religion:1 determines:1 acm:1 conditional:3 loosing:2 hard:15 argmini:1 determined:2 specifically:1 reducing:1 called:3 shannon:4 meaningful:1 newsgroup:3 puzicha:1 mark:1 relevance:3 evaluate:5 ex:1
713
1,652
Support Vector Method for Multivariate Density Estimation Vladimir N. Vapnik Royal Halloway College and AT &T Labs, 100 Schultz Dr. Red Bank, NJ 07701 vlad@research.att.com Sayan Mukherjee CBCL, MIT E25-201 Cambridge, MA 02142 sayan@ai.mit.edu Abstract A new method for multivariate density estimation is developed based on the Support Vector Method (SVM) solution of inverse ill-posed problems. The solution has the form of a mixture of densities. This method with Gaussian kernels compared favorably to both Parzen's method and the Gaussian Mixture Model method. For synthetic data we achieve more accurate estimates for densities of 2, 6, 12, and 40 dimensions. 1 Introduction The problem of multivariate density estimation is important for many applications, in particular, for speech recognition [1] [7]. When the unknown density belongs to a parametric set satisfying certain conditions one can estimate it using the maximum likelihood (ML) method. Often these conditions are too restrictive. Therefore, non-parametric methods were proposed. The most popular of these, Parzen's method [5], uses the following estimate given data Xl, ... , Xl: (1) where K'Yl(t - Xi) is a smooth function such that J K'Yl(t - xi)dt = 1. Under some conditions on "Yl and K'Yl (t - Xi), Parzen's method converges with a fast asymptotic rate. An example of such a function is a Gaussian with one free parameter "Y; (the width) The structure of the Parzen estimator is too complex: the number of terms in (1) is equal to the number of observations (which can be hundreds of thousands). V. N. Vapnik and S. Mukherjee 660 Researchers believe that for practical problems densities can be approximated by a mixture with few elements (Gaussians for Gaussian Mixture Models (GMM)). Therefore, the following parametric density model was introduced m m P(x, a,~) = 2: (}:iP(X, ai, ~i)' (}: ~ 0, 2: (}:i = 1, i=l (3) i=l where P(x, ai, ~i) are Gaussians with different vectors ai and different diagonal covariance matrices ~i; (}:i is the proportion of the i-th Gaussian in the mixture. It is known [9] that for general forms of Gaussian mixtures the ML estimate does not exist. To use the ML method two values are specified: a lower bound on diagonal elements of the covariance matrix and an upper bound on the number of mixture elements. Under these constraints one can estimate the mixture parameters using the EM algorithm. This solution, however, is based on predefined parameters. In this article we use an SVM approach to obtain an estimate in the form of a mixture of densities. The approach has no free parameters. In our experiments it performs better than the GMM method. 2 Density estimation is an ill-posed problem A density p(t) is defined as the solution of the equation i~ p(t) dt = (4) F(x), where F(x) is the probability distribution function. Estimating a density from data involves solving equation (4) on a given set of densities when the distribution function F(x) is unknown but a random i.i.d. sample Xl, ... , Xe is given. The empirical distribution function Fe(x) is a good approximation of the actual distribution, where O( u) is the step-function. In the univariate case, for sufficiently large l the distribution of the supremum error between F(x) and Ft(x) is given by the Kolmogorov-Smirnov distribution P{sup IF(x) - Fe(x)1 < c/Vi} = 1- 22:( _1)k-1 exp{ -2c 2 k 2 }. x (5) k=l Hence, the problem of density estimation can be restated as solving equation (4) but replacing the distribution function F(x) with the empirical distribution function Fe(x) which converges to the true one with the (fast) rate 0(1)' for univariate and multivariate cases. The problem of solving the linear operator equation Ap Ft(x) is ill-posed. =F with approximation In the 1960's methods were proposed for solving ill-posed problems using approximations Ft converging to F as l increases. The idea of these methods was to 661 Support Vector Method for Multivariate Density Estimation introduce a regularizing functional O(P) (a semi-continuous, positive functional for which O(p) ~ c, c > 0 is a compactum) and define the solution Pt which is a trade-off between O(p) and IIAp - Ftll. The following two methods which are asymptotically equivalent [11] were proposed by Tikhonov [8] and Phillips [6] min [IIAp - Ftll 2 p + ItO(P)] minO(p) s.t. IIAp - Fill p , It > 0, It -+ 0, < et, et > 0, et -+ O. (6) (7) For the stochastic case it can be shown for both methods that if Ft(x) converges in probability to F(x) and Ii -+ 0 then for sufficiently large f and arbitrary 1/ and J.L the following inequality holds [10] [9] [3] (8) where f > fO(1/, J.L) and PEl (p,Pi), PE2(F, Fi ) are metrics in the spaces p and F. Since Fi(X) -+ F(x) in probability with the rate O( ~), from equation (8) it follows that if Ii 3 > O( ~) the solutions of equation (4) are consistent. Choice of regularization parameters For the deterministic case the residual method [2] can be used to set the regularization parameters (,i in (6) and ei in (7)) by setting the parameter (,i or ei) such that Pi satisfies the following IIApi - Fill = IIF(x) - Fi(X) II = Ui, (9) where Ui is the known accuracy of approximation of F(x) by Fi(X). We use this idea for the stochastic case. The Kolmogorov-Smirnov distribution is used to set Ui, Ui = c/...[i, where c corresponds to an appropriate quantile. For the multivariate case one can either evaluate the appropriate quantile analytically [4] or by simulations. The density estimation problem can be solved using either regularization method (6) or (7). Using method (6) with a L2 norm in image space F and regularization functional O(p) = (Tp, Tp) where T is a convolution operator, one obtains Parzen's method [10] [9] with kernels defined by operator T 4 SVM for density estimation We apply the SVM technique to equation (7) for density estimation. We use the C norm in (7) and solve equation (4) in a set of functions belonging to a Reproducing Kernel Hilbert Space (RKHS). We use the regularization functional O(P) = IIplit = (p,p)1i. (10) A RKHS can be defined by a positive definite kernel K (x, y) and an inner product (f, g)1i in Hilbert space 1-l such that (f(x),K(x,Y))1i = f(y) "If E 1-l. (11) V. N. Vapnik and S. Mukherjee 662 Note that any positive definite function K(x,y) has an expansion 00 K(x,y) = LAi<!)i(x)<Pi(Y) (12) i=l where Ai and <Pi(X) are eigenvalues and eigenfunctions of K(x, y). Consider the set of functions 00 (13) f(x,c) = LCi<Pi(X) i=l and the inner product 00 * ** ci Ci (f( X,C *) , f( X,C **) = ""' ~~. i=l (14) t Kernel (12), inner product (14), and set (13) define a RKHS and For functions from a RKHS the functional (10) has the form 00 2 C? O(p) = ""' ~ At. , i=l (15) t where Ai is the i-th eigenvalue of the kernel K(x, y). The choice of the kernel defines smoothness properties on the solution. To use method (7) to solve for the density in equation (4) in a RKHS with a solution satisfying condition (9) we minimize O(p) = (p,p)ll subject to the constraint We look for a solution of equation (4) with the form l p(t) = L fiiKl'l (Xi, t). (16) i=l Accounting for (16) and (11) minimizing (10) is equivalent to minimizing l O(p,p) = (p,p)ll = L i,j=l fiifijKl'l(Xi,Xj) (17) Support Vector Method for Multivariate Density Estimation 663 subject to constraints (18) i f3i ~ 0, I:f3i = 1. (19) i=1 This optimization problem is closely related to the SV regression problem with an O"i-insensitive zone [9]. It can be solved using the standard SVM technique. Generally, only a few of the f3i will be nonzero, the Xi corresponding to these f3i are called support vectors. Note that kernel (2) has width parameter 'Yi. We call the value of this parameter admissible if it satisfies constraint (18) (the solution satisfies condition (9)). The admissible set 'Ymin :::; 'Yi :::; 'Ymax is not empty since for Parzen's method (which also has form (16)) such a value does exist. Among the 'Yi in this admissible set we select the one for which O(P) is smallest or the number of support vectors is minimum. Choosing other kernels (for example Laplacians) one can estimate densities using non-Gaussian mixture models which for some problems are more appropriate [1]. 5 Experiments Several trials of estimates constructed from sampling distributions were examined. Boxplots were made of the L1 (p) norm over the trials. The horizontal lines of the boxplot indicate the 5%, 25%, 50%, 75%, and 95% quantiles of the error distribution. For the SVM method we set O"i = c/V'i, where c = .36, .41, .936, and 1.75 for two, six, twelve and forty dimensions. For Parzen's method 'Yi was selected using a leave-one out procedure. The GMM method uses the EM algorithm and sets all parameters except n, the upper bound on the number of terms in the mixture [7]. Figure (1) shows plots of the SVM estimate using a Gaussian kernel and the GMM estimate when 60 points were drawn form a mixture of a Gaussian and Laplacian in two dimensions. Figure (2a) shows four boxplots of estimating a density defined by a mixture of two Laplacians in a two dimensional space using 200 observations. Each boxplot shows outcomes of 100 trials: for the SVM method, Parzen's method, and the GMM method with parameters n = 2, and n = 4. Figure (2c) shows the distribution of the number of terms for the SVM method. Figure (2b) shows boxplots of estimating a density defined by the mixture of four Gaussians in a six dimensional space using 600 observations. Each boxplot shows outcomes of 50 trials: for the SVM method, Parzen's method, and the G MM method with parameters n = 4, and n = 8. Figure (2c) shows the distribution of the number of terms for the SVM method. Figure (3a) shows boxplots of outcomes of estimating a density defined by the mixture of four Gaussians and four Laplacians in a twelve dimensional space using V. N. Vapnik and S. Mukherjee 664 400 observations. Each boxplot shows outcomes of 50 trials: for the SVM method, Parzen's method, and the GMM method with parameter n = 8. Figure (3c) shows the distribution of the number of terms for the SVM method. Figure (3b) shows boxplots of outcomes of estimating a density defined by the mixture of four Gaussians and four Laplacians in a forty dimensional space using 480 observations. Each box-plot shows outcomes of 50 trials: for the SVM method, Parzen's method, and the GMM method with parameter n = 8. Figure (3c) shows the distribution of the number of terms for the SVM method. 6 Summary A method for multivariate density estimation based on the SVM technique for solving ill-posed problems is introduced. This method has a form of a mixture of densities. The estimate in general has only a few terms. In experiments on synthetic data this method is more accurate than the GMM method. References [1] S. Basu and C.A. Micchelli. Parametric density estimation for the classification of acoustic feature vectors in speech recognition. In Nonlinear Modeling, Advanced Black-Box Techniques. Kluwer Publishers, 1998. [2] V.A. Morozov. Berlin, 1984. M ethods for solving incorrectly posed problems. Springer-Verlag, [3] S. Mukherjee and V. Vapnik. Multivariate density estimation: An svm approach. AI Memo 1653, Massachusetts Institute of Technology, 1999. [4] S. Paramasamy. On multivariate kolmogorov-smirnov distribution. Statistics ability Letters, 15:140-155, 1992. fj Prob- [5] E. Parzen. On estimation of a probability density function and mode. Ann. Math. Statis. , 33:1065-1076, 1962. [6] D.L. Phillips. A technique for the numerical solution of integral equations of the first kind. J.Assoc. Comput. Machinery, 9:84- 97, 1962. [7] D. Reynolds and R. Rose. Robust text-independent speaker identification using gaussian mixture speaker models. IEEE Trans on Speech and Audio Processing, 3(1) :1-27, 1995. [8] A. N. Tikhonov. Solution of incorrectly formulated problems and the regularization method. Soviet Math. Dokl., 4:1035-1038, 1963. [9] V. N. Vapnik. Statistical learning theory. J. Wiley, 1998. [10] V.N. Vapnik and A.R. Stefanyuk. Nonparametric methods for restoring probability densities. Avtomatika i Telemekhanika, (8) :38-52, 1978. [11] V.V. Vasin. Relationship of several variational methods for the approximate solution of ill-posed problems. Math Not es, 7:161- 166, 1970. 665 Support Vector Method/or Multivariate Density Estimation , .-r ~ ,??-r? .~ (a) . (c) (b) Figure 1: (a) The true distribution (b) the GMM case with 4 mixtures (c) the Parzen case (d) the SVM case for 60 points. X 10-3 0.16 --,- 2.1 0.14 0.12 e 1.9 88 0.1 ~ ~O.08 0.06 0.04 , El9 + $ :::; I .S 0 --r ~ B , 1.5 1.4 GMM2 15z 1.6 ~ Parzen ~ 1.7 0.02 SVM ~ ~, --,-- -+SVM GMM4 (a) ~ Parzen GMM4 2-"'" GMMS 6 dirnll'l!ilQ(ls (c) (b) Figure 2: (a) Boxplots of the L1 (P) error for the mixture of two Laplacians in two dimensions for the SVM method, Parzen's method, and the GMM method with 2 and 4 Gaussians. (b) Boxplots of the L 1 (P) error for mixture of four Gaussians in six dimensions with the SVM method, Parzen's method, and the GMM method with 4 mixtures. (c) Boxplots of distribution of the number of terms for the SVM method for the two and six dimensional cases. 0.16r-~---------' 0.12 B 0.14 0.1 0.12 g0.1 ~0 .08 ~O.08 ~O.oe 25 ~ & 0.04 0.08 0.02 0.04 SVM Parzen (a) GMMS -- SVM Parzen (b) GMMS 12 Dim 40 Dim (c) Figure 3: (a) Boxplots of the Ll (p) error for the mixture of four Laplacians and four Gaussians in twelve dimensions for the SVM method , Parzen's method, and the GMM method with 8 Gaussians . (b) Boxplots of the L1 (P) error for the mixture of four Laplacians and four Gaussians in forty dimensions for the SVM method, Parzen's method, and the GMM method with 8 Gaussians. (c) Boxplots of distribution of the number of terms for the SVM method for the twelve and forty dimensional cases.
1652 |@word trial:6 involves:1 indicate:1 true:2 norm:3 proportion:1 smirnov:3 hence:1 regularization:6 closely:1 analytically:1 nonzero:1 micchelli:1 simulation:1 stochastic:2 avtomatika:1 covariance:2 accounting:1 ll:3 parametric:4 width:2 diagonal:2 speaker:2 berlin:1 g0:1 att:1 rkhs:5 reynolds:1 performs:1 l1:3 fj:1 hold:1 com:1 mm:1 sufficiently:2 image:1 variational:1 cbcl:1 exp:1 fi:4 relationship:1 minimizing:2 vladimir:1 numerical:1 functional:5 fe:3 favorably:1 smallest:1 memo:1 plot:2 statis:1 insensitive:1 estimation:15 kluwer:1 selected:1 unknown:2 upper:2 observation:5 cambridge:1 convolution:1 ai:7 phillips:2 smoothness:1 incorrectly:2 mit:2 math:3 gaussian:10 ftll:2 reproducing:1 arbitrary:1 constructed:1 multivariate:11 introduced:2 specified:1 belongs:1 ethods:1 likelihood:1 introduce:1 tikhonov:2 certain:1 verlag:1 inequality:1 acoustic:1 dim:2 xe:1 yi:4 trans:1 dokl:1 minimum:1 laplacians:7 actual:1 forty:4 estimating:5 semi:1 ii:3 royal:1 among:1 ill:6 classification:1 pel:1 smooth:1 kind:1 developed:1 residual:1 advanced:1 equal:1 lai:1 nj:1 sampling:1 technology:1 iif:1 laplacian:1 converging:1 look:1 regression:1 ymin:1 metric:1 assoc:1 kernel:10 few:3 text:1 l2:1 positive:3 asymptotic:1 publisher:1 eigenfunctions:1 ap:1 subject:2 black:1 examined:1 gmms:3 consistent:1 article:1 call:1 mixture:24 bank:1 pi:5 summary:1 practical:1 predefined:1 accurate:2 restoring:1 integral:1 xj:1 free:2 definite:2 f3i:4 inner:3 idea:2 machinery:1 procedure:1 institute:1 basu:1 empirical:2 six:4 dimension:7 modeling:1 made:1 tp:2 speech:3 gmm4:2 operator:3 schultz:1 generally:1 approximate:1 obtains:1 equivalent:2 deterministic:1 hundred:1 supremum:1 nonparametric:1 ml:3 too:2 l:1 restated:1 xi:6 sv:1 exist:2 synthetic:2 estimator:1 continuous:1 density:32 twelve:4 fill:2 morozov:1 yl:4 off:1 robust:1 parzen:21 e25:1 pt:1 four:11 ilq:1 expansion:1 complex:1 us:2 drawn:1 gmm:13 element:3 dr:1 recognition:2 satisfying:2 approximated:1 boxplots:11 asymptotically:1 mukherjee:5 ft:4 inverse:1 letter:1 prob:1 solved:2 quantiles:1 thousand:1 wiley:1 oe:1 vi:1 mino:1 trade:1 lab:1 rose:1 xl:3 sup:1 red:1 ui:4 bound:3 comput:1 ito:1 sayan:2 admissible:3 solving:6 minimize:1 audio:1 accuracy:1 constraint:4 boxplot:4 svm:28 identification:1 vapnik:7 min:1 kolmogorov:3 soviet:1 ci:2 lci:1 researcher:1 fast:2 fo:1 choosing:1 outcome:6 belonging:1 univariate:2 em:2 posed:7 solve:2 pe2:1 ability:1 statistic:1 springer:1 corresponds:1 ip:1 satisfies:3 popular:1 massachusetts:1 vlad:1 eigenvalue:2 equation:11 ma:1 hilbert:2 formulated:1 ann:1 product:3 telemekhanika:1 dt:2 gaussians:11 except:1 achieve:1 ymax:1 apply:1 appropriate:3 box:2 called:1 e:1 vasin:1 empty:1 zone:1 horizontal:1 select:1 replacing:1 ei:2 nonlinear:1 converges:3 leave:1 college:1 support:7 defines:1 mode:1 evaluate:1 believe:1 regularizing:1 restrictive:1 quantile:2
714
1,653
Semiparametric Approach to Multichannel Blind Deconvolution of Nonminimum Phase Systems L.-Q. Zhang, S. Amari and A. Cichocki Brain-style Information Systems Research Group, BSI The Institute of Physical and Chemical Research Wako shi, Saitama 351-0198, JAPAN zha@open.brain.riken.go.jp {amari,cia }@brain.riken.go.jp Abstract In this paper we discuss the semi parametric statistical model for blind deconvolution. First we introduce a Lie Group to the manifold of noncausal FIR filters. Then blind deconvolution problem is formulated in the framework of a semiparametric model, and a family of estimating functions is derived for blind deconvolution. A natural gradient learning algorithm is developed for training noncausal filters. Stability of the natural gradient algorithm is also analyzed in this framework. 1 Introduction Recently blind separation/deconvolution has been recognized as an increasing important research area due to its rapidly growing applications in various fields, such as telecommunication systems, image enhancement and biomedical signal processing. Refer to review papers [7] and [13] for details. A semi parametric statistical model treats a family of probability distributions specified by a finite-dimensional parameter of interest and an infinite-dimensional nuisance parameter [12]. Amari and Kumon [10] have proposed an approach to semiparametric statistical models in terms of estimating functions and elucidated their geometric structures and efficiencies by information geometry [1]. Blind source separation can be formulated in the framework of semi parametric statistical models. Amari and Cardoso [5] applied information geometry of estimating functions to blind source separation and derived an admissible class of estimating functions which includes efficient estimators. They showed that the manifold of mixtures is m-curvature free, so that we can design algorithms of blind separation without taking much care of misspecification of source probability functions. The theory of estimating functions has also been applied to the case of instantaneous mixtures, where independent source signals have unknown temporal correlations [3]. It is also applied to derive efficiency and superefficiency of demixing learning algorithms [4]. Most of these theories treat only blind source separation of instantaneous mixtures. It is only recently that the natural gradient approach has been proposed for multichannel blind L.-Q. Zhang, S. Amari and A. Cichocki 364 deconvolution [8], [18]. The present paper extends the geometrical theory of estimating functions to the semiparametric model of multichannel blind deconvolution. For the limited space, the detailed derivations and proofs are left to a full paper. 2 Blind Deconvolution Problem In this paper, as a convolutive mixing model, we consider a multichannel linear timeinvariant (LTI) systems, with no poles on the unit circle, of the form 00 x(k) = L Hps(k - p), (1) p=-oo where s(k) is an n-dimensional vector of source signals which are spatially mutually independent and temporarily identically independently distributed, and x(k) is an n-dimensional sensor vector at time k, k = 1,2, . . '. We denote the unknown mixing filter by H(z) = 2::-00 Hpz-p. The goal of multichannel blind deconvolution is to retrieve source signals s(k) only using sensor signals x(k), k = 1,2"", and certain knowledge of the source signal distributions and statistics. We carry out blind deconvolution by using another multichannel LTI system of the form y(k) = W(z)x(k), (2) where W(z) = 2:~= -N Wpz-P, N is the length of FIR filter W(z), y(k) [Yl (k), ... ,Yn(k)V is an n-dimensional vector of the outputs, which is used to estimate the source signals. When we apply W(z) to the sensor signal x(k), the global transfer function from s(k) to y(k) is defined by G(z) = W(z)H(z). The goal of the blind deconvolution task is to find W(z) such that G(z) = PAD(z), where P E R nxn is a permutation matrix, D(z) = diag{z-d 1 , ??? ,z- dn }, and A E R n x n is a nonsingular diagonal scaling matrix. 3 Lie Group on M (N, N) In this section, we introduce a Lie group to the manifold of noncausal FIR filters. The Lie group operations playa crucial role in the following discussion. The set of all the noncausal FIR filters W (z) of length N, having the constraint that W is nonsingular, is denoted by M(N,N) = {W(Z) I W(z) = .tN W.z - ? , det(W) # o}, (3) where W is an N x N block matrix, ... W_N+ll ... W - N+2 . . .. (4) Wo M(N, N) is a manifold of dimension n 2 (2N + 1). In general, multiplication of two filters in M(N, N) will enlarge the filter length and the result does belong to M(N, N) anymore. This makes it difficult to introduce the Riemannian structure to the manifold of noncausal FIR filters. In order to explore possible geometrical structures of M(N, N) which will lead to effective learning algorithms for W (z) , we define algebraic operations of filters in the Lie group framework. First, we introduce a novel filter decomposition of noncausal filters in M (N, N) into a product of two one-sided FIR filters [19], which is illustrated in Fig. 1. Blind Deconvolution ofNonminimum Phase Systems 365 Unknown :s(k) n: u(k) R(Z") ~ L(z) x(k) H(z) n i y(k) n i Mixing model Demixing model Figure 1: Illustration of decomposition of noncausal filters in M (N, N) Lemma 1 [19] If the matrix W is nonsingular, any noncausalfilter W(z) in M(N,N) has the decomposition W(z) = R(z)L(z-l), where R(z) = L::=o Rpz-P, L(Z-l) = L::=o LpzP are one-sided FIR filters. In the manifold M(N, N), Lie operations, multiplication as follows: For B(z), C(z) E M(N, N), B(z) * C(z) = [B(z)C(z)]N' * and inverse t, are defined Bt(z) = Lt(Z-l)Rt(z), (5) where [B(Z)]N is the truncating operator that any terms with orders higher than N in the polynomial B (z) are truncated, and the inverse of one-side FIR filters is recurrently defined by ~ = RO l , = - L::=l Rt_qRqRO l , p = 1,'" ,N. Refer to [18] for the detailed derivation. With these operations, both B(z) * C(z) and Bt (z) still remain in the manifold M (N, N). It is easy to verify that the manifold M (N, N) with the above operations forms a Lie Group. The identity element is E(z) = I. at 4 Semiparametric Approach to Blind Deconvolution We first introduce the basic theory of semiparametric models, and formulate blind deconvolution problem in the framework of the semiparametric models. 4.1 Semiparametric model Consider a general statistical model {p( Xj 6, en, where x is a random variable whose probability density function is specified by two parameters, 6 and 6 being the parameter of interest, and being the nuisance parameter. When the nuisance parameter is of infinite dimensions or of functional degrees of freedom, the statistical model is called a semiparametric model [12]. The gradient vectors of the log likelihood u(Zj 6, e) = 81ogp(z;6.e) v(z,,,'it 6 ~) -- 81ogp(z;6,e) are called the score functions of the parameter 80 ' 8( , e, e e of interest or shortly 6-score and the nuisance score or shortly -score, respectively. In the semiparametric model, it is difficult to estimate both the parameters of interest and nuisance parameters at the same time, since the nuisance parameter is of infinite degrees of freedom. The semiparametric approach suggests to use an estimating function to estimate the parameters of interest, regardless of the nuisance parameters. The estimating function is a vector function z(x, 6), independent of nuisance parameters satisfying the following conditions e e, (6) 1) Eo,dz(x,6)] = 0, 2) det(lC) i= 0, where IC = Eo,d 8z(x,6) 88 ]. (7) L.-Q. Zhang, S. Amari and A. Cichocki 366 3) Eo ,dz(x, 8)zT (x, 8)] < 00, (8) e. for all 8 and Generally speaking, it is difficult to find an estimating function. Amari and Kawanabe [9] studied the information geometry of estimating functions and provided a novel approach to find all the estimating functions. In this paper, we follow the approach to find a family of estimating functions for bind deconvolution. 4.2 Semiparametric Formulation for Blind Deconvolution Now we tum to formulate the blind deconvolution problem in the framework of semi parametric models. From the statistical point of view, the blind deconvolution problem is to estimate H(z) or H- 1 (z) from the observed data VL = {x(k), k = 1, 2", .}. The estimate includes two unknowns: One is the mixing filter H(z) which is the parameter of interest, and the other is the probability density function p(s) of sources, which is the nuisance parameter in the present case. FOIf blind deconvolution problem, we usually assume that source signals are zero-mean, E[sil' = 0, for i = 1"", n. In addition, we generally impose constraints on the recovered signals to remove the indeterminacy, (9) A typical example of the constraint is k i ( Si) = sf -1. Since the source signals are spatially mutually independent and temporally iid, the pdfr(s) can be factorized into a product form r(s) = TI~l r(si). The purpose of this paper is to find a family of estimating functions for blind deconvolution. Remarkable progress has been made recently in the theory of the semiparametric approach [9],[12]. It has been shown that the efficient score itself is an estimating function for blind separation. 5 Estimating Functions In this section, we give an explicit form of the score function matrix of interest and the nuisance tangent space, by using a local nonholonomic reparameterization. We then derive a family of estimating functions from it. 5.1 Score function matrix and its representation Since the mixing model is a matrix filter, we write an estimating function in the same matrix filter format N F(x;H(z)) = L Fp(x;H(z))z-P, (10) p= -N where F p(x; H(z)) are n x n-matrices. In orderto derive the explicit form ofthe H-score, we reparameterize the filter in a small neighborhood of H (z) by using a new variable matrix filter as H(z) * (I - X(z)), where 1 is the identity element of the manifold M(N, N). The variation X(z) represents a local coordinate system at the neighborhoodNH of H(z) on the manifold M(N, N). The variation dH(z) of H(z) is represented as dH(z) = -H(z) * dX(z). Letting W(z) = Ht(z), we have dX(z) = dW(z) * wt(z) , (11 ) which is a nonholonomic differential variable [6] since (11) is not integrable. With this representation of the parameters, we can obtain learning algorithms having the equivariant property [14] since the deviation dX(z) is independent of a specific H(z). The relative or the natural gradient of a cost function on the manifold can be automatically derived from this representation [21, [14], [18]. Blind Deconvolution ofNonminimum Phase Systems 367 {p(x;e,;)} {p(x;9,e)} Figure 2: Illustration of orthogonal decomposition of score functions The derivative of any cost function l(H(z)) with respect to a noncausal filter X(z) E:==-N Xpz-P is defined by L N 8l(H(z? _ aX(z) 8l(H(z? z-p p==-N axp (12) Now we can easily calculate the score function matrix of noncausal filter X(z), alogp(XiH(z),r) _ aX(z) where lP(y) N '"' () T(k _ ) -p L.J lP Y Y Pz , p=-N (13) = ('Pi(Yi),"', 'Pn(Yn?T, 'Pi(Yi) = dlO~;:(II/). and y = Ht(z)x. S.2 Efficient scores The efficient scores, denoted by UE(s; H(z), r), can be obtained by projecting the score function to the space orthogonal to the nuisance tangent space TJ'{z},r' which is illustrated in figure 2. In this section, we give an explicit form of the efficient scores for blind deconvolution. Lemma 2 [5} The tangent nuisance space TJ'{z),r is a linear space spanned by the nui- sance score junctions, denoted by TJ'{z),r = {E:=I CiOi(Si)} , where Ci are coefficients, and ai(si) are arbitrary junctions, satisfying the/ollowing conditions E[ai(si)2] < 00, E[sai(si)] = 0, E[k(si)ai(si?) = O. (14) We rewrite the score function (13) into the form U(s; H(z), r) = E!-N Upz-P, where Up = (cp(si(k?sj(k - P?nxn. Lemma 3 The off-diagonal elements UO,ij(S; H(z), r), i =/: j, and the delay elements Up,ij(S; H(z), r), P =/: 0, 0/ the score junctions are orthogonal to the nuisance tangent space TJ(z),r' Lemma 4 The projection 0/ UO,ii to the space orthogonal to the nuisance tangent space TJ(z),r is o/the/orm W(Si) = Co + CISi + C2k(Si), where Cj are any constants. 368 L.-Q. Zhang, S. Amari and A. Cichocki In summary we have the following theorem Theorem 1 The efficient score, UE(s; H(z), r) U: = L::=- NU: z-P, is given by <p(s)sT(k - p), for p:f. 0; (15) for off diagonal elements, for diagonal elements. U~ (16) For the instantaneous mixture case, it has been proven [9] that the semiparametric model for blind separation is information m-curvature free. This is also true in the multichannel blind deconvolution case. As a result, the efficient score function is an estimating function for blind deconvolution. Using this result, we easily derive a family of estimating functions for blind deconvolution N F(x(k); W(z)) = L c.p(y(k))y(k - pf z-P - I, (17) p=-N where y(k) = W(z)x(k), and <p is a given function vector. The estimating function is the efficient score function, when Co = Cl = 0, C2 = 1 and ki(Si) = c.pi(sdsi - 1. 6 Natural Gradient Learning and its Stability Ordinary stochastic gradient methods for parameterized systems suffer from slow convergence due to the statistical correlations of the processes signals. While quasi-Newton and related methods can be used to improve convergence, they also suffer from the mass computation and numerical instability, as well as local convergence. The natural gradient approach was developed to overcome the drawback of the ordinary gradient algorithm in the Riemannian spaces [2, 8, 15]. It has been proven that the natural gradient algorithm is an efficient algorithm in blind separation and blind deconvolution [2]. The efficient score function ( the estimating function) gives an efficient search direction for updating filter X(z) . Therefore, the updating rule for X(z) is described by Xk+l(Z) = Xk(z) -1]F(x(k), Wk(Z)), (18) where 1] is a learning rate. Since the new parameterization X(z) is defined by a nonholonomic transformation dX (z) = dW (z) * wt (z ), the deviation of W (z) is given by ~ W(z) = ~X(z) * W(z). (19) Hence, the natural gradient learning algorithm for W (z) is described as (20) Wk+l(Z) = Wk(Z) -1]F(x(k), Wk(z)) * Wk(z) , where F(x, W (z)) is an estimating function in the form (17). The stability ofthe algorithm (20) is equivalent to the one of algorithm (18). Consider the averaged version of algorithm (18) (21) ~X(z) = -1]E[F(x(k), Wk(Z))] . Analyzing the variational equation of the above equation and using the mutual independence and i.i.d. properties of source signals, we derive the stability conditions of learning algorithm (21) at vicinity of the true solution mi for i , j = 1," ' , n, where mi + 1 > 0, K.i > 0, K.iK.ja;aJ > 1, = E(c.p'(Yi (k))y;(k)], K.i = E[c.p~(Yi)], (22) a; = E[IYiI 2 ]. Therefore, we have the following theorem: Theorem 2 If the conditions (22) are satisfied, then the natural gradient learning algorithm (20) is locally stable. Blind Deconvolution of Nonminimum Phase Systems 369 References [1] S. Amari. Differential-geometrical methods in statistics, Lecture Notes in Statistics, volume 28. Springer, Berlin, 1985. [2] S. Amari. Natural gradient works efficiently in learning. Neural Computation, 10:251-276, 1998. [3 J S. Amari. ICA of temporally correlated signals - Learning algorithm. In Proceeding of 1st Inter. Workshop on Independent Component Analysis and Signal Separation, pages 37-42, Aussois, France, January, 11-15 1999. [4] S. Amari. Superefficiency in blind source separation. IEEE Trans. on Signal Processing, 47(4):936-944, April 1999. [5] S. Amari and J.-F. Cardoso. Blind source separation- semiparametric statistical approach. IEEE Trans. Signal Processing, 45:2692-2700, Nov. 1997. [6] S. Amari, T. Chen, and A. Cichocki. Nonholonomic orthogonal constraints in blind source separation. Neural Comput., to be published. [7] S. Amari and A. Cichocki. Adaptive blind signal processing- neural network approaches. Proceedings of the IEEE, 86(10):2026-2048,1998. [81 S. Amari, S. Douglas, A. Cichocki, and H. Yang. Multichannel blind deconvolution and equalization using the natural gradient. In Proc. IEEE Workshop on Signal Processing Adv. in Wireless Communications, pages 101-104, Paris, France, April 1997. [9] S. Amari and M. Kawanabe. Estimating functions in semiparametric statistical models. In I. V. Basawa, v.P. Godambe, and R.L. Taylor, editors, Estimating Functions, volume 32 of Monograph Series, pages 65-81. IMS, 1998. [10] S. Amari and M. Kumon. Estimation in the presence of infinitely many nuisance parameters in semiparametric statistical models. Ann. Statistics, 16: 1044-1068, 1988. [11] A.l. Bell and T.l. Sejnowski. An information maximization approach to blind separation and blind deconvolution. Neural Computation, 7: 1129-1159, 1995. [12] P. Bickel, C. Klaassen, Y. Ritov, and J. Wellner. Efficient and Adaptive Estimation for Semiparametric Models. The Johns Hopkins Univ. Press, Baltimore and London, 1993. [13] J.-F Cardoso. Blind signal separation: Statistical principles. Proceedings of the IEEE, 86(10):2009-2025,1998. [14] J.-F. Cardoso and B. Laheld. Equivariant adaptive source separation. IEEE Trans. Signal Processing, SP-43: 30 17-3029, Dec 1996. [15] A. Cichocki and R. Unbehauen. Robust neural networks with on-line learning for blind identification and blind separation of sources. IEEE Trans Circuits and Systems I: Fundamentals Theory and Applications, 43(11):894-906, 1996. [16] L. Tong, R.W. Liu, v.c. Soon, and Y.F. Huang. Indeterminacy and identifiability of blind identification. IEEE Trans. Circuits, Syst., 38(5):499-509, May 1991. [17] H. Yang and S. Amari. Adaptive on-line learning algorithms for blind separation: Maximum entropy and minimal mutual infonnation. Neural Comput., 9: 1457-1482, 1997. [18] L. Zhang, A. Cichocki, and S. Amari. Geometrical structures of FIR manifold and their application to multichannel blind deconvolution. In Proceeding of NNSP'99, pages 303-312, Madison, Wisconsin, August 23-25 1999. [19] L. Zhang, A. Cichocki, and S. Amari. Multichannel blind deconvolution of nonminimum phase systems using information backpropagation. In Proceedings of the Fifth International Conference on Neural Information Processing(ICONIP'99), page 210-216, Perth, Australia, Nov. 16-20 1999.
1653 |@word version:1 polynomial:1 open:1 decomposition:4 carry:1 liu:1 series:1 score:21 rpz:1 wako:1 recovered:1 si:12 dx:4 john:1 numerical:1 remove:1 parameterization:1 xk:2 nnsp:1 zhang:6 dn:1 c2:1 differential:2 ik:1 introduce:5 inter:1 ica:1 equivariant:2 growing:1 brain:3 automatically:1 pf:1 increasing:1 provided:1 estimating:24 circuit:2 factorized:1 mass:1 developed:2 transformation:1 temporal:1 sai:1 ti:1 ro:1 unit:1 uo:2 yn:2 bind:1 treat:2 local:3 analyzing:1 studied:1 suggests:1 co:2 limited:1 averaged:1 block:1 backpropagation:1 area:1 laheld:1 bell:1 projection:1 orm:1 operator:1 godambe:1 instability:1 equalization:1 equivalent:1 shi:1 dz:2 go:2 regardless:1 independently:1 truncating:1 formulate:2 estimator:1 rule:1 spanned:1 retrieve:1 reparameterization:1 stability:4 dw:2 variation:2 coordinate:1 element:6 satisfying:2 updating:2 observed:1 role:1 calculate:1 adv:1 monograph:1 rewrite:1 efficiency:2 easily:2 various:1 represented:1 riken:2 derivation:2 univ:1 effective:1 london:1 sejnowski:1 neighborhood:1 whose:1 amari:21 statistic:4 itself:1 product:2 rapidly:1 mixing:5 convergence:3 enhancement:1 derive:5 oo:1 nui:1 dlo:1 ij:2 progress:1 indeterminacy:2 cisi:1 direction:1 drawback:1 filter:24 stochastic:1 australia:1 ogp:2 ja:1 ic:1 bickel:1 purpose:1 estimation:2 proc:1 infonnation:1 kumon:2 sensor:3 pn:1 derived:3 ax:2 likelihood:1 vl:1 bt:2 pad:1 quasi:1 france:2 denoted:3 mutual:2 field:1 having:2 enlarge:1 represents:1 phase:5 geometry:3 freedom:2 interest:7 analyzed:1 mixture:4 tj:5 noncausal:9 orthogonal:5 taylor:1 circle:1 minimal:1 maximization:1 ordinary:2 cost:2 pole:1 deviation:2 saitama:1 delay:1 st:2 density:2 fundamental:1 international:1 yl:1 off:2 hopkins:1 satisfied:1 huang:1 fir:9 derivative:1 style:1 japan:1 syst:1 wk:6 includes:2 coefficient:1 blind:47 view:1 zha:1 identifiability:1 efficiently:1 nonsingular:3 ofthe:2 nonholonomic:4 identification:2 iid:1 published:1 proof:1 riemannian:2 mi:2 knowledge:1 cj:1 tum:1 higher:1 axp:1 follow:1 april:2 formulation:1 ritov:1 biomedical:1 c2k:1 correlation:2 aj:1 verify:1 true:2 hence:1 vicinity:1 chemical:1 bsi:1 spatially:2 illustrated:2 ll:1 ue:2 nuisance:15 iconip:1 tn:1 cp:1 geometrical:4 image:1 variational:1 instantaneous:3 novel:2 recently:3 functional:1 physical:1 jp:2 volume:2 belong:1 ims:1 refer:2 ai:3 hp:1 stable:1 playa:1 curvature:2 showed:1 certain:1 yi:4 integrable:1 care:1 impose:1 eo:3 recognized:1 signal:21 semi:4 ii:2 full:1 basic:1 nonminimum:3 dec:1 addition:1 semiparametric:18 baltimore:1 source:18 crucial:1 yang:2 presence:1 identically:1 easy:1 xj:1 independence:1 det:2 wellner:1 wo:1 suffer:2 algebraic:1 speaking:1 generally:2 detailed:2 cardoso:4 locally:1 multichannel:10 zj:1 write:1 aussois:1 group:7 douglas:1 lti:2 ht:2 inverse:2 parameterized:1 telecommunication:1 extends:1 family:6 klaassen:1 separation:17 scaling:1 ki:1 alogp:1 elucidated:1 constraint:4 superefficiency:2 reparameterize:1 format:1 remain:1 lp:2 projecting:1 sided:2 equation:2 mutually:2 discus:1 letting:1 junction:3 operation:5 apply:1 kawanabe:2 cia:1 anymore:1 shortly:2 perth:1 newton:1 madison:1 parametric:4 rt:1 diagonal:4 gradient:14 berlin:1 manifold:12 length:3 illustration:2 difficult:3 design:1 zt:1 unknown:4 finite:1 unbehauen:1 truncated:1 january:1 communication:1 misspecification:1 arbitrary:1 august:1 paris:1 specified:2 xih:1 nu:1 trans:5 usually:1 convolutive:1 fp:1 natural:11 improve:1 temporally:2 cichocki:10 review:1 geometric:1 tangent:5 multiplication:2 relative:1 nxn:2 wisconsin:1 lecture:1 permutation:1 proven:2 remarkable:1 sil:1 degree:2 principle:1 editor:1 pi:3 summary:1 wireless:1 free:2 soon:1 side:1 institute:1 taking:1 fifth:1 distributed:1 overcome:1 dimension:2 made:1 adaptive:4 sj:1 nov:2 global:1 search:1 transfer:1 robust:1 cl:1 diag:1 sp:1 fig:1 en:1 slow:1 tong:1 lc:1 explicit:3 sf:1 comput:2 lie:7 admissible:1 theorem:4 specific:1 recurrently:1 pz:1 deconvolution:31 demixing:2 workshop:2 ci:1 chen:1 entropy:1 lt:1 explore:1 infinitely:1 temporarily:1 springer:1 dh:2 goal:2 formulated:2 identity:2 ann:1 infinite:3 typical:1 wt:2 lemma:4 called:2 timeinvariant:1 correlated:1
715
1,654
Learning the Similarity of Documents: An Information-Geometric Approach to Document Retrieval and Categorization Thomas Hofmann Department of Computer Science Brown University, Providence, RI hofmann@cs.brown.edu, www.cs.brown.edu/people/th Abstract The project pursued in this paper is to develop from first information-geometric principles a general method for learning the similarity between text documents. Each individual document is modeled as a memoryless information source. Based on a latent class decomposition of the term-document matrix, a lowdimensional (curved) multinomial subfamily is learned. From this model a canonical similarity function - known as the Fisher kernel - is derived. Our approach can be applied for unsupervised and supervised learning problems alike. This in particular covers interesting cases where both, labeled and unlabeled data are available. Experiments in automated indexing and text categorization verify the advantages of the proposed method. 1 Introduction The computer-based analysis and organization of large document repositories is one oftoday's great challenges in machine learning, a key problem being the quantitative assessment of document similarities. A reliable similarity measure would provide answers to questions like: How similar are two text documents and which documents match a given query best? In a time, where searching in huge on-line (hyper-)text collections like the World Wide Web becomes more and more popular, the relevance of these and related questions needs not to be further emphasized. The focus of this work is on data-driven methods that learn a similarity function based on a training corpus of text documents without requiring domain-specific knowledge. Since we do not assume that labels for text categories, document classes, or topics, etc. are given at this stage, the former is by definition an unsupervised learning problem. In fact, the general problem of learning object similarities precedes many "classical" unsupervised learning methods like data clustering that already presuppose the availability of a metric or similarity function. In this paper, we develop a framework for learning similarities between text documents from first principles. In doing so, we try to span a bridge from the foundations of statistics in information geometry [13, 1] to real-world applications in information retrieval and text learning, namely ad hoc retrieval and text categorization. Although the developed general methodology is not limited to text documents, we will for sake of concreteness restrict our attention exclusively to this domain. Learning the Similarity ofDocuments 2 915 Latent Class Decomposition Memoryless Information Sources Assume we have available a set of documents V = {d l , ..? , dN} over some fixed vocabulary of words (or terms) W = {WI, ... , WM}. In an information-theoretic perspective, each document di can be viewed as an information source, i.e. a probability distribution over word sequences. Following common practice in information retrieval, we will focus on the more restricted case where text documents are modeled on the level of single word occurrences. This means that we we adopt the bag-of- words view and treat documents as memoryless information sources. I A. Modeling assumption: Each document is a memoryless information source. This assumption implies that each document can be represented by a multinomial probability distribution P(wjldi), which denotes the (unigram) probability that a generic word occurrence in document d i will be Wj. Correspondingly, the data can be reduced to some simple sufficient statistics which are counts n(di , Wj) of how often a word Wj occurred in a document d j ? The rectangular N x M matrix with coefficients n(di , Wj) is also called the term-document matrix. Latent Class Analysis Latent class analysis is a decomposition technique for contingency tables (cf. [5, 3] and the references therein) that has been applied to language modeling [15] ("aggregate Markov model") and in information retrieval [7] ("probabilistic latent semantic analysis"). In latent class analysis, an unobserved class variable Zk E Z = {zt, ... , ZK} is associated with each observation, i.e. with each word occurrence (d i , Wj). The joint probability distribution over V x W is a mixture model that can be parameterized in two equivalent ways P(di , Wj) = K K k=l k=l 2: P(zk)P(dilzk)P(wjlzk) = P(di) 2: P( WjIZk)P(Zk Idd . (1) The latent class model (1) introduces a conditional independence assumption, namely that di and Wj are independent conditioned on the state of the associated latent variable. Since the cardinality of Zk is typically smaller than the number of documents/words in the collection, Zk acts as a bottleneck variable in predicting words conditioned on the context of a particular document. To give the reader a more intuitive understanding of the latent class decomposition, we have visualized a representative subset of 16 "factors" from a K = 64 latent class model fitted from the Reuters2I578 collection (cf. Section 4) in Figure 1. Intuitively, the learned parameters seem to be very meaningful in that they represent identifiable topics and capture the corresponding vocabulary quite well. By using the latent class decomposition to model a collection of memory less sources, we implicitly assume that the overall collection will help in estimating parameters for individual sources, an assumption which has been validated in our experiments. B. Modeling assumption: Parameters for a collection of memoryless information sources are estimated by latent class decomposition. Parameter Estimation The latent class model has an important geometrical interpretation: the parameters ?1 == P( Wj IZk) define a low-dimensional subfamily of the multinomial family, S(?) == {11" E [0; I]M : 1I"j = :Ek 1/;k?1 for some1/; E [0; I]K, :Ek 1/;k = I}, i.e. all multinomials 11" that can be obtained by convex combinations from the set of "basis" vectors {?k : 1 :::; k :::; K}. For given ?-parameters, 1 Extensions to the more general case are possible, but beyond the scope of this paper. T. Hofmann 916 government presIdent tax budget cut s pending chairman cuts vice deficit company taxes reform named billion trading director america.n exchange general futures stock options index contracts market london exchanges executive chief officer board motors chrysler gm car ford test cars motor banks debt brazil new loans dlrs bankers b .. nk payments billion tr .. de japan j .. panese ec pct january february rise rose 1986 december year fell prices oil crude united officials community energy petroleum prices bpd barrels barrel european imports exploration price states marks currency dollar german airlines bundesbank aircraft central port mark boeing west employees dollars airline dealers vs areas cts weather net area loss normal min good shr crop qtr damage revs caused profit affected note people gold steel plant mining copper unlon air workers strike ton s silver metal production ounCeS food drug study aid s prod uct bllhon dlrs year surplu s deficit foreign current trade account reserves house rea.gan president administration congress trea.tment white company environmental products approval s ecretary told volcker reagans Figure 1: 16 selected factors from a 64 factor decomposition ofthe Reuters21578 collection. The displayed terms are the 10 most probable words in the class-conditional distribution P (Wj IZk) for 16 selected states Zk after the exclusion of stop words. each 1/;i , 1/;i == P(zkldi), will define a unique multinomial distribution rri E S(?). Since S( ?) defines a submanifold on the multinomial simplex, it corresponds to a curved exponential subfamily. 2 We would like to emphasis that we propose to learn both, the parameters within the family (the 1/;'s or mixing proportions P(Zk Idi )) and the parameters that define the subfamily (the ?'s or class-conditionals P(WjIZk)). The standard procedure for maximum likelihood estimation in latent variable models is the Expectation Maximization (EM) algorithm. In the E-step one computes posterior probabilities for the latent class variables, P( Zk )P( di IZk )P( Wj IZk) 2:1 P(zI)P(dilzt)P(wjlz/) P(Zk) P( di IZk )P( Wj IZk) P(di' Wj) (2) The M-step formulae can be written compactly as P(diIZk)} N M P(WjIZk) ex n(dn, wm)P(zkldn, wm) P(Zk) n=l m=l 2: 2: { X din djm (3) 1 where 6 denotes the Kronecker delta. Related Models As demonstrated in [7], the latent class model can be viewed as a probabilistic variant of Latent Semantic Analysis [2], a dimension reduction technique based on Singular Value Decomposition. It is also closely related to the non-negative matrix decomposition discussed in [12] which uses a Poisson sampling model and has been motivated by imposing non-negativity constraints on a decomposition by PCA. The relationship of the latent class model to clustering models like distributional clustering [14] has been investigated in [8]. [6] presents yet another approach to dimension reduction for multinomials which is based on spherical models, a different type of curved exponential subfamilies than the one presented here which is affine in the mean-value parameterization. 2Notice that graphical models with latent variable are in general stratified exponential families [4], yet in our case the geometry is simpler. The geometrical view also illustrates the well-known identifiability problem in latent class analysis. The interested reader is referred to [3]. As a practical remedy, we have used a Bayesian approach with conjugate (Dirichlet) prior distributions over all multinomials which for the sake of clarity is not described in this paper since it is very technical but nevertheless rather straightforward. Learning the Similarity of Documents 3 917 Fisher Kernel and Information Geometry The Fisher Kernel We follow the work of [9] to derive kernel functions (and hence similarity functions) from generative data models. This approach yields a uniquely defined and intrinsic (i. e. coordinate invariant) kernel, called the Fisher kernel. One important implication is that yardsticks used for statistical models carryover to the selection of appropriate similarity functions. In spite of the purely unsupervised manner in which a Fisher kernel can be learned, the latter is also very useful in supervised learning, where it provides a way to take advantage of additional unlabeled data. This is important in text learning, where digital document databases and the World Wide Web offer a huge background text repository. As a starting point, we partition the data log-likelihood into contributions from the various documents. The average log-probability of a document d i , i.e. the probability of all the word occurrences in d i normalized by document length is given by, M l(dd = L F(wjld j=l K i) log L P(WjIZk)P(Zkldi), F(wj Idi ) == 2: n(d(, ~j)) n d" k=l m (4) Wm which is up to constants the negative Kullback-Leibler divergence between the empirical distribution F(wjldi ) and the model distribution represented by (1). In order to derive the Fisher kernel, we have to compute the Fisher scores u(d i ; 0), i.e. the gradient of l(dd with respect to 0, as well as the Fisher information 1(0) in some parameterization 0 [13]. The Fisher kernel at {) is then given by [9] (5) Computational Considerations For computational reasons we propose to approximate the (inverse) information matrix by the identity matrix, thereby making additional assumptions about information orthogonality. More specifically, we use a variance stabilizing parameterization for multinomials - the square-root parameterization - which yields an isometric embedding of multinomial families on the positive part of a hypersphere [11]. In this parameterization, the above approximation will be exact for the multinomial family (disregarding the normalization constraint). We conjecture that it will also provide a reasonable approximation in the case of the subfamily defined by the latent class model. c. Simplifying assumption: The Fisher information in the square-root parameterization can be approximated by the identity matrix. Interpretation of Results Instead of going through the details of the derivation which is postponed to the end of this section, it is revealing to relate the results back to our main problem of defining a similarity function between text documents. We will have a closer look at the two contributions reSUlting from different sets of parameters. The contribution which stems from (square-root transformed) parameters P(Zk) is (in a simplified version) given by L P(Zk Idi)P(Zk Id n )/ P(Zk) . (6) k J( is a weighted inner product in the low-dimensional factor representation of the documents by mixing weights P(zkldi). This part of the kernel thus computes a "topical" overlap between documents and is thereby able to capture synonyms, i.e., words with an identical or similar meaning, as well as words referring to the same T. Hofmann 918 topic. Notice, that it is not required that di and dn actually have (many) terms in common in order to get a high similarity score. The contribution due to the parameters P(WjIZk) is of a very different type. Again using the approximation of the Fisher matrix, we arrive at the inner product K:(di, do) = l( P(Wj Idi ) I'>(wj Id o) ~ P(zkldi';(2~;:; Ido , Wj) ? (7) j( has also a very appealing interpretation: It essentially computes an inner product between the empirical distributions of di and dn , a scheme that is very popular in the context of information retrieval in the vector space model. However, common words only contribute, if they are explained by the same factor(s), i.e., if the respective posterior probabilities overlap. This allows to capture words with multiple meanings, so-called polysems. For example, in the factors displayed in Figure 1 the term "president" occurs twice (as the president of a company and as the president of the US). Depending on the document the word occurs in, the posterior probability will be high for either one of the factors, but typically not for both. Hence, the same term used in different context and different meanings will generally not increase the similarity between documents, a distinction that is absent in the naive inner product which corresponds to the degenerate case of K = 1. Since the choice of K determines the coarseness of the identified "topics" and different resolution levels possibly contribute useful information, we have combined models by a simple additive combination of the derived inner products. This combination scheme has experimentally proven to be very effective and robust. D. Modeling assumption: Similarities derived from latent class decompositions at different levels of resolution are additively combined. In summary, the emergence of important language phenomena like synonymy and polysemy from information-geometric principles is very satisfying and proves in our opinion that interesting similarity functions can be rigorously derived, without specific domain knowledge and based on few, explicitly stated assumptions (A-D). Technical Derivation Define Pjk == 2v'P(wjlzk), then 8l(d j ) oP(WjIZk) = . fp(wjlzk) P(wjldi ) P(zkJdd oP(WjIZk) OPjk V P(wjld i ) P(wjlddP(Zkldi' Wj) v'P(WjIZk) Similarly we define Pk = 2v'P(Zk). Applying Bayes' rule to substitute P(zkldd in l(dd (i.e. P(zkldd = P(zk)P(di/zk)/P(di)) yields 8l(dd OPk 8l(d;) OP(Zk) = v'P(Zk) P(dilzk) ~ P(wjldd P(W ' IZk) OP(Zk) OPk P(dd ~ P(WjJdd J J P(zkJdd ~ P(wj1di)p( I ) P( zkld j ) L...J W? Zk ~ ~==== v'P(Zk) j P(wjldi) J v'P(Zk) . The last (optional) approximation step makes sense whenever P(wjld j ) ~ P(wjldi ). Notice that we have ignored the normalization constraints which would yield a (reactive) term that is constant for each multinomial. Experimentally, we have observed no deterioration in performance by making these additional simplifications. Learning the Similarity ofDocuments VSM VSM++ Medline 44.3 67.2 919 Cranfield 29.9 37.9 CACM 17.9 27.5 CISI 12.7 20.3 Table 1: Average precision results for the vector space baseline method (VSM) and the Fisher kernel approach (VSM ++) for 4 standard test collections, Medline, Cranfield, CACM, and CIS!. I earn 20x sub lOx sub 5x sub all data lOx cv SVM SVM++ kNN kNN++ SVM SVM++ kNN kNN++ SVM SVM++ kNN kNN++ SVM SVM++ kNN kNN++ 5.51 4.56 5.91 5.05 4.88 4.11 5.51 4.94 4.09 3.64 5.13 4.74 2.92 2.98 4.17 4.07 acq 7.67 5.37 9.64 7.80 5.54 4.84 9.23 7.47 4.40 4.15 8.70 6.99 3.21 3.15 6.69 5.34 money 3.25 2.08 3.24 3.11 2.38 2.08 2.64 2.42 2.10 1.78 2.27 2.22 1.20 1.21 1.78 1.73 grain 2.06 1.71 2.54 2.35 1.71 1.42 2.55 2.28 1.32 0.98 2.40 2.18 0.77 0.76 1.73 1.58 crude 2.50 1.53 2.42 1.95 1.88 1.45 2.42 1.88 1.46 1.19 2.23 1.74 0.92 0.86 1.42 1.18 average 4.20 3.05 4.75 4.05 3.27 2.78 4.47 3.79 2.67 2.35 4.14 3.57 1.81 1.79 3.16 2.78 lmprov. +27.4% - +14.7% - +15.0% +15.2% - +12 .1% +13.7% - +0.6% +12.0% Table 2: Classification errors for k-nearest neighbors (kNN) SVMs (SVM) with the naive kernel and with the Fisher kernel(++) (derived from J( = 1 and J( = 64 models) on the 5 most frequent categories of the Reuters21578 corpus (earn, acq, monex-fx, grain, and crude) at different subsampling levels. 4 Experimental Results We have applied the proposed method for ad hoc information retrieval , where the goal is to return a list of documents, ranked with respect to a given query. This obviously involves computing similarities between documents and queries. In a follow-up series of experiments to the ones reported in [7] - where kernels K(d i , dn ) = ~k P(Zk Idi)P( Zk Idn ) and JC(di ' dn ) = ~j P(Wj Idi )P(wjldn ) have been proposed in an ad hoc manner - we have been able to obtain a rigorous theoretical justification as well as some additional improvements. Average precision-recall values for four standard test collections reported in Table 1 show that substantial performance gains can be achieved with the help of a generative model (cf. [7] for details on the conducted experiments). To demonstrate the utility of our method for supervised learning problems, we have applied it to text categorization, using a standard data set in the evaluation, the Reuters21578 collections of news stories. We have tried to boost the performance of two classifiers that are known to be highly competitive for text categorization: the k- nearest neighbor method and Support Vector Machines (SVMs) with a linear kernel [10]. Since we are particularly interested in a setting, where the generative model is trained on a larger corpus of unlabeled data, we have run experiments where the classifier was only trained on a subsample (at subsampling factors 20x,10x,5x). The results are summarized in Table 2. Free parameters of the base classifiers have been optimized in extensive simulations with held-out data. The results indicate T. Hofmann 920 that substantial performance gains can be achieved over the standard k-nearest neighbor method at all subsampling levels. For SVMs the gain is huge on the subsampled data collections, but insignificant for SVMs trained on all data. This seems to indicate that the generative model does not provide any extra information, if the SVM classifier is trained on the same data. However, notice that many interesting applications in text categorization operate in the small sample limit with lots of unlabeled data. Examples include the definition of personalized news categories by just a few example, the classification and/or filtering of email, on-line topic spotting and tracking, and many more. 5 Conclusion We have presented an approach to learn the similarity of text documents from first principles. Based on a latent class model, we have been able to derive a similarity function, that is theoretically satisfying, intuitively appealing, and shows substantial performance gains in the conducted experiments. Finally, we have made a contribution to the relationship between unsupervised and supervised learning as initiated in [9] by showing that generative models can help to exploit unlabeled data for classification problems. References [1] Shun'ichi Amari. Differential-geometrical methods in statistics. Springer-Verlag, Berlin, New York, 1985. [2] S. Deerwester, S. T. Dumais, G. W. Furnas, T. K. Landauer, and R. Harshman. Indexing by latent semantic analysis. Journal of the American Society for Information Science, 41:391-407, 1990. [3] M. J. Evans, Z. Gilula, and I. Guttman. Latent class analysis of two-way contingency tables by Bayesian methods. Biometrika, 76(3):557-563, 1989. [4] D. Geiger, D. Heckerman, H. King, and C. Meek. Stratified exponential families: Graphical models and model selection. Technical Report MSR-TR-98-31, Microsoft Research, 1998. [5] Z. Gilula and S. J . Haberman. Canonical analysis of contingency tables by maximum likelihood. Journal of the American Statistical Association, 81(395):780-788, 1986. [6] A. Gous. Exponential and Spherical Subfamily Models. PhD thesis, Stanford, Statistics Department, 1998. [7] T. Hofmann. Probabilistic latent semantic indexing. In Proceedings of the 22th International Conference on Research and Development in Information Retrieval (SIGIR), pages 50-57, 1999. [8] T . Hofmann, J. Puzicha, and M. I. Jordan. Unsupervised learning from dyadic data. In Advances in Neural Information Processing Systems 11. MIT Press, 1999. [9] T. Jaakkola and D. Haussler. Exploiting generative models in discriminative classifiers. In Advances in Neural Information Processing Systems 11. MIT Press, 1999. [lO] T. Joachims. Text categorization with support vector machines: Learning with many relevant features. In International Conference on Machine Learning (ECML), 1998. [ll] R.E. Kass and P. W. Vos. Geometrical foundations of asymptotic inference. Wiley, New York, 1997. [12] D. Lee and S. Seung. Learning the parts of objects by non-negative matrix factorization. Nature, 401:788-791, 1999. [13] M. K. Murray and J. W. Rice. Differential geometry and statistics. Chapman & Hall, London, New York, 1993. [14] F.C.N. Pereira, N.Z. Tishby, and L. Lee. Distributional clustering of English words. In Proceedings of the ACL, pages 183- 190, 1993. [15] L. Saul and F . Pereira. Aggregate and mixed-order Markov models for statistical language processing. In Proceedings of the 2nd International Conference on Empirical Methods in Natural Language Processing, 1997.
1654 |@word aircraft:1 msr:1 repository:2 version:1 proportion:1 coarseness:1 seems:1 nd:1 additively:1 simulation:1 tried:1 dealer:1 decomposition:11 simplifying:1 thereby:2 profit:1 tr:2 reduction:2 series:1 exclusively:1 united:1 score:2 document:37 current:1 ka:1 yet:2 import:1 written:1 grain:2 evans:1 additive:1 partition:1 hofmann:7 motor:2 v:1 pursued:1 selected:2 generative:6 parameterization:6 hypersphere:1 provides:1 idi:6 contribute:2 simpler:1 dn:6 differential:2 director:1 manner:2 theoretically:1 market:1 approval:1 spherical:2 company:3 food:1 cardinality:1 haberman:1 becomes:1 project:1 estimating:1 barrel:2 developed:1 unobserved:1 quantitative:1 act:1 biometrika:1 classifier:5 harshman:1 positive:1 treat:1 congress:1 limit:1 id:2 initiated:1 acl:1 emphasis:1 therein:1 twice:1 limited:1 factorization:1 stratified:2 unique:1 practical:1 practice:1 procedure:1 area:2 empirical:3 drug:1 weather:1 revealing:1 word:18 synonymy:1 spite:1 get:1 unlabeled:5 selection:2 context:3 applying:1 www:1 equivalent:1 demonstrated:1 straightforward:1 attention:1 starting:1 convex:1 rectangular:1 resolution:2 stabilizing:1 sigir:1 rule:1 haussler:1 embedding:1 searching:1 coordinate:1 fx:1 president:5 brazil:1 rri:1 justification:1 gm:1 exact:1 us:1 approximated:1 satisfying:2 particularly:1 cut:2 distributional:2 labeled:1 database:1 observed:1 capture:3 wj:18 news:2 trade:1 qtr:1 rose:1 substantial:3 seung:1 rigorously:1 trained:4 purely:1 basis:1 compactly:1 joint:1 stock:1 represented:2 america:1 various:1 derivation:2 effective:1 london:2 query:3 precedes:1 presuppose:1 aggregate:2 hyper:1 cacm:2 quite:1 larger:1 stanford:1 amari:1 statistic:5 knn:9 emergence:1 ford:1 obviously:1 hoc:3 advantage:2 sequence:1 net:1 propose:2 lowdimensional:1 product:6 frequent:1 relevant:1 mixing:2 degenerate:1 tax:2 gold:1 intuitive:1 billion:2 exploiting:1 categorization:7 silver:1 object:2 help:3 derive:3 develop:2 depending:1 nearest:3 op:4 cisi:1 c:2 involves:1 implies:1 trading:1 indicate:2 closely:1 exploration:1 opinion:1 shun:1 exchange:2 government:1 pjk:1 probable:1 extension:1 hall:1 normal:1 great:1 scope:1 reserve:1 adopt:1 estimation:2 bag:1 label:1 bridge:1 vice:1 reuters21578:3 weighted:1 mit:2 rather:1 jaakkola:1 derived:5 focus:2 validated:1 joachim:1 improvement:1 likelihood:3 rigorous:1 baseline:1 dollar:2 sense:1 inference:1 foreign:1 typically:2 going:1 transformed:1 interested:2 overall:1 classification:3 reform:1 development:1 sampling:1 chapman:1 identical:1 look:1 unsupervised:6 future:1 simplex:1 carryover:1 report:1 few:2 divergence:1 individual:2 subsampled:1 geometry:4 microsoft:1 organization:1 huge:3 mining:1 highly:1 evaluation:1 introduces:1 mixture:1 held:1 implication:1 closer:1 worker:1 respective:1 theoretical:1 fitted:1 modeling:4 cover:1 opk:2 maximization:1 subset:1 trea:1 submanifold:1 conducted:2 tishby:1 reported:2 providence:1 answer:1 ido:1 combined:2 referring:1 dumais:1 international:3 told:1 guttman:1 lee:2 probabilistic:3 contract:1 earn:2 again:1 central:1 medline:2 thesis:1 possibly:1 ek:2 american:2 return:1 japan:1 account:1 de:1 summarized:1 availability:1 coefficient:1 jc:1 caused:1 explicitly:1 ad:3 try:1 view:2 root:3 lot:1 doing:1 wm:4 bayes:1 option:1 competitive:1 identifiability:1 contribution:5 acq:2 air:1 square:3 variance:1 yield:4 ofthe:1 bayesian:2 whenever:1 email:1 definition:2 shr:1 energy:1 associated:2 di:15 stop:1 gain:4 popular:2 recall:1 knowledge:2 car:2 actually:1 back:1 supervised:4 follow:2 methodology:1 isometric:1 just:1 stage:1 uct:1 web:2 assessment:1 defines:1 oil:1 normalized:1 brown:3 verify:1 requiring:1 remedy:1 former:1 hence:2 din:1 memoryless:5 leibler:1 semantic:4 white:1 ll:1 uniquely:1 theoretic:1 demonstrate:1 geometrical:4 meaning:3 consideration:1 common:3 multinomial:12 discussed:1 occurred:1 interpretation:3 association:1 employee:1 imposing:1 cv:1 similarly:1 vos:1 language:4 similarity:22 money:1 etc:1 base:1 posterior:3 exclusion:1 perspective:1 driven:1 verlag:1 postponed:1 additional:4 strike:1 currency:1 multiple:1 stem:1 technical:3 match:1 offer:1 retrieval:8 variant:1 crop:1 essentially:1 metric:1 expectation:1 poisson:1 kernel:15 represent:1 normalization:2 deterioration:1 achieved:2 rea:1 background:1 conditionals:1 singular:1 source:8 extra:1 operate:1 airline:2 fell:1 december:1 idd:1 seem:1 jordan:1 automated:1 independence:1 zi:1 restrict:1 bpd:1 identified:1 inner:5 administration:1 absent:1 bottleneck:1 motivated:1 pca:1 utility:1 york:3 ignored:1 useful:2 generally:1 visualized:1 category:3 svms:4 reduced:1 canonical:2 notice:4 djm:1 estimated:1 delta:1 affected:1 officer:1 key:1 four:1 ichi:1 nevertheless:1 clarity:1 concreteness:1 year:2 deerwester:1 run:1 inverse:1 parameterized:1 named:1 arrive:1 family:6 reader:2 reasonable:1 geiger:1 ct:1 pct:1 meek:1 simplification:1 identifiable:1 kronecker:1 constraint:3 orthogonality:1 ri:1 personalized:1 sake:2 span:1 min:1 conjecture:1 department:2 combination:3 idn:1 conjugate:1 smaller:1 heckerman:1 em:1 wi:1 appealing:2 rev:1 making:2 alike:1 intuitively:2 restricted:1 indexing:3 invariant:1 explained:1 payment:1 count:1 german:1 end:1 available:2 gous:1 izk:7 generic:1 appropriate:1 occurrence:4 thomas:1 substitute:1 denotes:2 clustering:4 cf:3 dirichlet:1 gan:1 graphical:2 subsampling:3 include:1 exploit:1 prof:1 murray:1 february:1 classical:1 society:1 question:2 already:1 occurs:2 damage:1 chairman:1 gradient:1 deficit:2 berlin:1 topic:5 reason:1 length:1 modeled:2 index:1 relationship:2 relate:1 negative:3 rise:1 boeing:1 steel:1 stated:1 zt:1 observation:1 markov:2 petroleum:1 curved:3 tment:1 january:1 displayed:2 defining:1 optional:1 ecml:1 topical:1 community:1 copper:1 namely:2 required:1 extensive:1 optimized:1 learned:3 distinction:1 boost:1 beyond:1 able:3 spotting:1 fp:1 challenge:1 reliable:1 memory:1 debt:1 overlap:2 ranked:1 natural:1 predicting:1 scheme:2 negativity:1 lox:2 naive:2 text:19 prior:1 geometric:3 understanding:1 asymptotic:1 loss:1 plant:1 mixed:1 interesting:3 filtering:1 proven:1 digital:1 foundation:2 contingency:3 executive:1 affine:1 sufficient:1 metal:1 principle:4 port:1 bank:1 dd:5 story:1 production:1 lo:1 summary:1 cranfield:2 last:1 free:1 english:1 wide:2 neighbor:3 saul:1 correspondingly:1 dimension:2 vocabulary:2 world:3 computes:3 collection:11 made:1 simplified:1 ec:1 approximate:1 implicitly:1 kullback:1 corpus:3 discriminative:1 landauer:1 latent:26 prod:1 chief:1 table:7 nature:1 learn:3 zk:26 pending:1 robust:1 investigated:1 european:1 domain:3 official:1 polysemy:1 pk:1 main:1 synonym:1 subsample:1 dyadic:1 representative:1 west:1 referred:1 board:1 vsm:4 aid:1 furnas:1 precision:2 sub:3 wiley:1 pereira:2 exponential:5 crude:3 house:1 formula:1 specific:2 emphasized:1 unigram:1 showing:1 ton:1 disregarding:1 svm:10 list:1 insignificant:1 intrinsic:1 ci:1 phd:1 conditioned:2 budget:1 illustrates:1 nk:1 subfamily:7 tracking:1 springer:1 corresponds:2 environmental:1 determines:1 rice:1 conditional:2 viewed:2 identity:2 goal:1 king:1 banker:1 price:3 fisher:13 experimentally:2 loan:1 specifically:1 called:3 experimental:1 meaningful:1 puzicha:1 people:2 mark:2 latter:1 support:2 yardstick:1 relevance:1 reactive:1 phenomenon:1 ex:1
716
1,655
Search for Information Bearing Components in Speech Howard Hua Yang and Hynek Hermansky Department of Electrical and Computer Engineering Oregon Graduate Institute of Science and Technology 20000 NW, Walker Rd., Beaverton, OR97006, USA {hyang,hynek}@ece.ogi.edu, FAX:503 7481406 Abstract In this paper, we use mutual information to characterize the distributions of phonetic and speaker/channel information in a timefrequency space. The mutual information (MI) between the phonetic label and one feature, and the joint mutual information (JMI) between the phonetic label and two or three features are estimated . The Miller's bias formulas for entropy and mutual information estimates are extended to include higher order terms. The MI and the JMI for speaker/channel recognition are also estimated. The results are complementary to those for phonetic classification. Our results show how the phonetic information is locally spread and how the speaker/channel information is globally spread in time and frequency. 1 Introduction Speech signals typically carry information about number of target sources such as linguistic message, speaker identity, and environment in which the speech was produced. In most realistic applications of speech technology, only one or a few information targets are important. For example, one may be interested in identifying the message in the signal regardless of the speaker or the environments in which the speech was produced, or the identification of the speaker is needed regardless of the words the targeted speaker is saying. Thus, not all components of the signal may be equally relevant for a decoding of the targeted information in the signal. The speech research community has at its disposal rather large speech databases which are mainly used for training and testing automatic speech recognition (ASR) systems. There have been relatively few efforts to date to use such databases for deriving reusable knowledge about speech and speech communication processes which could be used for improvements of ASR technology. In this paper we apply information-theoretic approaches to study a large hand-labeled data set of fluent speech to learn about the information structure of the speech signal including the distribution of speech information in frequency and in time. Based on the labeled data set, we analyze the relevancy of the features for phonetic 804 H. H. YangandH Hermansky classifications and speaker/channel variability. The features in this data set are labeled with respect to underlying phonetic classes and files from which the features come from. The phoneme labels relate to the linguistic message in the signal, and the file labels carry the information about speakers and communication channels (each file contains speech of a single speaker transmitted through one telephone channel). Thus, phoneme and file labels are two target variables for statistical inference. The phoneme labels take 19 different values corresponding to 19 broad phoneme categories in the OGI Stories database [2]. The file labels take different values representing different speakers in the OGI Stories database. The relevancy of a set of features is measured by the joint mutual information (JMI) between the features and a target variable. The phoneme target variable represents in our case the linguistic message . The file target variable represents both different speakers and different telephone channels. The joint mutual information between a target variable and the features quantifies the relevancy of the features for that target variable. Mutual information measure the statistical dependence between random variables. Morris et al (1993) used mutual information to find the critical points of information for classifying French Vowel-Plosive-Vowel utterances. Bilmes(1998) showed recently that the information appears to be spread over relatively long temporal spans. While Bilmes used mutual information between two variables on nonlabeled data to reveal the mutual dependencies between the components of the spectral energies in time and frequency, we focused on joint mutual information between the phoneme labels or file labels and one, two or three feature variables in the time-frequency plane[7, 6] and used this concept to gain insight into how information about phonemes and speaker/channel variability is distributed in the time-frequency plane. 2 Data Set and Preprocessing The data set used in this paper is 3-hour phonetically labeled telephone speech, a subset of the English portion (Stories) ofthe OGI multi-lingual database [2] containing approximately 50 seconds of extemporaneous speech from each of 210 different speakers. The speech data is labeled by a variable Y taking 19 values representing 19 most often occurring phoneme categories. The average phoneme duration is about 65 ms and the number of phoneme instances is 6542l. Acoustic features X (fk, t) for the experiments are derived from a short-time analysis of the speech signal with a 20 ms analysis window (Hamming) at the frame t advanced in 10 ms steps. The logarithmic energy at a frequency fk is computed from the squared magnitude FFT using a critical-band spaced (log-like in the frequency variable) weighting function in a manner similar to that of the computation of Perceptual Linear Prediction coefficients [3]. In particular, the 5-th, 8-th and 12th bands are centered around 0.5, 1 and 2 kHz respectively. Each feature X(fk, t) is labeled by a phoneme label YP(t) and a file label Y J (t). We use mutual information to measure the relevancy of X(/k, t - d) across all frequencies fk and in a context window - D ::; d ::; + D for the phoneme classification and the speaker/channel identification. 3 Estimation of MI and Bias Correction In this paper, we only consider the mutual information (MI) between discrete random variables. The phoneme label and the file label are discrete random variables. Search for Information Bearing Components in Speech 805 However, the feature variables are bounded continuous variables. To obtain the quantized features, we divide the maximum range of the observed features into cells of equal volume so that we can use histogram to estimate mutual information defined by " p(x, y) I(Xi Y) = L...Jp(x, y) log2 (x) ( )' x,y P PY If X and Yare jointly Gaussian, then I(Xi Y) = -~ In(1 - p2) where p is the correlation coefficient between X and Y. However, for speech data the feature variables are generally non-Gaussian and target variables are categorical type variables. Correlations involving a categorical variable are meaningless. The MI can also be written as I(XiY) = H(X) = + H(Y) - H(X, Y) H(Y) - H(YIX) = H(X) - H(XIY) (1) where H (Y IX) is a conditional entropy defined by - H(YIX) = - L:p(x) L:p(Ylx) log2P(ylx). x y The two equations in (1) mean that the MI is the uncertainty reduction about Y give X or the uncertainty reduction about X give Y. Based on the histogram, H(X) is estimated by H(X) = - L: ni log2 ni . n n ~ where ni is the number of data points in the i-th cell and n is the data size. And I(X i Y) is estimated by i(Xi Y) = H(X) + H(Y) - H(X, Y). Miller(1954)[4] has shown that H(X) is an underestimate of H(X) and i(Xi Y) is an overestimate of I (X i Y) . The biases are r - A 1 E[H(X)] - H(X) = - 2In(2)n E[i(X;Y)]-I(X;Y) 1 (2) + O( n 2 ) = (r-l)(c-l) +O(~) 2In(2)n n2 (3) where rand c are the number of cells for X and Y respectively. Interestingly, the first order terms in (2) and (3) do not depend on the probability distribution. After using these formulas to correct the'estimates, the new estimates have the same variances as the old estimates but with reduced biases. However, these formulas break down when rand n are of the same order. Extending Miller's approach, we find a high order correction for the bias. Let {pd be the probability distribution of X, then E[H(X)] - H(X) r-l 1 = - 2In(2)n + 6In(2)n2 (S( {Pi}) 1 1 --(S({pd) - 1) + 0(-) 3 4 4n n 3r + 2) (4) H. H. Yang and H. Hermansky 806 The last two terms in the bias (4) depend on the unknown probabilities {pd. In practice they are approximated by the relative frequency estimates. Similarly, we can find the bias formulas of the high order terms O(nl:2) and O(n\) for the MI estimate. When X is evenly distributed, Pi = 1/r, so S( {pd) = r2 and , r-1 E[H(X)]- H(X) = - 2ln(2)n 1 + 6ln(2)n 2 (r 2 1 2 - 3r + 2) - 4n 3 (r -1) 1 + O(n 4 ). Theoretically S( {Pi}) has no upper bound when one of the probabilities is close to zero. However, in practice it is hard to collect a sample to estimate a very small probability. For this reason, we assume that Pi is either zero or greater than 6/r where 6 > 0 is a small constant does not depend on nor r . Under this assumption S( {pd) ::; r 2/6 and the amplitUde of the last term in (4) is less than (r2 /6 - 1) . 4!3 4 MI in Speech for Phonetic Classification The three hour telephone speech in the OGI database gives us a sample size greater than 1 million, n = 1050000. To estimate the mutual information between three features and a target variable, we need to estimate the entropy H(Xl' X 2, X3, Y). Take B = 20 as the number of bins for each feature variable and C = 19 is the number of phoneme categories. Then the total number of cells is r = B3 * C. After a constant adjustment, assuming 6 = 1, the bias is O( :2) = 6ln(12)n2 (r2 - 3r + 2) = 0.005(bits). It is shown in Fig. 1(a) that X(/4,t) and X(/5,t) are most relevant features for phonetic classification. From Fig. 1(b), at 5 Bark the MI spread around the current frame is 200 ms. Given one feature Xl, the information gain due to the second feature is the difference I(Xl,X2;Y)- I(Xl;Y) = I(X 2;YIXd where I(X2; YIXd is called the information gain of X 2 given Xl. It is a conditional mutual information defined by It is shown in Fig. 1(c)-(d) that given X(/5, t) across different bands the maximum information gain is achieved by X(/g, t), and within 5 Bark band the maximum information gain is achieved by X (/5 , t - 5). The mutual informations I(X(/4' t), X(/k, t + d); Y) for k 1, ... ,15, k ? 4, and d ?1, ... ,?1O, the information gain from the second feature in the vicinity of the first one, are shown in Fig. 2. The asymmetric distribution of the MI around the neighborhood (/5, d = 0) indicates that the phonetic information is spread asymmetrically through time but localized in about 200 ms around the current frame. = = Based on our data set, we have H(Y) = 3.96 (bits). The JMI for three frequency features and three temporal features are shown in Fig. 1{e)-{f). Based on these estimates, the three frequency features give 28% reduction in uncertainty about Y while the three temporal features give 19% reduction. Search for Information Bearing Components in Speech 807 0.55 0 .5 0.5 0.45 0.4 0.4 :II .Iii :II 0 . 35 0.3 .S; :iii! :iii! 0.3 0 .2 0 . 25 0 .1 0 .2 0,'5 6 0 earl< 9 12 0 -400 15 200 )~ 0 0.9 0 . 85 O .BS 0.8 0 .8 0 . 75 0 . 75 :II .'" :iii 0.7 0 .85 07 ~~ 0 .65 0.8 0 .6 0.55 0.5 0 .55 0 3 9 8 12 0 .5 -400 15 Bark -200 0 400 (d) 1.2 1 .2 1 .1 1 .1 :II .& :iii 0.9 200 time shift In rna (C) :II .Iii :iii 400 (b) (a) 0 .9 :II . S; :iii 200 tim_ shift In rna 0.9 0 .8 0 .8 0.7 o.71=--------..,. ---------------- o . eO~-~3--~B,.------:9:----:,:'::2---:',S? ea'" (e) ~~0~0~---:2~0~0--~0-----2~0~0~-~400 tl",. shirt In rna (f) Figure 1: (a) MIs of individual features in different bands. (b) MIs of individual feature at 5 Bark with different lOms-frame shifts. (c) JMIs of two features: at 5 Bark and in other bands. (d) JMIs of two features: current frame and shifted frames, both at 5 Bark. (e) JMIs of three features: at 5 Bark, 9 Bark and in other bands. The dashed line is th~ JMI level achieved by the two features X (15, t) and X (19, t). (f) JMIs of three features: current frame, 5th frame before current frame, and other shifted frames , all at 5 Bark. The dashed line is the JMI level achieved by X (15 , t) and X (15 , t - 5) . The size of our data set is n = 1050000. Therefore, we can reliably estimate the joint H H Yang and H Hermansky 808 MI between three features and the phoneme label. However, to estimate the JMI for more than 3 features we have the problem of curse of dimensionality since for k features, r = Bk * C is exponential increasing. For example, when k 4, B = 20, and C = 19, the second order bias is O(1/n2) = 2.02 (bits) which is too high to be ignored. To extend our approach beyond the current three-feature level, we need either to enlarge our data set or to find an alternative to the histogram based MI estimation. = . .. ... 0.95 0.9 0.85 ~ 0.8 .. ...; .Ii 0.75 ~ 0.7 ~0.65 0.6 0.55 0.5 15 100 o Bark -100 frame shill In ms Figure 2: The 3-D plot of joint mutual information around X(!4 , t). An asymmetric distribution is apparent especially around 4 Bark and 5 Bark. 1 .3 .---~-~--~-~----, 1.5 .----~--~--~----, 1.4 1 .2 1.3 1 .1 1 .2 1.1 ~ 0 .9 . S; 2 0 .8 0.9 0 .7 O.B I~ '" '" , 1.,t band 15-th band ,, , \ 0 .5 3 6 9 Bark (a) 12 15 ___ _ I \ I \ 0 .6 I ;' 0 .7 I I I I I 0 .6 ~ I, ~~O~0~--~'0~0---0~--'~00~-~ 200 time shift In rna (b) Figure 3: (a) The MI between one frequency feature and the file label. (b) The JMI between two features and the file identity labels. 5 MI in Speech for Speaker/Channel Recognition The linguistic variability expressed by phoneme labels is not the only variability present in speech. We use the mutual information to evaluate relevance to other Search for Information Bearing Components in Speech 809 sources of variabilities such as speaker/channel variability. Taking the file label as a target variable , we estimated the mutual information for one and two features. It is shown in Fig. 3(a) that the most relevant features are in the very low frequency channels, which in our case of telephone speech carry only very little speech information. Fig. 3(b) shows that the second most relevant feature for speaker/channel recognition is at least 150 ms apart from the first most relevant feature. These results suggest that the information about the speaker and the communication channel is not localized in time. These results are complementary to the results for phonetic classification shown in Fig. 1 (a) and (d) . 6 CONCLUSIONS Our results have shown that the information theoretic analysis of labeled speech data is feasible and useful for obtaining reusable knowledge about speech/channel variabilities. The joint mutual information of two features for phonetic classification is asymmetric around the current frame. We also estimated the joint mutual information between the phoneme labels and three feature variables . The uncertainty about the phonetic classification is reduced by adding more features. The maximum uncertainty reductions due to three frequency features and three temporal features are 28% and 19% respectively. The mutual informations of one and two features for speaker/channel recognition are estimated. The results show that the most relevant features are in the very low frequency bands. At 1 Bark and 5 Bark, the second most relevant temporal feature for speaker/channel recognition is at least 150 ms apart from the first most relevant feature. These results suggest that the information about the speaker and the communication channel is not localized in time. These results are complementary to the results for phonetic classification for which the mutual information is generally localized with some time spread. References [1] J. A. Bilmes. Maximum mutual information based reduction strategies for crosscorrelation based joint distribution modeling . In ICASSP98 , pages 469-472, April 1998. [2] R. Cole, M. Fanty, M. Noel, and T . Lander . Telephone speech corpus development at CSLU. In ICSLP, pages 1815-1818, Yokohama, Sept. 1994. [3] H. Hermansky. Perceptual linear predictive (PLP) analysis of speech. 1. A coust. Soc. Am., 87(4):1738-1752, April 1990 . [4] G. A. Miller. Note on the bias of information estimates. In H. Quastler, editor, Information Theory and Psychology , pages 95-100. The Free Press, Illinois, 1954. [5] Andrew Morris, Jean-Luc Schwartz , and Pierre Escudier . An information theoretical investigation into the distribution of phonetic information across the auditory spectogram. Computer Speech fj Language, 7(2):121-136, April 1993. [6] H. H. Yang, S. Van Vuuren, , S. Sharma, and H. Hermansky. Relevancy of timefrequency features for phonetic classification and speaker-channel recognition. Accepted by Speech Communication , 1999 . [7] H. H . Yang , S. Van Vuuren, and H. Hermansky. Relevancy of time-frequency features for phonetic classification measured by mutual information. In ICASSP99, pages 1:225-228 , Phoenix, March 1999. PART VII VISUAL PROCESSING
1655 |@word timefrequency:2 relevancy:6 carry:3 reduction:6 contains:1 xiy:2 interestingly:1 current:7 written:1 realistic:1 plot:1 vuuren:2 plane:2 short:1 quantized:1 manner:1 theoretically:1 nor:1 multi:1 shirt:1 globally:1 little:1 curse:1 window:2 increasing:1 underlying:1 bounded:1 temporal:5 schwartz:1 overestimate:1 before:1 engineering:1 approximately:1 collect:1 graduate:1 range:1 testing:1 practice:2 x3:1 word:1 suggest:2 close:1 context:1 py:1 regardless:2 duration:1 fluent:1 focused:1 identifying:1 insight:1 deriving:1 target:11 recognition:7 approximated:1 asymmetric:3 database:6 labeled:7 observed:1 electrical:1 environment:2 pd:5 depend:3 predictive:1 joint:9 neighborhood:1 apparent:1 jean:1 jointly:1 fanty:1 relevant:8 date:1 extending:1 andrew:1 measured:2 p2:1 soc:1 come:1 correct:1 centered:1 bin:1 icslp:1 investigation:1 correction:2 around:7 nw:1 estimation:2 label:19 cole:1 gaussian:2 yix:2 rna:4 rather:1 linguistic:4 derived:1 improvement:1 indicates:1 mainly:1 am:1 inference:1 typically:1 cslu:1 interested:1 classification:11 development:1 mutual:26 equal:1 asr:2 enlarge:1 represents:2 broad:1 hermansky:7 few:2 individual:2 vowel:2 message:4 nl:1 divide:1 old:1 theoretical:1 instance:1 modeling:1 subset:1 too:1 characterize:1 dependency:1 decoding:1 squared:1 containing:1 crosscorrelation:1 yp:1 coefficient:2 oregon:1 break:1 analyze:1 portion:1 ni:3 phonetically:1 phoneme:17 variance:1 miller:4 spaced:1 ofthe:1 identification:2 produced:2 bilmes:3 underestimate:1 energy:2 frequency:16 mi:16 hamming:1 gain:6 auditory:1 knowledge:2 dimensionality:1 amplitude:1 ea:1 appears:1 disposal:1 higher:1 rand:2 april:3 correlation:2 hand:1 french:1 reveal:1 escudier:1 b3:1 usa:1 concept:1 vicinity:1 ogi:5 plp:1 speaker:23 m:8 theoretic:2 fj:1 recently:1 log2p:1 phoenix:1 khz:1 jp:1 volume:1 million:1 extend:1 rd:1 automatic:1 fk:4 similarly:1 illinois:1 language:1 showed:1 apart:2 phonetic:17 transmitted:1 greater:2 eo:1 sharma:1 signal:7 ii:7 dashed:2 long:1 equally:1 prediction:1 involving:1 histogram:3 achieved:4 cell:4 lander:1 walker:1 source:2 meaningless:1 file:12 jmi:8 yang:5 iii:8 fft:1 psychology:1 shift:4 effort:1 speech:34 ignored:1 generally:2 useful:1 ylx:2 band:10 locally:1 morris:2 category:3 hynek:2 reduced:2 shifted:2 estimated:7 discrete:2 reusable:2 uncertainty:5 saying:1 bit:3 bound:1 icassp99:1 x2:2 span:1 relatively:2 department:1 march:1 lingual:1 across:3 b:1 ln:3 equation:1 needed:1 apply:1 yare:1 spectral:1 pierre:1 alternative:1 yokohama:1 include:1 beaverton:1 log2:2 especially:1 strategy:1 dependence:1 evenly:1 reason:1 assuming:1 relate:1 reliably:1 unknown:1 upper:1 howard:1 extended:1 communication:5 variability:7 frame:12 community:1 bk:1 acoustic:1 hour:2 beyond:1 including:1 critical:2 advanced:1 representing:2 technology:3 categorical:2 utterance:1 sept:1 fax:1 bark:15 relative:1 localized:4 earl:1 editor:1 story:3 classifying:1 pi:4 last:2 free:1 english:1 loms:1 bias:10 institute:1 taking:2 distributed:2 van:2 preprocessing:1 corpus:1 xi:4 search:4 continuous:1 quantifies:1 channel:19 learn:1 obtaining:1 plosive:1 bearing:4 spread:6 n2:4 complementary:3 fig:8 tl:1 exponential:1 xl:5 perceptual:2 weighting:1 ix:1 formula:4 down:1 r2:3 adding:1 magnitude:1 occurring:1 vii:1 entropy:3 logarithmic:1 visual:1 expressed:1 adjustment:1 hua:1 conditional:2 identity:2 targeted:2 noel:1 luc:1 feasible:1 hard:1 telephone:6 hyang:1 total:1 called:1 asymmetrically:1 ece:1 accepted:1 relevance:1 evaluate:1
717
1,656
Resonance in a Stochastic Neuron Model with Delayed Interaction Toru Ohira* Sony Computer Science Laboratory 3-14-13 Higashi-gotanda Shinagawa, Tokyo 141, Japan ohira@csl.sony.co.jp Yuzuru Sato Institute of Physics, Graduate School of Arts and Science, University of Tokyo 3-8-1 Komaba, Meguro, Tokyo 153 Japan ysato@sacral.c.u-tokyo.ac.jp Jack D. Cowan Department of Mathematics University of Chicago 5734 S. University, Chicago, IL 60637, U.S.A cowan@math.uchicago.edu Abstract We study here a simple stochastic single neuron model with delayed self-feedback capable of generating spike trains. Simulations show that its spike trains exhibit resonant behavior between "noise" and "delay". In order to gain insight into this resonance, we simplify the model and study a stochastic binary element whose transition probability depends on its state at a fixed interval in the past. With this simplified model we can analytically compute interspike interval histograms, and show how the resonance between noise and delay arises. The resonance is also observed when such elements are coupled through delayed interaction. 1 Introd uction "Noise" and "delay" are two elements which are associated with many natural and artificial systems and have been studied in diverse fields. Neural networks provide representative examples of information processing systems with noise and delay. Though much research has gone into the investigation of these two factors in the community, they have mostly been separately studied (see e.g. [1]). Neural * Affiliated also with the Laboratory for Information Synthesis, RIKEN Brain Science Institute, Wako, Saitama, Japan 315 Resonance in a Stochastic Neuron Model with Delayed Interaction models incorporating both noise and delay are more realistic [2], but their complex characteristics have yet to be explored both theoretically and numerically. The main theme of this paper is the study of a simple stochastic neural model with delayed interaction which can generate spike trains. The most striking feature of this model is that it can show a regular spike pattern with suitably "tuned" noise and delay [3]. Stochastic resonance in neural information processing has been investigated by others (see e.g. [4]). This model, however, introduces a different type of such resonance, via delay rather than through an external oscillatory signal. It can be classified with models of stochastic resonance without an external signal [5] . The novelty of this model is the use of delay as the source of its oscillatory dynamics. To gain insight into the resonance, we simplify the model and study a stochastic binary element whose transition probability depends on its state at a fixed interval in the past. With this model, we can analytically compute interspike interval histograms, and show how the resonance between noise and delay arises. We further show that the resonance also occurs when such stochastic binary elements are coupled through delayed interaction. 2 Single Delayed-feedback Stochastic Neuron Model Our model is described by the following equations: d Jl dt Vet) ?(V( t)) -Vet) + W?(V(t - r)) + eL(t) 2 -1 1 + e-1)(V(t)-9) (1) where 11 and () are constants, and V is the membrane potential of the neuron. The noise term eL has the following probability distribution. pee = u) 1 (-L 5: u 5: L) 2L o (u<-L,u>L) , (2) i.e., eL is a time uncorrelated uniformly distributed noise in the range (-L , L). It can be interpreted as a fluctuation that is much faster than the membrane relaxation time Jl. The model can be interpreted as a stochastic neuron model with delayed self-feedback of weight W, which is an extension of a model with no delay previously studied using the Fokker- Planck equation [6]. We numerically study the following discretized version: 2 Vet + 1) = 1 + e-1)(V(t-T ) - 9) - 1 + eL (3) We fix 11 and () so that this map has two basins of attractors of differing size with no delay, as shown in Figure l(A) . We have simulated the map (3) with various noise widths and delays and find regular spiking behavior as shown in Fig l(C) for tuned noise width and delay. In case the noise width is too large or too small given self-feedback delay, this rhythmic behavior does not emerge, as shown in Fig1(B) and (D). We argue that the delay changes the effective shape of the basin of attractors into an oscillatory one, just like that due to an external oscillating force which, as is well-known, leads to stochastic resonance with a tuned noise width. The analysis of the dynamics given by (1) or (3), however, is a non- trivial task, particularly with T. Ohira, Y. Sato and J. D. Cowan 316 respect to the spike trains. A previous analysis using the Fokker-Planck equation cannot capture this emergence of regular spiking behavior. This difficulty motivates us to further simplify our model, as described in the next section. (E) ,cp (A) I ... .~. (0) (b) ., ,.p (F) (8) X(t) V(t) .. , 200 ... m .oo 100 1000 t t L-- (G) (e) V(t) X(t) , , , , .. .. ,>--- I - I - L.. ~ , " I - L.. ' - - (8) (D) V(t) X(t) Figure 1: (A) The shape of the sigmoid function 4> (b) for", = 4 and 0 = 0.1. The straight line (a) is 4> = V and the crossings of the two lines indicate the stationary point of the dynamics. Also, the typical dynamics of V (t) from the map model are shown as we change noise width L. The values of L are (B) L = 0.2, (C) L = 0.4, (D) L = 0.7. The data is taken with T = 20, '" = 4.0, 0 = 0.1 and the initial condition V(t) = 0.0 for t E [-r,O]. The plots are shown between t = a to 1000. (E) Schematic view of the single binary model. Some typical dynamics from the binary model are also shown. The values of parameters are r = 10, q = 0.5, and (F) p = 0.005, (G) p = 0.05, and (H) p = 0.2. 3 Delayed Stochastic Binary N enron Model The model we now discuss is an approximation of the dynamics that retains the asymmetric stochastic transition and delay. The state X(t) of the system at time step t is either -lor 1. With the same noise the model is described as follows: X(t + 1) O[f(X(t + 1 f(n) = 2?a + b) + n(a - b?, eL, T? eLl, O[n] 1 (0 ~ n), -1 (0) n), (4) where a and b are parameters such that lal ~ L and Ibl ~ L, and r is the delay. This model is an approximate discretization of the space of map (3) into two states Resonance in a Stochastic Neuron Model with Delayed Interaction 317 with a and b controlling the bias of transition depending on the state of X r steps earlier. When a i- b, the transition between the two states is asymmetric, reflecting the two differing sized basins of attractors. We can describe this model more concisely in probability space (Figure I(E)). The formal definition is given as follows: P(I, t + 1) P(-I,t+l) p q p, X(t 1- q, X(t q, X(t 1- p, X(t 1 b 2(1 + L)' 1 a 2(1 - L)' r) = -1, - r) = 1, r) = 1, - r) = -1, (5) where P(s, t) is the probability that X(t) = s. Hence, the transition probability of the model depends on its state r steps in the past, and is a special case of a delayed random walk [7]. We randomly generate X(t) for the interval t = (-r, 0) . Simulations are performed in which parameters are varied and X(t) is recorded for up to 10 6 steps. They appear to be qualitatively similar to those generated by the map dynamics (Figure I(F),(G),(H)). ;,From the trajectory X(t), we construct a residence time histogram h( u) for the system to be in the state -1 for u consecutive steps. Some examples of the histograms are shown in Figure 2 (q = 1 - q = 0.5, r = 10). We note that with p ? 0.5, as in Figure 2(A), the model has a tendency to switch or spike to the X = 1 state after the time step interval of r. But the spike trains do not last long and result in a small peak in the histogram. For the case of Figure 2(C) where p is closer to 0.5, we observe less regular transitions and the peak height is again small. With appropriate p as in Figure 2(B), spikes tend to appear at the interval T more frequently, resulting in higher peaks in the histogram. This is what we mean by stochastic resonance (Figure 2(D)). Choosing an appropriate p is equivalent to "tuning" the noise width L, with other parameters appropriately fixed. In this sense, our model exhibits stochastic resonance. This model can be treated analytically. The first observation to make with the model is that given r, it consists of statistically independent r + 1 Markov chains. Each Markov chain has its state appearing at every r+l interval. With this property of the model, we label time step t by the two integers sand k as follows t=s(r+l)+k, (O::;s,O::;k::;r) (6) Let P?(t) == P?(s, k) be the probability for the state to be in the ?1 state at time t or (s, k). Then, it can be shown that P+(s, k) P_ (s, k) a {3 , ,s + ,s a(1 - ,S) + {3(1- ,S) p , p+q q --, p+q 1 - (p + q). P+(s = 0, k), P_(s = 0, k), (7) In the steady state, P+(s --+ oo,k) == P+ = a and P_(s --+ oo,k) == P_ = {3. The steady state residence time histogram can be obtained by computing the following T. Ohira, Y. Safo and J. D. Cowan 318 = quantity, h(u) P(+;-,Uj+), which is the probability that the system takes consecutive -1 states U times between two +1 states. With the definition of the model and statistical independence between Markov chains in the sequence, the following expression can be derived: P(+;-,Uj+) P+(P_)Up+ = (,8)U(a)2 (1 ~ U < r) (8) = p+(p-r(1- q) = (,8r(a)(1- q) (u = r) (9) = p+(p-r(q)(1- p)U-T(p) = (,8)U(p)2 (u > r) (10) With appropriate normalization, this expression reflects the shape of the histogram obtained by numerical simulations, as shown in Figure 2. Also, by differentiating equation (9) with respect to p, we derive the resonant condition for the peak to reach maximum height as (11) q=pr or, equivalently, L - a = (L + b)r. (12) In Figure 2(D), we see that maximum peak amplitude is reached by choosing parameters according to equation (11). We note that this analysis for the histogram is exact in the stationary limit, which makes this model unique among those showing stochastic resonance. (M "':L b(a) ~ Ol~ ." 'M. j ~ ~ ~ ~ l~ I~ ~ I'!. I '!.:O ? (at "...': I~ ." hCD) . . Ol~ . ~. 1 ~ , " 10 12 ~ I' 17 ~ jO ? (et ~ ..."'~ (I Ol~ b(ot .." ,', !> ~ "L I" us ? )0 UP!> 20 ? ", h(tt . "" ." .... ~. 10 10 to "'a ...." Figure 2: Residence time histogram and dynamics of X(t) as we change p. The 0.005, (B) p = 0.05, (C) p = 0.2. The solid line in the values of p are (A) p histogram is from the analytical expression given in equations (8-10). Also, in (D) we show a plot of peak height by varying p. The solid line is from equation (9). The parameters are r = 10, q = 0.5. = 4 Delay Coupled Two Neuron Case We now consider a circuit comprising two such stochastic binary neurons coupled with delayed interaction. We observe again that resonance between noise and delay Resonance in a Stochastic Neuron Model with Delayed Interaction 319 takes place. The coupled two neuron model is a simple extension of the model in the previous section. The transition probability of each neuron is dependent on the other neuron's state at a fixed interval in the past. Formally, it can be described in probability space as follows. Pl(l, t + 1) X 2(t - 72) = -1, X 2(t - 72) = 1, X 2(t - 72) = 1, ql, 1 - PI, X 2(t - 72) = -1, Xl(t - 7d = -1, P2, 1- q2, Xl(t - 7d = 1, q2, Xl(t - 71) = 1, 1- P2, Xl(t - 71) = -1 PI! 1- q!, Pl(-I,t+l) P 2(1, t + 1) P2(-I,t+l) (13) Pi(S, t) is the probability that the state of the neuron i is Xi(t) = s. We have performed simulation experiments on the model and have again found resonance between noise and delay. Though more intricate than the single neuron model, we can perform a similar theoretical analysis of the histograms and have obtained approximate results for some cases. For example, we obtain the following approximate analytical results for the peak height of the interspike histogram of Xl for the case 71 = 72 == 7. ( The peak occurs at 71 + 72 + 1.) H(Pl' P2, qI, q2) J.tl (PI, P2, qI, q2) J.t2 (PI, P2, ql, q2) J.t3 (PI, P2, ql , q2 ) J.t4 (PI! P2, qI, q2 ) II (PbP2, ql, q2) 12(PI,P2, qI, q2) S(PI,P2, ql, q2) = {J.t3(P!, P2, ql, q2 )ql + J.t4(PI, P2, ql, q2)(1 - pd Y (14) {J.tl(P!'P2, ql, q2)(qlq2PI + ql(l - q2)(1 - ql)) +J.t2(Pl,P2,ql,q2)((I- pdq2Pl + (1- Pl)(I- q2)(I- qd)} h (PI, P2, ql, q2)!2(PI, P2, ql, q2) (15) S(PI , P2 , ql , q2) II (PI ,P2, qI, q2) (16) S(PI,P2, q!, q2) !2(PI,P2, ql, q2) (17) s (PI! P2, ql, q2) 1 (18) S(PI,P2, ql, q2) PI(I - P2) + P2(1 - ql) (19) ql (1 - q2) + q2 (1 - qd P2 + PI (1 - P2 - q2) (20) q2 + ql(l - P2 -q2) h (PI, P2, ql, q2)!2(PI ,P2, ql, q2) (21) + h(PI,P2, ql, q2) + !2(PI,P2, ql, q2) + 1 These analytical results are compared with the simulation experiments, examples of which are shown in Figure 3. A detailed analysis, particularly for the case of 71 =I 72, is quite intricate and is left for the future. 5 Discussion There are two points to be noted. Firstly, although there are examples which may indicate that stochastic resonance is utilized in biological information processing, it is yet to be explored if the resonance between noise and delay has some role in T. Ohira, Y. Sato and J. D. Cowan 320 (A) (C) (8) 0 . 01 h(t) . .. h(t) ::::~ 0 . 008 ~ ..00< o.oo:z 0 . 00:3: 0 .2 pi . 0.3 0.1 0 ,' .. 0 .1 0.' 0 .5 Figure 3: A plot of peak height by varying P2. The solid line is from equation (1420). The parameters are T1 = T2 = 10, q1 = q2 = 0.5, (A)P1 = P2, (B) P1 = 0.005, (C) P1 = 0.025. neural information processing. Secondly, there are many investigations of spiking neural models and their applications (see e.g., [8]). Our model can be considered as a new mechanism for generating controlled stochastic spike trains. One can predict its application to weak signal transmission analogous to recent research using stochastic resonance with a larger number of units in series [9]. Investigations of the network model with delayed interactions are currently underway. References [1) Hertz, J. A., Krogh, A., & Palmer, R . G. (1991). Introduction to the Theory of Neural Computation. Redwood City: Addison-Wesley. [2) Foss, J., Longtin, A., Mensour, B., & Milton, J . G. (1996). Multistability and Delayed Recurrent Loops. Physical Review Letters, 76, 708-711; Pham, J., Pakdaman, K., Vibert, J.-F. (1998). Noise-induced coherent oscillations in randomly connected neural networks. Physical Review E, 58, 3610-3622; Kim, S., Park, S. H., Pyo, H.-B. (1999). Stochastic Resonance in Coupled Oscillator Systems with Time Delay. Physical Review Letters, 8!, 1620-1623; Bressloff, P. C. (1999). Synaptically Generated Wave Propagation in Excitable Neural Media. Physical Review Letters, 8!, 2979-2982. [3) Ohira, T. & Sato, Y. (1999). Resonance with Noise and Delay. Physical Review Letters, 8!, 2811-2815. [4) Gammaitoni, L., Hii.nggi, P., Jung, P., & Marchesoni, F.(1998). Stochastic Resonance. Review of Modem Physics, 70, 223-287. [5) Gang, H., Ditzinger, T., Ning, C. Z., & Haken, H.(1993) Stochastic Resonance without External Periodic Force. Physical Review Letters, 71, 807-810; Rappel, W-J. & Strogatz, S. H. (1994). Stochastic resonance in an autonomous system with a nonuniform limit cycle. Physical Review E, 50,3249-3250; Longtin, A. (1997). Autonomous stochastic resonance in bursting neurons. Physical Review E, 55, 868-876. [6) Ohira, T. & Cowan J . D. (1995). Stochastic Single Neurons, Neural Communication, 7518-528. [7) Ohira, T. & Milton, J. G. (1995) . Delayed Random Walks. Physical Review E, 5!, 3277-3280; Ohira, T. (1997). Oscillatory Correlation of Delayed Random Walks, Physical Review E, 55, RI255-1258. [8) Maas, W. (1997). Fast Sigmoidal Network via Spiking Neurons. Neural Computation, 9(2), 279-304; Maas, W. (1996). Lower Bounds for the Computational Power of Networks of Spiking Neurons. Neural Computation, 8(1), 1-40. [9) Locher, M., Cigna, D., and Hunt, E. R. (1998). Noise Sustained Propagation of a Signal in Coupled Bistable Electric Elements Physical Review Letters, 80, 5212-5215.
1656 |@word version:1 suitably:1 simulation:5 q1:1 solid:3 initial:1 series:1 tuned:3 wako:1 past:4 discretization:1 yet:2 realistic:1 numerical:1 chicago:2 interspike:3 shape:3 plot:3 stationary:2 math:1 firstly:1 sigmoidal:1 lor:1 height:5 consists:1 sustained:1 theoretically:1 intricate:2 behavior:4 p1:3 frequently:1 brain:1 discretized:1 ol:3 csl:1 circuit:1 medium:1 what:1 interpreted:2 q2:34 differing:2 every:1 unit:1 appear:2 planck:2 t1:1 limit:2 fluctuation:1 studied:3 bursting:1 co:1 hunt:1 p_:5 graduate:1 gone:1 range:1 statistically:1 palmer:1 unique:1 regular:4 cannot:1 equivalent:1 map:5 insight:2 gammaitoni:1 autonomous:2 analogous:1 controlling:1 exact:1 crossing:1 element:6 particularly:2 utilized:1 asymmetric:2 observed:1 role:1 higashi:1 capture:1 connected:1 cycle:1 pd:1 dynamic:8 various:1 riken:1 train:6 fast:1 effective:1 describe:1 artificial:1 choosing:2 whose:2 quite:1 larger:1 emergence:1 sequence:1 analytical:3 interaction:9 loop:1 meguro:1 transmission:1 oscillating:1 generating:2 oo:4 depending:1 ac:1 derive:1 recurrent:1 school:1 krogh:1 p2:33 indicate:2 qd:2 ning:1 tokyo:4 stochastic:29 bistable:1 sand:1 fix:1 investigation:3 biological:1 secondly:1 extension:2 pl:5 pham:1 considered:1 predict:1 consecutive:2 pee:1 label:1 currently:1 city:1 reflects:1 rather:1 varying:2 fig1:1 derived:1 rappel:1 ibl:1 kim:1 sense:1 dependent:1 el:5 comprising:1 among:1 resonance:28 art:1 special:1 ell:1 field:1 construct:1 park:1 future:1 others:1 t2:3 komaba:1 simplify:3 randomly:2 delayed:17 attractor:3 introduces:1 chain:3 capable:1 closer:1 walk:3 theoretical:1 earlier:1 retains:1 saitama:1 delay:23 too:2 periodic:1 peak:9 physic:2 synthesis:1 jo:1 again:3 recorded:1 external:4 japan:3 potential:1 depends:3 performed:2 view:1 reached:1 wave:1 il:1 characteristic:1 t3:2 weak:1 trajectory:1 straight:1 classified:1 oscillatory:4 reach:1 fo:1 definition:2 associated:1 gain:2 amplitude:1 reflecting:1 wesley:1 higher:1 dt:1 though:2 just:1 correlation:1 propagation:2 analytically:3 hence:1 laboratory:2 self:3 width:6 noted:1 steady:2 tt:1 haken:1 cp:1 jack:1 sigmoid:1 spiking:5 physical:11 jp:2 jl:2 numerically:2 tuning:1 mathematics:1 gotanda:1 recent:1 binary:7 novelty:1 signal:4 ii:2 faster:1 long:1 controlled:1 schematic:1 qi:5 longtin:2 histogram:13 normalization:1 synaptically:1 separately:1 interval:9 source:1 appropriately:1 ot:1 enron:1 induced:1 tend:1 cowan:6 integer:1 ohira:9 switch:1 independence:1 expression:3 introd:1 detailed:1 generate:2 diverse:1 shinagawa:1 relaxation:1 letter:6 striking:1 place:1 resonant:2 residence:3 oscillation:1 bound:1 vibert:1 sato:4 gang:1 department:1 according:1 membrane:2 hertz:1 pr:1 taken:1 equation:8 previously:1 discus:1 mechanism:1 addison:1 sony:2 milton:2 multistability:1 observe:2 appropriate:3 hii:1 appearing:1 uj:2 quantity:1 spike:9 occurs:2 exhibit:2 simulated:1 argue:1 trivial:1 equivalently:1 ql:25 mostly:1 affiliated:1 motivates:1 perform:1 neuron:19 observation:1 markov:3 modem:1 communication:1 varied:1 redwood:1 nonuniform:1 community:1 lal:1 concisely:1 coherent:1 pattern:1 power:1 difficulty:1 natural:1 force:2 treated:1 excitable:1 coupled:7 review:12 underway:1 basin:3 uncorrelated:1 pi:25 maas:2 jung:1 last:1 bias:1 uchicago:1 formal:1 institute:2 bressloff:1 emerge:1 rhythmic:1 differentiating:1 distributed:1 feedback:4 transition:8 qualitatively:1 simplified:1 yuzuru:1 approximate:3 xi:1 vet:3 investigated:1 complex:1 electric:1 main:1 noise:22 fig:1 representative:1 tl:2 theme:1 xl:5 showing:1 explored:2 incorporating:1 uction:1 t4:2 strogatz:1 toru:1 fokker:2 sized:1 oscillator:1 change:3 typical:2 uniformly:1 tendency:1 formally:1 arises:2
718
1,657
Information Capacity and Robustness of Stochastic Neuron Models Elad Schneidman Idan Segev N aftali Tishby Institute of Computer Science, Department of Neurobiology and Center for Neural Computation, Hebrew University Jerusalem 91904, Israel { elads, tishby} @cs.huji.ac.il, idan@lobster.ls.huji.ac.il Abstract The reliability and accuracy of spike trains have been shown to depend on the nature of the stimulus that the neuron encodes. Adding ion channel stochasticity to neuronal models results in a macroscopic behavior that replicates the input-dependent reliability and precision of real neurons. We calculate the amount of information that an ion channel based stochastic Hodgkin-Huxley (HH) neuron model can encode about a wide set of stimuli. We show that both the information rate and the information per spike of the stochastic model are similar to the values reported experimentally. Moreover, the amount of information that the neuron encodes is correlated with the amplitude of fluctuations in the input, and less so with the average firing rate of the neuron. We also show that for the HH ion channel density, the information capacity is robust to changes in the density of ion channels in the membrane, whereas changing the ratio between the Na+ and K+ ion channels has a considerable effect on the information that the neuron can encode. Finally, we suggest that neurons may maximize their information capacity by appropriately balancing the density of the different ion channels that underlie neuronal excitability. 1 Introduction The capacity of neurons to encode information is directly connected to the nature of spike trains as a code. Namely, whether the fine temporal structure of the spik~ train carries information or whether the fine structure of the train is mainly noise (see e.g. [1, 2]). Experimental studies show that neurons in vitro [3, 4] and in vivo [5, 6, 7], respond to fluctuating inputs with repeatable and accurate spike trains, whereas slowly varying inputs result in lower repeatability and 'jitter' in the spike timing. Hence, it seems that the nature of the code utilized by the neuron depends on the input that it encodes [3, 6]. Recently, we suggested that the biophysical origin of this behavior is the stochas- Capacity and Robustness oJStochastic Neuron Models 179 ticity of single ion channels. Replacing the average conductance dynamics in the Hodgkin-Huxley (HH) model [8], with a stochastic channel population dynamics [9, 10, 11], yields a stochastic neuron model which replicates rather well the spike trains' reliability and precision of real neurons [12]. The stochastic model also shows subthreshold oscillations, spontaneous and missing spikes, all observed experimentally. Direct measurement of membranal noise has also been replicated successfully by such stochastic models [13]. Neurons use many tens of thousands of ion channels to encode the synaptic current that reaches the soma into trains of spikes [14]. The number of ion channels that underlies the spike generation mechanism, and their types, depend on the activity of the neuron [15, 16]. It is yet unclear how such changes may affect the amount and nature of the information that neurons encode. Here we ask what is the information encoding capacity of the stochastic HH model neuron and how does this capacity depend on the densities of different of ion channel types in the membrane. We show that both the information rate and the information per spike of the stochastic HH model are similar to the values reported experimentally and that neurons encode more information about highly fluctuating inputs. The information encoding capacity is rather robust to changes in the channel densities of the HH model. Interestingly, we show that there is an optimal channel population size, around the natural channel density of the HH model. The encoding capacity is rather sensitive to changes in the distribution of channel types, suggesting that changes in the population ratios and adaptation through channel inactivation may change the information content of neurons. 2 The Stochastic HH Model The stochastic HH (SHH) model expands the classic HH model [8], by incorporating the stochastic nature of single ion channels [9, 17]. Specifically, the membrane voltage dynamics is given by the HH description, namely, dV cmTt = -gLCV - VL) - gK(V, t)(V - VK) - gNa(V, t)(V - VNa) +I (1) where V is the membrane potential, VL, VK and VNa are the reversal potentials of the leakage, potassium and sodium currents, respectively, gL, gK(V, t) and gNa(V, t) are the corresponding ion conductances, Cm is the membrane capacitance and I is the injected current. The ion channel stochasticity is introduced by replacing the equations describing the ion channel conductances with explicit voltage-dependent Markovian kinetic models for single ion channels [9, 10]. Based on the activation and inactivation variables of the deterministic HH model, each K+ channel can be in one of five different states, and the rates for transition between these states are given in the following diagram, [~ ] 4Qn ~ f3n [ ~ ] 3a n ~ 2f3n [ ~ ] 2a n ~ 3f3n [ ~ ] an ~ 4f3n [ ~ ] (2) where [nj] refers to the number of channels which are currently in the state nj. Here [n4] labels the single open state of a potassium channel, and an, i3n, are the voltage-dependent rate-functions in the HH formalism. A similar_model is used for the Na+ channel (The Na+ kinetic model has 8 states, with only one open state, see [12] for details). The potassium and sodium membrane conductances are given by, (3) gK(V, t) = ,K [114] gNa(V, t) = ,Na [mahl] where and ,Na are the conductances of an ion channel for the K+ and Na+ respectively. We take the conductance of a single channel to be 20pS [14] for both the ,K E. Schneidman. l. Segev and N. Tishby 180 K+ and Na+ channel types 1. Each of the ion channels will thus respond stochastically by closing or opening its 'gates' according to the kinetic model, fluctuating around the average expected behavior. Figure 1 demonstrates the effect of the ion A B Figure 1: Reliability of firing patterns in a model of an isopotential Hodgkin-Huxley membrane patch in response to different current inputs. (A) Injecting a slowly changing current input (low-pass Gaussian white noise with a mean TJ = 8I1A/cm 2 , and standard deviation a = 1 p,A/ cm 2 which was convolved with an 'alpha-function' with a time constant To = 3 msec, top frame), results in high 'jitter' in the timing of the spikes (raster plots of spike responses, bottom frame). (B) The same patch was again stimulated repeatedly, with a highly fluctuating stimulus (TJ = 8 p,A/cm 2 , a = 7 p,A/cm 2 and To = 3 msec, top frame) The 'jitter' in spike timing is significantly smaller in B than in A (i.e. increased reliability for the fluctuating current input). Patch area used was 200 p,m 2 , with 3,600 K+ channels and 12,000 Na+ channels. (Compare to Fig.l in see [3]). (C) Average firing rate in response to DC current input of both the HH and the stochastic HH model. (D) Coefficient of variation of the inter spike interval of the SHH model in response to DC inputs, giving values which are comparable to those observed in real neurons channel stochasticity, showing the response of a 200 J.Lm 2 SHH isopotential membrane patch (with the 'standard' SHH channel densities) to repeated presentation of suprathreshold current input. When the same slowly varying input is repeatedly presented (Fig. lA) , the spike trains are very different from each other, i.e. , spike firing time is unreliable. On the other hand, when the input is highly fluctuating (Fig. IB), the reliability of the spike timing is relatively high. The stochastic model thus replicates the input-dependent reliability and precision of spike trains observed in pyramidal cortical neurons [3] . As for cortical neurons, the Repeatability and Precision of the spike trains of the stochastic model (defined in [3]) are strongly correlated with the fluctuations in the current input and may get to sub-millisecond precision [12]. The f-I curve of the stochastic model (Fig. lC) and the coefficient of variation (CV) of the inter-spike intervals (lSI) distribution for DC inputs (Fig. ID) are both similar to the behavior of cortical neurons in vivo [18], in clear contrast to the deterministic model 2 1 The number of channels is thus the ratio between the total conductance of a single type of ion channels and the single channel conductance, and so the 'standard' SHH densities will be 60 Na+ and 18 Na+ channels per p,m 2 . 2 Although the total number of channels in the model is very large, the microscopic level ion channel noise has a macroscopic effect on the spike train reliability, since the number Capacity and Robustness ofStochastic Neuron Models 3 181 The Information Capacity of the SHH Neuron Expanding the Repeatability and Precision measures [3], we turn to quantify how much information the neuron model encodes about the stimuli it receives. We thus present the model with a set of 'representative' input current traces, and the amount of information that the respective spike trains encode is calculated. Following Mainen and Sejnowski [3], we use a set of input current traces which imitate the synaptic current that reaches the soma from the dendritic tree. We convolve a Gaussian white noise trace (with a mean current 1} and standard deviation 0') with an alpha function (with a To: = 3 msec). Six different mean current values are used (1} = 0,2,4,6,8,10 pA/cm 2 ) , and five different std values (0' = 1,3,5,7, 9pA/cm 2 ), yielding a set of 30 input current traces (each is 10 seconds long). This set of inputs is representative of the wide variety of current traces that neurons might encounter under in vivo conditions in the sense that the average firing rates for this set of inputs which range between 2 - 70 Hz (not shown). We present these input traces to the model, and calculate the amount of information that the resulting spike trains convey about each input, following [6, 19]. Each input is presented repeatedly and the resulting spike trains are discretized in D..T bins, using a sliding 'window' of size T along the discretized sequence. Each train of spikes is thus transformed into a sequence of K-letter 'words' (K = T/D..T) , consisting of O's (no spike) and l's (spike). We estimate P(W), the probability of the word W to appear in the spike trains, and then compute the entropy rate of its total word distribution, Htotal = - L P(W) log2 P(W) bits/word (4) W which measures the capacity of information that the neuron spike trains hold [20, 6, 19]. We then examine the set of words that the neuron model used at a particular time t over all the repeated presentations of the stimulus, and estimate P(Wlt), the time-dependent word probability distribution. At each time t we calculate the time-dependent entropy rate, and then take the average of these entropies Hnoise = (- LP(Wlt)lOg2 P(Wlt))t w bits/word (5) where ( .. .)t denotes the average over all times t. Hnoise is the noise entropy rate, which measures how much of the fine structure of the spike trains of the neuron is just noise. After performing the calculation for each of the inputs, using different word sizes 3, we estimate the limit of the total entropy and noise entropy rates at T --* 00, where the entropies converge to their real values (see [19] for details) . Figure 2A shows the total entropy rate of the responses to the set of stimuli, ranging from 10 to 170 bits/sec. The total entropy rate is correlated with the firing rates of the neuron (not shown). The noise entropy rate however, depends in a different way on the input parameters: Figure 2B shows the noise entropy rate of the responses to the set of stimuli, which may get up to 100 bits/sec. Specifically, for inputs with high mean current values and low fluctuation amplitude, many of the spikes are of ion channels which are open near the spike firing threshold is rather small [12). The fluctuations in this small number of open channels near firing threshold give rise to the input-dependent reliability of the spike timing. 3 t he bin size T = 2 msec has been set to be small enough to keep the fine temporal structure of the spike train within the word sizes used, yet large enough to avoid undersampling problems 182 E. Schneidman, /. Segev and N. Tishby just noise, even if the mean firing rate is high. The difference between the neuron's entropy rate (the total capacity of information of the neuron's spike train) and the noise entropy rate, is exactly the average rate of information that the neuron's spike trains encode about the input , I(stimulus , spike train) = Htotal - Hnoise [20, 6], this is shown in Figure 2C. The information rate is more sensitive to the size of A .~: 200 B ;f. ':'~ 150 150 100 100 50 50 10 0 0 10 0 a fl.LA/cm 2 ) '1 ,. 100 :~ 60 40 20 5 a lllAIcm2 ) 0 0 a 0 .~ ~ ~ i 0 a fl.LA/cm 2 ) D ~;) 2 ~ )~r . 1 0 10 3 2 .5 3 ~ 0 0 fl.LA/cm 2 ) ;;; 80 10 200 2 . 1.5 10 5 '1 [llAIcm 2 ) 0 .5 5 0 0 a lllAIcm 2 ) Figure 2: Information capacity of the SHH model. (A) The total spike train entropy rate of the SHH model as a function of 'TI, the current input mean, and a, the standard deviation (see text for details). Error bar values of this surface as well as for the other frames range between 1 - 6% (not shown). (B) Noise entropy rate as a function of the current input parameters. (C) The information rate about the stimulus in the spike trains, as a function of the input parameters, calculated by subtracting noise entropy from the total entropy (note the change in grayscale in C and D). (D) Information per spike as a function of the input parameters, which is calculated by normalizing the results shown in C by the average firing rate of the responses to each of the inputs. fluctuations in the input than to the mean value of the current trace (as expected, from the reliability and precision of spike timing observed in vitro [3] and in vivo [6] as well as in simulations [12]). The dependence of the neural code on the input parameters is better reflected when calculating the average amount of information per spike that the model gives for each of the inputs (Fig. 2D) (see for comparison the values for the Fly's HI neuron [6]). 4 The effect of Changing the Neuron Parameters on the Information Capacity Increasing the density of ion channels in the membrane compared to the 'standard' SHH densities, while keeping the ratio between the K+ and Na+ channels fixed, only diminishes the amount of information that the neuron encodes about any of the inputs in the set. However, the change is rather small: Doubling the channel density decreases the amount of information by 5 - 25% (Fig. 3A), depending on the specific input. Decreasing the channel densities of both types, results in encoding more information about certain stimuli and less about others. Figure 3B shows that having half the channel densities would result with in 10% changes in the information in both directions. Thus, the information rates conveyed by the stochastic model are robust to changes in the ion channel density. Similar robustness (not shown) has been observed for changes in the membrane area (keeping channel Capacity and Robustness ofStochastic Neuron Models 183 density fixed) and in the temperature (which effects the channel kinetics). However, A .... ., 1.2 B 1.2 jl.2 1 .1 .21.2 :5 1 1 ~ 0.9 ..s0.8 10 "1 '~o 1.s 10 5 o c 1.1 ~ 0 0 .9 10 O.S 10 5 a[~em2) '1 [pAIem2) $A O.S 5 o 0 a[~Alem2) 0.4 D >~ 3 .5 0.3 3 2.5 0.2 2 1.5 0 .1 10 0 .5 o 0 a[~Alem2) 0 Figure 3: The effect of changing the ion channel densities on the information capacity. (A) The ratio of the information rate of the SHH model with twice the density of the 'standard' SHH densities divided by the information rate of the mode with 'standard' SHH densities. (B) As in A, only for the SHH model with half the 'standard' densities. (C) The ratio of the info rate of the SHH model with twice as many Na+ channels, divided by the info rate of the standard SHH Na+ channel density, where the K+ channel density remains untouched (note the change in graycale in C and D). (D) As in C, only for the SHH model with the number of Na+ channels reduced by half. changing the density of the Na+ channels alone has a larger impact on the amount of information that the neuron conveys about the stimuli. Increasing Na+ channel density by a factor of two results in less information about most of the stimuli, and a gain in a few others (Fig. 3C). However, reducing the number of Na+ channels by half results in drastic loss of information for all of the inputs (Fig. 3D). 5 Discussion We have shown that the amount of information that the stochastic HH model encodes about its current input is highly correlated with the amplitude of fluctuations in the input and less so with the mean value of the input. The stochastic HH model, which incorporates ion channel noise, closely replicates the input-dependent reliability and precision of spike trains observed in cortical neurons. The information rates and information per spike are also similar to those of real neurons. As in other biological systems (e.g., [21]), we demonstrate robustness of macroscopic performance to changes in the cellular properties - the information coding rates of the SHH model are robust to changes in the ion channels densities as well as in the area of the excitable membrane patch and in the temperature (kinetics) of the channel dynamics. However, the information coding rates are rather sensitive to changes in the ratio between the densities of different ion channel types, suggests that the ratio between the density of the K+ channels and the Na+ channels in the 'standard' SHH model may be optimal in terms of the information capacity. This may have important implications on the nature of the neural code under adaptation and learning. We suggest that these notions of optimality and robustness may be a key biophysical principle of the operation of real neurons. Further investigations should take into account the activity-dependent nature of the channels and the 184 E. Schneidman, I. Segev and N. nshby neuron [15, 16] and the notion of local learning rules which could modify neuronal and suggest loca1learning rules as in [22]. Acknowledgements This research was supported by a grant from the Ministry of Science, Israel. References [1] Rieke F ., Warland D., de Ruyter van Steveninck R., and Bialek W. Spike: Exploring the Neural Code. MIT Press, 1997. [2] Shadlen M. and Newsome W. Noise, neural codes and cortical organization. Curro Opin. Neurobiol., 4:569-579, 1994. [3] Mainen Z. and Sejnowski T . Reliability of spike timing in neocortical neurons. Science, 268:1503-1508, 1995. [4] Nowak L., Sanches-Vives M., and McCormick D. Influence of low and high frequency inputs on spike timing in visual cortical neurons. Cerebral Cortex, 7:487-501, 1997. [5] Bair W. and Koch C. Temporal precision of spike trains in extrastriate cortex of the behaving macaque monkey. Neural Comp., 8:1185-1202, 1996. [6] de Ruyter van Steveninck R ., Lewen G., Strong S., Koberle R., and Bialek W. Reproducibility and variability in neural spike trains. Science, 275:1805-1808, 1997. [7] Reich D., Victor J., Knight B., Ozaki T., and Kaplan E. Response variability and timing precision of neuronal spike trains in vivo. J. Neurophysiol., 77:2836:2841, 1997. [8] Hodgkin A. and Huxley A. A quantitative description of membrane current and its application to conduction and excitation in nerve. J. Physiol., 117:500-544, 1952. [9] Fitzhugh R . A kinetic model of the conductance changes in nerve membrane. J. Cell. Comp o Physiol., 66:111-118, 1965. [10] DeFelice L. Introduction to Membrane Noise. Perseus Books, 1981. [11] Skaugen E. and Wall0e L. Firing behavior in a stochastic nerve membrane model based upon the Hodgkin-Huxley equations. Acta Physiol. Scand., 107:343-363, 1979. [12] Schneidman E ., Freedman B., and Segev I. Ion channel stochasticity may be critical in determining the reliability and precision of spike timing. Neural Comp., 10:1679-1704, 1998. [13] White J., Klink R., Alonso A., and Kay A. Noise from voltage-gated channels may influence neuronal dynamics in the entorhinal cortex. J Neurophysiol, 80:262-9, 1998. [14] Hille B. Ionic Channels of Excitable Membrane. Sinauer Associates, 2nd ed., 1992. [15] Marder E., Abbott L., Turrigiano G., Liu Z., and Golowasch J . Memory from the dynamics of intrinsic membrane currents. Proc. Natl. Acad. Sci., 93:13481-6, 1996. [16] Toib A., Lyakhov V., and Marom S. Interaction between duration of activity and rate of recovery from slow inactivation in mammalian brain Na+ channels. J Neurosci., 18:1893-1903, 1998. [17] Strassberg A. and DeFelice L. Limits of the HH formalism: Effects of single channel kinetics on transmembrane voltage dynamics. Neural Comp., 5:843-856, 1993. [18] Softky W. and Koch C . The highly irregular firing of cortical cells is inconsistent with temporal integration of random EPSPs. J. Neurosci ., 13:334-350, 1993. [19] Strong S., Koberle R., de Ruyter van Steveninck R., and Bialek W. Entropy and information in neural spike trains. Phys. Rev. Lett., 80:197-200, 1998. [20] Cover T.M. and Thomas J.A. Elements of Information Theory. Wiley, 1991. [21] Barkai N. and Leibler S. Robustness in simple biochemical networks. Nature, 387:913917, 1997. [22] Stemmler M. and Koch C. How voltage-dependent conductances can adapt to maximize the information encoded by neuronal firing rate. Nat. Neurosci., 2:521-7, 1999.
1657 |@word seems:1 nd:1 open:4 simulation:1 carry:1 extrastriate:1 liu:1 mainen:2 interestingly:1 current:23 i3n:1 activation:1 yet:2 physiol:3 opin:1 plot:1 alone:1 half:4 imitate:1 five:2 along:1 direct:1 ozaki:1 elads:1 inter:2 expected:2 behavior:5 examine:1 brain:1 discretized:2 decreasing:1 window:1 increasing:2 moreover:1 israel:2 what:1 cm:10 neurobiol:1 monkey:1 perseus:1 nj:2 temporal:4 quantitative:1 expands:1 ti:1 exactly:1 demonstrates:1 underlie:1 grant:1 appear:1 htotal:2 timing:10 local:1 modify:1 limit:2 acad:1 encoding:4 id:1 fluctuation:6 firing:13 might:1 twice:2 acta:1 suggests:1 range:2 steveninck:3 area:3 significantly:1 word:9 refers:1 suggest:3 get:2 influence:2 deterministic:2 center:1 missing:1 jerusalem:1 l:1 duration:1 recovery:1 rule:2 kay:1 population:3 classic:1 notion:2 variation:2 rieke:1 spontaneous:1 origin:1 pa:2 associate:1 element:1 utilized:1 std:1 mammalian:1 observed:6 bottom:1 fly:1 calculate:3 thousand:1 connected:1 decrease:1 knight:1 transmembrane:1 stochas:1 dynamic:7 depend:3 upon:1 neurophysiol:2 stemmler:1 train:29 sejnowski:2 encoded:1 elad:1 larger:1 sequence:2 biophysical:2 turrigiano:1 subtracting:1 interaction:1 adaptation:2 reproducibility:1 description:2 potassium:3 p:1 depending:1 ac:2 strong:2 epsps:1 c:1 quantify:1 direction:1 closely:1 stochastic:20 suprathreshold:1 bin:2 investigation:1 dendritic:1 biological:1 exploring:1 kinetics:3 hold:1 around:2 koch:3 lm:1 diminishes:1 injecting:1 proc:1 label:1 currently:1 sensitive:3 successfully:1 mit:1 gaussian:2 rather:6 inactivation:3 avoid:1 varying:2 voltage:6 encode:8 vk:2 mainly:1 contrast:1 sense:1 dependent:10 hille:1 vl:2 biochemical:1 transformed:1 integration:1 having:1 wlt:3 stimulus:12 others:2 opening:1 few:1 consisting:1 conductance:10 organization:1 highly:5 replicates:4 yielding:1 tj:2 natl:1 implication:1 accurate:1 nowak:1 respective:1 tree:1 increased:1 formalism:2 markovian:1 cover:1 newsome:1 deviation:3 shh:18 tishby:4 reported:2 conduction:1 density:27 huji:2 golowasch:1 na:19 again:1 slowly:3 stochastically:1 book:1 suggesting:1 potential:2 account:1 de:3 sec:2 coding:2 coefficient:2 depends:2 i1a:1 vivo:5 il:2 accuracy:1 yield:1 subthreshold:1 repeatability:3 ionic:1 comp:4 reach:2 phys:1 synaptic:2 ed:1 raster:1 lobster:1 frequency:1 conveys:1 gain:1 ask:1 amplitude:3 nerve:3 marom:1 reflected:1 response:9 strongly:1 just:2 hand:1 receives:1 replacing:2 mode:1 barkai:1 effect:7 hence:1 excitability:1 leibler:1 white:3 excitation:1 neocortical:1 demonstrate:1 temperature:2 ranging:1 recently:1 vitro:2 cerebral:1 jl:1 untouched:1 he:1 measurement:1 cv:1 closing:1 stochasticity:4 reliability:13 reich:1 cortex:3 surface:1 behaving:1 certain:1 victor:1 ministry:1 converge:1 maximize:2 schneidman:5 sliding:1 adapt:1 calculation:1 long:1 divided:2 impact:1 underlies:1 ion:28 cell:2 irregular:1 whereas:2 fine:4 interval:2 diagram:1 pyramidal:1 macroscopic:3 appropriately:1 hz:1 incorporates:1 inconsistent:1 near:2 enough:2 variety:1 affect:1 whether:2 six:1 bair:1 repeatedly:3 clear:1 amount:10 ten:1 reduced:1 lsi:1 millisecond:1 per:6 key:1 soma:2 threshold:2 undersampling:1 changing:5 abbott:1 letter:1 jitter:3 respond:2 hodgkin:5 injected:1 patch:5 oscillation:1 curro:1 comparable:1 bit:4 fl:3 hi:1 activity:3 marder:1 huxley:5 segev:5 encodes:6 optimality:1 performing:1 fitzhugh:1 relatively:1 department:1 according:1 membrane:17 smaller:1 lp:1 n4:1 rev:1 dv:1 equation:2 remains:1 describing:1 turn:1 mechanism:1 hh:18 drastic:1 reversal:1 operation:1 fluctuating:6 robustness:8 encounter:1 gate:1 convolved:1 thomas:1 top:2 convolve:1 denotes:1 log2:2 calculating:1 warland:1 giving:1 leakage:1 capacitance:1 spike:52 dependence:1 bialek:3 unclear:1 microscopic:1 softky:1 sci:1 capacity:18 alonso:1 cellular:1 gna:3 code:6 scand:1 ratio:8 hebrew:1 gk:3 trace:7 info:2 rise:1 kaplan:1 gated:1 mccormick:1 neuron:46 neurobiology:1 variability:2 frame:4 dc:3 introduced:1 namely:2 macaque:1 suggested:1 bar:1 pattern:1 memory:1 critical:1 natural:1 sodium:2 excitable:2 koberle:2 text:1 lewen:1 acknowledgement:1 determining:1 sinauer:1 loss:1 generation:1 conveyed:1 s0:1 shadlen:1 principle:1 balancing:1 gl:1 supported:1 keeping:2 institute:1 wide:2 van:3 curve:1 calculated:3 cortical:7 cmtt:1 transition:1 lett:1 qn:1 replicated:1 alpha:2 em2:1 unreliable:1 keep:1 vna:2 grayscale:1 stimulated:1 nature:8 channel:68 robust:4 expanding:1 ruyter:3 neurosci:3 noise:18 freedman:1 repeated:2 convey:1 neuronal:6 fig:9 representative:2 slow:1 wiley:1 lc:1 precision:11 sub:1 explicit:1 msec:4 ib:1 specific:1 repeatable:1 showing:1 normalizing:1 incorporating:1 intrinsic:1 adding:1 entorhinal:1 nat:1 entropy:18 idan:2 visual:1 doubling:1 kinetic:4 presentation:2 considerable:1 experimentally:3 change:16 content:1 specifically:2 reducing:1 f3n:4 total:9 pas:1 isopotential:2 experimental:1 la:4 defelice:2 correlated:4
719
1,658
Spike-based learning rules and stabilization of persistent neural activity Xiaohui Xie and H. Sebastian Seung Dept. of Brain & Cog. Sci., MIT, Cambridge, MA 02139 {xhxie, seung}@mit.edu Abstract We analyze the conditions under which synaptic learning rules based on action potential timing can be approximated by learning rules based on firing rates. In particular, we consider a form of plasticity in which synapses depress when a presynaptic spike is followed by a postsynaptic spike, and potentiate with the opposite temporal ordering. Such differential anti-Hebbian plasticity can be approximated under certain conditions by a learning rule that depends on the time derivative of the postsynaptic firing rate. Such a learning rule acts to stabilize persistent neural activity patterns in recurrent neural networks. 1 INTRODUCTION Recent experiments have demonstrated types of synaptic ~re 11111111111 111111111111111111 plasticity that depend on the post 11111111111 11111111? ?11. A temporal ordering of presynaptic and postsynaptic spiking. At B 0L-.i cortical [ I] and hippocampal[2] t .:oLl____ o synapses, long-term potenti1000 2000 tpost - tpre time (ms) ation is induced by repeated pairing of a presynaptic spike Figure I: (A) Pairing function for differential Heband a succeeding postsynaptic bian learning. The change in synaptic strength is plotspike, while long-term deprested versus the time difference between postsynaptic sion results when the order and presynaptic spikes. (B) Pairing function for difis reversed. The dependence ferential anti-Hebbian learning. (C) Differential antiof the change in synaptic Hebbian learning is driven by changes in firing rates. strength on the difference The synaptic learning rule of Eq. (l) is applied to two l:!..t = tpost - tpre between Poisson spike trains. The synaptic strength remains postsynaptic and presynaptic roughly constant in time, except when the postsynapspike times has been measured tic rate changes. quantitatively. This pairing function, sketched in Figure lA, has positive and negative lobes correspond to potentiation and depression. and a width of tens of milliseconds. We will refer to synaptic plasticity associated with this pairing function as differential Hebbian plasticity-Hebbian because the conditions for o~i=~=::====: _:3/_-----' ? \:;J __~ X Xie and H. S. Seung 200 potentiation are as predicted by Hebb[3], and differential because it is driven by the difference between the opposing processes of potentiation and depression. The pairing function of Figure IA is not characteristic of all synapses. For example, an opposite temporal dependence has been observed at electrosensory lobe synapses of electric fish[4]. As shown in Figure IB, these synapses depress when a presynaptic spike is followed by a postsynaptic one, and potentiate when the order is reversed. We will refer to this as differential anti-Hebbian plasticity. According to these experiments, the maximum ranges of the differential Hebbian and antiHebbian pairing functions are roughly 20 and 40 ms, respectively. This is fairly short, and seems more compatible with descriptions of neural activity based on spike timing rather than instantaneous firing rates[5, 6]. In fact, we will show that there are some conditions under which spike-based learning rules can be approximated by rate-based learning rules. Other people have also studied the relationship between spike-based and rate-based learning rules[7, 8]. The pairing functions of Figures IA and IB lead to rate-based learning rules like those traditionally used in neural networks, except that they depend on temporal derivatives of firing rates as well as firing rates themselves. We will argue that the differential antiHebbian learning rule of Figure IB could be a general mechanism for tuning the strength of positive feedback in networks that maintain a short-term memory of an analog variable in persistent neural activity. A number of recurrent network models have been proposed to explain memory-related neural activity in motor [9] and prefrontal [ 10] cortical areas, as well as the head direction system [11] and oculomotor integrator[ 12, 13, 14]. All of these models require precise tuning of synaptic strengths in order to maintain continuously variable levels of persistent activity. As a simple illustration of tuning by differential antiHebbian learning, a model of persistent activity maintained by an integrate-and-fire neuron with an excitatory autapse is studied. 2 SPIKE-BASED LEARNING RULE Pairing functions like those of Figure 1 have been measured using repeated pairing of a single presynaptic spike with a single postsynaptic spike. Quantitative measurements of synaptic changes due to more complex patterns of spiking activity have not yet been done. We will assume a simple model in which the synaptic change due to arbitrary spike trains is the sum of contributions from all possible pairings of presynaptic with postsynaptic spikes. The model is unlikely to be an exact description of real synapses, but could turn out to be approximately valid. We will write the spike train of the ith neuron as a series of Dirac delta functions, Si (t) = <5(t - Tr), where Tr is the nth spike time of the ith neuron. The synaptic weight from neuron j to i at time t is denoted by W ij (t). Then the change in synaptic weight induced by presynaptic spikes occurring in the time interval [0, Tj is modeled as Ln Wij(T + >.) - Wij(>') = [T dtj io foo dti f(ti - tj)Si(ti) Sj(tj) (1) -00 Each presynaptic spike is paired with all postsynaptic spikes produced before and after. For each pairing, the synaptic weight is changed by an amount depending on the pairing function f. The pairing function is assumed to be nonzero inside the interval [-T, Tj, and zero outside. We will refer to T as the pairing range. According to our model, each presynaptic spike results in induction of plasticity only after a latency>.. Accordingly, the arguments T + >. and >. of W ij on the left hand side of the equation are shifted relative to the limits T and 0 of the integral on the right hand side. We 201 Spike-based Learning and Stabilization ofPersistent Neural Activity will assume that the latency>. is greater than the pairing range T, so that Wi} at any time is only influenced by presynaptic and postsynaptic spikes that happened before that time, and therefore the learning rule is causal. 3 RELATION TO RATE-BASED LEARNING RULES The learning rule of Eq. (1) is driven by correlations between presynaptic and postsynaptic activities. This dependence can be made explicit by making the change of variables u = ti - t j in Eq. (I), which yields Wij(T + >.) - W ij (>.) = iTT duf(u)Cij(u) (2) where we have defined the cross-correlation Cij(u) = !aT dt Si(t + u) Sj(t) . (3) and made use of the fact that f vanishes outside the interval [-T, T]. Our immediate goal is to relate Eq. (2) to learning rules that are based on the cross-correlation between firing rates, Crre(u) = !aT dt Vi(t + u) Vj(t) (4) There are a number of ways of defining instantaneous firing rates. Sometimes they are computed by averaging over repeated presentations of a stimulus. In other situations, they are defined by temporal filtering of spike trains. The following discussion is general, and should apply to these and other definitions of firing rates. The "rate correlation" is commonly subtracted from the total correlation to obtain the "spike correlation" C:r ke = Cij - Cijate. To derive a rate-based approximation to the learning rule (2), we rewrite it as Wij(T + >.) - Wij(>') = iTT du f(u)Cijate(u) + iTT du f(u)C:r ke (u) (5) and simply neglect the second term. Shortly we will discuss the conditions under which this is a good approximation. But first we derive another form for the first term by applying the approximation Vi(t + u) ~ Vi(t) + UVi(t) to obtain j T -T duf(u)Crre(u) ~ iT dt[fiovi(t) + 131Vi(t)]VJ (t) (6) 0 where we define (7) This approximation is good when firing rates vary slowly compared to the pairing range T . The learning rule depends on the postsynaptic rate through fio Vi + 131 Vi . When the first term dominates the second, then the learning rule is the conventional one based on correlations between firing rates, and the sign of fio determines whether the rule is Hebbian or anti-Hebbian. In the remainder of the paper, we will discuss the more novel case where 130 = O. This holds for the pairing functions shown in Figures lA and IB, which have positive and negative lobes with areas that exactly cancel in the definition of 130. Then the dependence on X Xie and H. S. Seung 202 postsynaptic activity is purely on the time derivative of the firing rate. Differential Hebbian learning corresponds to /31 > 0 (Figure IA), while differential anti-Hebbian learning leads to /31 < 0 (Figure IB). To summarize the /30 = 0 case, the synaptic changes due to rate correlations are approximated by W ij ex: -ViVj (diff. anti-Hebbian) (8) for slowly varying rates. These formulas imply that a constant postsynaptic firing rate causes no net change in synaptic strength. Instead, changes in rate are required to induce synaptic plasticity. To illustrate this point, Figure lC shows the result of applying differential anti-Hebbian learning to two spike trains. The presynaptic spike train was generated by a 50 Hz Poisson process, while the postsynaptic spike train was generated by an inhomogeneous Poisson process with rate that shifted from 50 Hz to 200 Hz at 1 sec. Before and after the shift, the synaptic strength fluctuates but remains roughly constant. But the upward shift in firing rate causes a downward shift in synaptic strength, in accord with the sign of the differential anti-Hebbian rule in Eq. (8). The rate-based approximation works well for this example, because the second term of Eq. (5) is not so important. Let us return to the issue of the general conditions under which Pike (u) are this term can be neglected. With Poisson spike trains, the spike correlations zero in the limit T -7 00, but for finite T they fluctuate about zero. The integr~l over u in the second term of (5) dampens these fluctuations. The amount of dampening depends on the pairing range T, which sets the limits of integration. In Figure 1C we used a relatively long pairing range of 100 ms, which made the fluctuations small even for small T. On the other hand, if T were short, the fluctuations would be small only for large T_ Averaging over large T is relevant when the amplitUde of f is small, so that the rate of learning is slow. In this case, it takes a long time for significant synaptic changes to accumulate, so that plasticity is effectively driven by integrating over long time periods T in Eq. (l). C: In the brain, nonvanishing spike correlations are sometimes observed even in the T -7 00 limit, unlike with Poisson spike trains. These correlations are often roughly symmetric about zero, in which case they should produce little plasticity if the pairing functions are antisymmetric as in Figures lA and lB. On the other hand, if the spike correlations are asymmetric, they could lead to substantial effects[6]. 4 EFFECTS ON RECURRENT NETWORK DYNAMICS The learning rules of Eq. (8) depend on both presynaptic and postsynaptic rates, like learning rules conventionally used in neural networks. They have the special feature that they depend on time derivatives, which has computational consequences for recurrent neural networks of the form Xi + Xi = L Wiju(Xj) + bi (9) j Such classical neural network equations can be derived from more biophysically realistic models using the method of averaging[ 15] or a mean field approximation[ 16]. The firing rate of neuron j is conventionally identified with Vj = u(Xj). v; The cost function E( {Xi}; {Wij}) = ~ Li quantifies the amount of drift in firing rate at the point Xl , ... , X N in the state space of the network. If we consider Vi to be a function of Xi and W ij defined by (9), then the gradient ofthe cost function with respect to W ij is given by BE / BWij = U' (Xi)ViVj. Assuming that U is a monotonically increasing function so that u' (xd > 0, it follows that the differential Hebbian update of (8) increases the cost function, Spike-based Learning and Stabilization ofPersistent Neural Activity 203 and hence increases the magnitude of the drift velocity. In contrast, the differential antiHebbian update decreases the drift velocity. This suggests that the differential anti-Hebbian update could be useful for creating fixed points of the network dynamics (9). 5 PERSISTENT ACTIVITY IN A SPIKING AUTAPSE MODEL The preceding arguments about drift velocity were based on approximate rate-based descriptions of learning and network dynamics. It is important to implement spike-based learning in a spiking network dynamics, to check that our approximations are valid. Therefore we have numerically simulated the simple recurrent circuit of integrate-and-fire neurons shown in Figure 2. The core of the circuit is the "memory neuron," which makes an excitatory autapse onto itself. It also receives synaptic input from three input neurons: a tonic neuron, an excitatory burst neuron, and an inhibitory burst neuron. It is known that this circuit can store a shortterm memory of an analog variable in persistent activity, if the strengths of the autapse and tonic synapse are precisely INHIBITORY BURST tuned[ 17]. Here we show that this tun? ing can be accomplished by the spikebased learning rule of Eq. (1), with a dFigure 2: Circuit diagram for autapse model ifferential anti-Hebbian pairing function like that of Figure 1B. The memory neuron is described by the equations C m dr Tsyn dt dV dt = (10) (1) +r n where V is the membrane potential. When V reaches V'thres, a spike is considered to have occurred, and V is reset to Vreset. Each spike at time Tn causes a jump in the synaptic activation r of size CY.r/Tsyn, after which r decays exponentially with time constant Tsyn until the next spike. The synaptic conductances of the memory neuron are given by (12) The term W r is recurrent excitation from the autapse, where W is the strength of the autapse. The synaptic activations ro, r +, and r _ of the tonic, excitatory burst, and inhibitory burst neurons are governed by equations like (10) and (1), with a few differences. These neurons have no synaptic input; their firing patterns are instead determined by applied currents lapp,o, lapp,+ and lapp,_. The tonic neuron has a constant applied current, which makes it fire repetitively at roughly 20 Hz (Figure 3). For the excitatory and inhibitory burst neurons the applied current is normally zero, except for brief 100 ms current pulses that cause bursts of action potentials. As shown in Figure 3, if the synaptic strengths W and Wo are arbitrarily set before learning, the burst neurons cause only transient changes in the firing rate of the memory neuron. After applying the spike-based learning rule (1) to tune both W and W o, the memory X Xie and H. S. Seung 204 111111111111 I ~IIIIIIIII I IUIIIIIIII I /untuned 111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111 I I~ ____~I~____~______=-_____ I " tuned 1IIIIIIIIIIIIIIIIIIIIIIIIIIINIIIIII'.tl I 1 sec 1111111111'"111111111111111111111111111 Figure 3: Untuned and tuned autapse activity. The middle three traces are the membrane potentials of the three input neurons in Figure 2 (spikes are drawn at the reset times of the integrate-and-fire neurons). Before learning, the activity of the memory neuron is not persistent, as shown in the top trace. After the spike-based learning rule (1) is applied to the synaptic weights Wand W o, then the burst inputs cause persistent changes in activity. em = 1 nF, gL = 0.025 J-lS, VL = -70 mY, VE = 0 mY, VI = -70 mY, vthres = -52 mY, Vr eset = -59 mY, a s = 1, Tsyn = 100 ms, Iapp,o = 0.5203 nA, I app ,? = 0 or 0.95 nA, Ts yn ,O = 100 ms, Tsyn,+ = Ts yn,- = 5 ms, W+ = 0.1, W_ = 0.05. neuron is able to maintain persistent activity. During the interburst intervals (from A after one burst until A before the next), we made synaptic changes using the differential antiHebbian pairing function f(t) = -Asin(7l'tjT) for spike time differences in the range [-T, T] with A = 1.5 X 10- 4 and T=A=120 ms. The resulting increase in persistence time can be seen in Figure 4A, along with the values of the synaptic weights versus time. To quantify the performance of the system at maintaining persistent activity, we determined the relationship between dv / dt and v using a long sequence of interburst intervals, where v was defined as the reciprocal of the interspike interval. If Wand Wo are fixed at optimally tuned values, there is still a residual drift, as shown in Figure 4B. But if these parameters are allowed to adapt continuously, even after good tuning has been achieved, then the residual drift is even smaller in magnitude. This is because the learning rule tweaks the synaptic weights during each interburst interval, reducing the drift for that particular firing rate. Autapse learning is driven by the autocorrelation of the spike train, rather than a crosscorrelation. The peak in the autocorrelogram at zero lag has no effect, since the pairing function is zero at the origin. Since the autocorrelation is zero for small time lags, we used a fairly large pairing range in our simulations. In a recurrent network of many neurons, a shorter pairing range would suffice, as the cross-correlation does not vanish near zero. 6 DISCUSSION We have shown that differential anti-Hebbian learning can tune a recurrent circuit to maintain persistent neural activity. This behavior can be understood by reducing the spike-based learning rule (l) to the rate-based learning rules ofEqs. (6) and (8). The rate-based approximations are good if two conditions are satisfied. First, the pairing range must be large, or the rate of learning must be slow. Second, spike synchrony must be weak, or have little effect on learning due to the shape of the pairing function. The differential anti-Hebbian pairing function results in a learning rule that uses -Vi as a negative feedback signal to reduce the amount of drift in firing rate, as illustrated by our simulations of an integrate-and-fire neuron with an excitatory autapse. More generally, the learning rule could be relevant for tuning the strength of positive feedback in networks that maintain a short-term memory of an analog variable in persistent neural activity. Spike-based Learning and Stabilization of Persistent Neural Activity A 250 B c ' 200 0.395 ~150 W WO 0.16 0385 ~ i! 0 ?100 "" 0.12 10 20 "me Is) I 6 I 4 1 5 10 15 tlme(s) 20 .~~ 2 1~ :I: ~r? 0 ~-2 ~~! 50 00 205 25 ~~~ -4f -at -8' 20 40 60 rate (Hzl '0 80 i 100 Figure 4: Tuning the autapse. (A) The persistence time of activity increases as the weights Wand Wo are tuned. Each transition is driven by pseudorandom bursts of input (B) Systematic relationship between drift dv/dt in firing rate and v, as measured from a long sequence of interburst intervals. If the weights are continuously fine-tuned ('*') the drift is less than with fixed well-tuned weights ('0'). For example, the learning rule could be used to improve the robustness of the oculomotor integrator[12, 13, 14] and head direction system[l1] to mistuning of parameters. In deriving the differential forms of the learning rules in (8), we assumed that the areas under the positive and negative lobes of the pairing function are equal, so that the integral defining 130 vanishes. In reality, this cancellation might not be exact. Then the ratio of 131 and 130 would limit the persistence time that can be achieved by the learning rule. Both the oculomotor integrator and the head direction system are also able to integrate vestibular inputs to produce changes in activity patterns. The problem of finding generalizations of the present learning rules that train networks to integrate is still open . References [1] H. Markram, J. Lubke, M. Frotscher, and B. Sakmann. Science, 275(5297):213-5, 1997. [2] G. Q. Bi and M. M. Poo. 1 Neurosci, 18(24):10464-72,1998. [3] D. O . Hebb. Organization of behavior. Wiley, New York, 1949. [4] C. C. Bell, V. Z. Han, Y. Sugawara, and K. Grant. Nature , 387(6630):278-81 , 1997. [5] w. Gerstner, R . Kempter, 1. L. van Hemmen, and H. Wagner. Nature, 383(6595):76-81, 1996. [6] L. F. Abbott and S. Song. Adv. Neural Info. Proc. Syst., 11, 1999. [7] P. D. Roberts . 1. Comput. Neurosci., 7:235-246, 1999. [8] R. Kempter, W. Gerstner, and J. L. van Hemmen. Phys. Rev. E, 59(4):4498-4514, 1999. [9] A. P. Georgopoulos, M. Taira, and A. Lukashin. Science, 260:47-52, 1993. [10] M. Camperi and X. J. Wang. 1 Comput Neurosci, 5(4):383-405, 1998. [11] K. Zhang. 1. Neurosci., 16:2112-2126, 1996. [12] S. C. Cannon, D. A. Robinson, and S. Shamma. Bio!. Cybern., 49:127-136,1983. [13] H. S. Seung. Proc. Nat!. A cad. Sci. USA, 93:13339-13344, 1996. [14] H. S. Seung, D. D. Lee, B. Y. Reis, and D. W. Tank. Neuron, 2000. [15] B. Ermentrout. Neural Comput., 6:679-695, 1994. [16] O. Shriki, D. Hansel, and H. Sompolinsky. Soc. Neurosci. Abstr., 24:143, 1998. [17] H. S. Seung, D. D. Lee, B. Y. Reis, and D. W. Tank. 1. Comput. Neurosci., 2000. III THEORY PART
1658 |@word middle:1 seems:1 open:1 simulation:2 pulse:1 lobe:4 electrosensory:1 t_:1 thres:1 tr:2 series:1 tuned:7 current:4 cad:1 si:3 yet:1 activation:2 must:3 realistic:1 interspike:1 plasticity:10 shape:1 motor:1 succeeding:1 update:3 accordingly:1 ith:2 reciprocal:1 short:4 core:1 zhang:1 burst:11 along:1 differential:20 persistent:14 pairing:30 inside:1 autocorrelation:2 behavior:2 themselves:1 roughly:5 brain:2 integrator:3 little:2 autapse:11 increasing:1 suffice:1 circuit:5 tic:1 finding:1 temporal:5 quantitative:1 dti:1 nf:1 act:1 ti:3 xd:1 exactly:1 ro:1 asin:1 bio:1 normally:1 grant:1 yn:2 positive:5 before:6 understood:1 timing:2 limit:5 io:1 consequence:1 firing:21 fluctuation:3 approximately:1 might:1 studied:2 suggests:1 shamma:1 range:10 bi:2 sugawara:1 implement:1 area:3 bell:1 persistence:3 induce:1 integrating:1 spikebased:1 onto:1 applying:3 cybern:1 conventional:1 xiaohui:1 demonstrated:1 poo:1 l:1 ke:2 rule:35 deriving:1 traditionally:1 exact:2 us:1 origin:1 velocity:3 approximated:4 asymmetric:1 observed:2 wang:1 cy:1 adv:1 sompolinsky:1 ordering:2 decrease:1 substantial:1 vanishes:2 seung:8 ermentrout:1 neglected:1 dynamic:4 depend:4 rewrite:1 purely:1 lapp:3 train:11 outside:2 fluctuates:1 lag:2 itself:1 sequence:2 net:1 reset:2 remainder:1 relevant:2 description:3 dirac:1 abstr:1 produce:2 depending:1 recurrent:8 derive:2 illustrate:1 measured:3 ij:6 eq:9 soc:1 predicted:1 quantify:1 direction:3 inhomogeneous:1 hzl:1 stabilization:4 transient:1 require:1 potentiation:3 generalization:1 tjt:1 hold:1 considered:1 shriki:1 vary:1 proc:2 hansel:1 mit:2 rather:2 cannon:1 fluctuate:1 sion:1 varying:1 derived:1 check:1 contrast:1 vl:1 unlikely:1 relation:1 wij:6 sketched:1 upward:1 issue:1 tank:2 denoted:1 integration:1 fairly:2 special:1 frotscher:1 field:1 equal:1 cancel:1 stimulus:1 quantitatively:1 few:1 ve:1 taira:1 fire:5 opposing:1 maintain:5 dampening:1 conductance:1 organization:1 tj:4 integral:2 shorter:1 re:1 xhxie:1 causal:1 cost:3 tweak:1 optimally:1 my:5 peak:1 systematic:1 lee:2 continuously:3 nonvanishing:1 na:2 satisfied:1 prefrontal:1 slowly:2 dr:1 creating:1 derivative:4 return:1 crosscorrelation:1 li:1 syst:1 potential:4 w_:1 sec:2 stabilize:1 depends:3 vi:9 analyze:1 synchrony:1 contribution:1 lubke:1 characteristic:1 correspond:1 yield:1 ofthe:1 weak:1 biophysically:1 produced:1 app:1 explain:1 synapsis:6 influenced:1 reach:1 phys:1 sebastian:1 synaptic:29 definition:2 associated:1 camperi:1 amplitude:1 dt:7 xie:4 bian:1 synapse:1 done:1 correlation:13 until:2 hand:4 receives:1 usa:1 effect:4 hence:1 symmetric:1 nonzero:1 illustrated:1 during:2 width:1 maintained:1 excitation:1 m:8 hippocampal:1 tn:1 l1:1 instantaneous:2 novel:1 spiking:4 exponentially:1 analog:3 occurred:1 numerically:1 accumulate:1 refer:3 potentiate:2 measurement:1 cambridge:1 significant:1 tuning:6 cancellation:1 han:1 recent:1 driven:6 store:1 certain:1 arbitrarily:1 accomplished:1 seen:1 greater:1 preceding:1 period:1 monotonically:1 signal:1 hebbian:19 ing:1 adapt:1 cross:3 long:7 repetitively:1 post:1 paired:1 poisson:5 sometimes:2 accord:1 achieved:2 fine:1 interval:8 diagram:1 unlike:1 induced:2 hz:4 near:1 iii:1 xj:2 identified:1 opposite:2 reduce:1 shift:3 whether:1 depress:2 wo:4 song:1 york:1 cause:6 pike:1 action:2 depression:2 useful:1 latency:2 generally:1 tune:2 amount:4 ten:1 mistuning:1 millisecond:1 fish:1 shifted:2 happened:1 sign:2 delta:1 inhibitory:4 write:1 drawn:1 abbott:1 interburst:4 sum:1 wand:3 dampens:1 followed:2 autocorrelogram:1 activity:24 strength:12 precisely:1 georgopoulos:1 argument:2 pseudorandom:1 relatively:1 uvi:1 according:2 membrane:2 fio:2 smaller:1 em:1 postsynaptic:17 wi:1 rev:1 making:1 dv:3 ln:1 equation:4 remains:2 lukashin:1 turn:1 discus:2 mechanism:1 integr:1 apply:1 subtracted:1 robustness:1 shortly:1 top:1 maintaining:1 neglect:1 classical:1 crre:2 iapp:1 spike:46 dependence:4 gradient:1 reversed:2 sci:2 simulated:1 me:1 presynaptic:15 argue:1 induction:1 assuming:1 modeled:1 relationship:3 illustration:1 ratio:1 cij:3 robert:1 relate:1 info:1 trace:2 negative:4 vivj:2 sakmann:1 neuron:26 finite:1 anti:12 t:2 immediate:1 defining:2 situation:1 tonic:4 head:3 precise:1 arbitrary:1 lb:1 drift:10 required:1 vestibular:1 tpre:2 antihebbian:5 able:2 robinson:1 pattern:4 summarize:1 oculomotor:3 tun:1 memory:10 ia:3 ation:1 residual:2 nth:1 improve:1 brief:1 imply:1 conventionally:2 shortterm:1 vreset:1 relative:1 kempter:2 filtering:1 versus:2 untuned:2 integrate:6 compatible:1 excitatory:6 changed:1 gl:1 side:2 markram:1 wagner:1 van:2 feedback:3 cortical:2 valid:2 transition:1 tpost:2 made:4 commonly:1 jump:1 tlme:1 sj:2 approximate:1 assumed:2 xi:5 quantifies:1 reality:1 nature:2 itt:3 du:2 gerstner:2 complex:1 electric:1 vj:3 antisymmetric:1 neurosci:6 repeated:3 dtj:1 allowed:1 tl:1 hemmen:2 hebb:2 slow:2 vr:1 foo:1 lc:1 wiley:1 explicit:1 xl:1 comput:4 governed:1 ib:5 vanish:1 formula:1 cog:1 decay:1 dominates:1 effectively:1 magnitude:2 nat:1 downward:1 occurring:1 simply:1 corresponds:1 determines:1 ma:1 goal:1 presentation:1 vthres:1 change:16 determined:2 except:3 diff:1 reducing:2 averaging:3 total:1 la:3 people:1 dept:1 ex:1
720
1,659
Boosting with Multi-Way Branching in Decision Trees Yishay Mansour David McAllester AT&T Labs-Research 180 Park Ave Florham Park NJ 07932 {mansour, dmac }@research.att.com Abstract It is known that decision tree learning can be viewed as a form of boosting. However, existing boosting theorems for decision tree learning allow only binary-branching trees and the generalization to multi-branching trees is not immediate. Practical decision tree algorithms, such as CART and C4.5, implement a trade-off between the number of branches and the improvement in tree quality as measured by an index function . Here we give a boosting justification for a particular quantitative trade-off curve. Our main theorem states, in essence, that if we require an improvement proportional to the log of the number of branches then top-down greedy construction of decision trees remains an effective boosting algorithm. 1 Introduction Decision trees have been proved to be a very popular tool in experimental machine learning. Their popularity stems from two basic features - they can be constructed quickly and they seem to achieve low error rates in practice. In some cases the time required for tree growth scales linearly with the sample size. Efficient tree construction allows for very large data sets. On the other hand, although there are known theoretical handicaps of the decision tree representations, it seem that in practice they achieve accuracy which is comparable to other learning paradigms such as neural networks. While decision tree learning algorithms are popular in practice it seems hard to quantify their success ,in a theoretical model. It is fairly easy to see that even if the target function can be described using a small decision tree, tree learning algorithms may fail to find a good approximation. Kearns and, Mansour [6] used the weak learning hypothesis to show that standard tree learning algorithms perform boosting. This provides a theoretical justification for decision tree learning similar Boosting with Multi-Way Branching in Decision Trees 301 to justifications that have been given for various other boosting algorithms, such as AdaBoost [4]. Most decision tree learning algorithms use a top-down growth process. Given a current tree the algorithm selects some leaf node and extends it to an internal node by assigning to it some "branching function" and adding a leaf to each possible output value of this branching function. The set of branching functions may differ from one algorithm to another, but most algorithms used in practice try to keep the set of branching functions fairly simple. For example, in C4.5 [7], each branching function depends on a single attribute. For categorical attributes, the branching is according to the attribute's value, while for continuous attributes it performs a comparison of the attribute with some constant. Of course such top-down tree growth can over-fit the data - it is easy to construct a (large) tree whose error rate on the training data is zero. However, if the class of splitting functions has finite VC dimension then it is possible to prove that, with high confidence of the choice of the training data, for all trees T the true error rate of T is bounded by f(T) + 0 (JITI/m) where f(T) is the error rate of T on the training sample, ITI is the number of leaves of T, and m is the size of the training sample. Over-fitting can be avoided by requiring that top-down tree growth produce a small tree. In practice this is usually done by constructing a large tree and then pruning away some of its nodes. Here we take a slightly different approach. We assume a given target tree size s and consider the problem of constructing a tree T with ITI = sand f(T) as small as possible. We can avoid over-fitting by selecting a small target value for the tree size. A fundamental question in top-down tree growth is how to select the branching function when growing a given leaf. We can think of the target size as a "budget" . A four-way branch spends more of the tree size budget than does a two-way branch - a four-way branch increases the tree size by roughly the same amount as two twoway branches. A sufficiently large branch would spend the entire tree size budget in a single step. Branches that spend more of the tree size budget should be required to achieve more progress than branches spending less ofthe budget. Naively, one would expect that the improvement should be required to be roughly linear in the number of new leaves introduced - one should get a return proportional to the expense. However, a weak learning assumption and a target tree size define a nontrivial game between the learner and an adversary. The learner makes moves by selecting branching functions and the adversary makes moves by presenting options consistent with the weak learning hypothesis. We prove here that the learner achieve a better value in this game by selecting branches that get a return considerably smaller than the naive linear return. Our main theorem states, in essence, that the return need only be proportional to the log of the number of branches. 2 Preliminaries We assume a set X of instances and an unknown target function f mapping X to {O,l}. We assume a given "training set" S which is a set of pairs of the form (x, f(x)). We let 1l be a set of potential branching functions where each hE 1l is a function from X to a finite set Rh - we allow different functions in 1l to have different ranges. We require that for any h E 1l we have IRhl ~ 2. An 1l-tree is Y. Mansour and D. McAllester 302 a tree where each internal node is labeled with an branching function h E 1i and has children corresponding to the elements of the set Rh. We define ITI to be the number ofleafnodes ofT. We let L(T) be the set ofleafnodes ofT. For a given tree T, leaf node f of T and sample S we write Sl to denote the subset of the sample S reaching leaf f. For f E T we define Pl to be the fraction of the sample reaching leaf f, i.e., ISll/ISI. We define ql to be the fraction of the pairs (x, f(x? in Sl for which f(x) = 1. The training error ofT, denoted i(T), is L:lEL(T)Plmin(ql, 1- ql). 3 The Weak Learning Hypothesis and Boosting Here, as in [6], we view top-down decision tree learning as a form of Boosting [8, 3]. Boosting describes a general class of iterative algorithms based on a weak learning hypothesis. The classical weak learning hypothesis applies to classes of Boolean functions. Let 1i2 be the subset of branching functions h E 1i with IRhl = 2. For c5 > the classical c5-weak learning hypothesis for 1i2 states that for any distribution on X there exists an hE 1i2 with PrD(h(x) f f(x)) ~ 1/2-c5. Algorithms designed to exploit this particular hypothesis for classes of Boolean functions have proved to be quite useful in practice [5]. ? Kearns and Mansour show [6] that the key to using the weak learning hypothesis for decision tree learning is the use of an index function I : [0, 1] ~ [0,1] where I(q) ~ 1, I(q) ~ min(q, (1- q)) and where I(T) is defined to be L:lEL(T) PlI(ql). Note that these conditions imply that i(T) ~ I(T). For any sample W let qw be the fraction of pairs (x, f(x)) E W such that f(x) = 1. For any h E 1i let Th be the decision tree consisting of a single internal node with branching function h plus a leaf for each member of IRh I. Let Iw (Th) denote the value of I(Th) as measured with respect to the sample W. Let ~ (W, h) denote I (qW ) - Iw (Th). The quantity ~(W, h) is the reduction in the index for sample W achieved by introducing a single branch. Also note that Pt~(Sl, h) is the reduction in I(T) when the leaf f is replaced by the branch h. Kearns and Mansour [6] prove the following lemma. Lemma 3.1 (Kearns & Mansour) Assuming the c5-weak learning hypothesis for 1i2, and taking I(q) to be 2Jq(1- q), we have that for any sample W there exists an h E 1i2 such that ~(W,h) ~ ~:I(qw). This lemma motivates the following definition. Definition 1 We say that 1i2 and I satisfies the "I-weak tree-growth hypothesis if for any sample W from X there exists an hE 1i2 such that ~(W, h) ~ "II(qw). Lemma 3.1 states, in essence, that the classical weak learning hypothesis implies the weak tree growth hypothesis for the index function I(q) = 2Jq(l - q). Empirically, however, the weak tree growth hypothesis seems to hold for a variety of index functions that were already used for tree growth prior to the work of Kearns and Mansour. The Ginni index I(q) 4q(1 - q) is used in CART [1] and the entropy I(q) = -q log q - (1- q) log(l- q) is used in C4.5 [7]. It has long been empirically observed that it is possible to make steady progress in reducing I(T) for these choices of I while it is difficult to make steady progress in reducing i(T). = We now define a simple binary branching procedure. For a given training set S and target tree size s this algorithm grows a tree with ITI = s. In the algorithm Boosting with Multi-Way Branching in Decision Trees 303 odenotes the trivial tree whose root is a leaf node and Tl ,h denotes the result of replacing the leaf l with the branching function h and a new leaf for each element of Rh. T=0 WHILE (ITI l f- < s) DO argmaxl 'ftl1(til) h f- argmaxhEl?:l~(Sl' h) T f- Tl,h; END-WHILE We now define e(n) to be the quantity TI~:ll(l-;). Note that e(n) ~ TI~:/ e- 7 ~ .. - l /" e--Y Wi"'l 1 S < e--Y Inn = n--Y. = Theorem 3.2 (Kearns & Mansour) 1f1l2 and I satisfy the ,-weak tree growth hypothesis then the binary branching procedure produces a tree T with i(T) ~ I(T) ~ e(ITI) ~ ITI--Y? Proof: The proof is by induction on the number of iterations of the procedure. We have that 1(0) ~ 1 e(l) so the initial tree immediately satisfies the condition. We now assume that the condition is satisfied by T at the begining of an iteration and prove that it remains satisfied by Tl,h at the end of the iteration. Since I(T) = LlET Ih1(til) we have that the leaf l selected by the procedure is such that Pl1(til) 2: II~)? By the ,-weak tree growth assumption the function h selected by the procedure has the property that ~(Sl, h) 2: ,1(ql). We now l " This implies that have that I(T) - I(Tl,h) = Pl~(Sl' h) 2: P1I1(til) 2: I(Tl,h) ~ I(T) - rh1(T) = (1- j;)I(T) ~ (1- rh)e(ITI) = e(ITI + 1) = e(ITl,hl). = ,II?i o 4 Statement of the Main Theorem We now construct a tree-growth algorithm that selects multi-way branching functions. As with many weak learning hypotheses, the ,-weak tree-growth hypothesis can be viewed as defining a game between the learner and an adversary. Given a tree T the adversary selects a set of branching functions allowed at each leaf of the tree subject to the constraint that at each leaf l the adversary must provide a binary branching function h with ~(Sl' h) 2: ,1(til). The learner then selects a leaf land a branching function h and replaces T by Tl,h. The adversary then again selects a new set of options for each leaf subject to the ,-weak tree growth hypothesis. The proof of theorem 3.2 implies that even when the adversary can reassign all options at every move there exists a learner strategy, the binary branching procedure, guaranteed to achieves a final error rate of ITI--Y. Of course the optimal play for the adversary in this game is to only provide a single binary option at each leaf. However, in practice the "adversary" will make mistakes and provide options to the learner which can be exploited to achieve even lower error rates. Our objective now is to construct a strategy for the learner which can exploit multi-way branches provided by the adversary. We first say that a branching function h is acceptable for tree T and target size Y. Mansour and D. MeAl/ester 304 = s if either IRhl 2 or ITI < e(IRh!)s"Y/(2IRh!). We also define g(k) to be the quantity (1 - e(k?/"Y. It should be noted that g(2) = 1. It should also be noted that e( k) '" e -'Y Ink and hence for "Y In k small we have e( k) '" 1 - "Y In k and hence g(k) '" Ink. We now define the following multi-branch tree growth procedure. T=0 WHILE (ITI < s) DO l +- argm~ Ptl(qt) h +- argmaxhEll , h acceptable for T and s ~(St, h)/g(IRhl) T +- Tt,h; END-WHILE A run of the multi-branch tree growth procedure will be called "Y-boosting if at each iteration the branching function h selected has the property that ~(St, h) / g(lRh I) ~ "YI(qt). The "Y-weak tree growth hypothesis implies that ~(St,h)/g(IRhl) ~ "YI(qt)/g(2) = "YI(qt). Therefore, the "Y-weak tree growth hypothesis implies that every run of the multi-branch growth procedure is "Y-bootsing. But a run can be "Y-bootsing by exploiting mutli-way branches even when the "Y-weak tree growth hypothesis fails. The following is the main theorem of this paper. Theorem 4.1 1fT is produced by a "Y-boosting run of the multi-branch tree-growth procedure then leT) 5 ~ e(ITI) ~ ITI-'Y? Proof of Theorem 4.1 To prove the main theorem we need the concept of a visited weighted tree, or VWtree for short. A VW-tree is a tree in which each node m is assigned both a rational weight Wm E [0,1] and an integer visitation count Vm ~ 1. We now define the following VW tree growth procedure. In the procedure Tw is the tree consisting of a single root node with weight wand visitation count 1. The tree Tt.w1 .... .w/c is the result of inserting k new leaves below the leaf l where the ith new leaf has weight Wi and new leaves have visitation count 1. W +- any rational number in [0,1] T+-Tw FOR ANY NUMBER OF STEPS REPEAT THE FOLLOWING l +- argmaxt e(tI:~wl Vt +- Vt + 1 OPTIONALLY T +- Tt.Wl .. .. ,W lll WITH WI + .. . Wtll ~ e(vt)wt We first prove an analog of theorem 3.2 for the above procedure. For a VW-tree T we define ITI to be LtEL(T) Vt and we define leT) to be LtEL(T) e( Vt)Wt. Lemma 5.1 The VW procedure maintains the invariant that leT) ~ e(ITI). Proof: The proof is by induction on the number of iterations of the algorithm. The result is immediate for the initial tree since eel) 1. We now assume that leT) ~ e(IT!) at the start of an iteration and show that this remains true at the end of the iteration. = 305 Boosting with Multi- Way Branching in Decision Trees We can associate each leaf l with Vt "subleaves" each of weight e(vt)wt/Vt. We have that ITI is the total number of these subleaves and I(T) is the total weight of these subleaves. Therefore there must exist a subleaf whose weight is at least I(T)/ITI. Hence there must exist a leaf l satisfying e(vt)wt/Vt 2': I(T)/ITI. Therefore this relation must hold of the leaf l selected by the procedure. = Let T' be the tree resulting from incrementing Vt. We now have I(T) - I(T') e(vt)wt- e(vt + l)wt = e(vt)wt- (1- ;;)e(vt)wt ;;e(vt)wt 2': "/I~)' So we have I(T') ~ (1 )I(T) ~ (1 )e(ITI) = e(IT'I). ffl ffl = Finally, if the procedure grows new leaves we have that the I(T) does not increase and that ITI remains the same and hence the invariant is maintained. 0 For any internal node m in a tree T let C(m) denote the set of nodes which are children of m. A VW-tree will be called locally-well-formed if for every internal node m we have that Vm = IC(m)l, that I:nEC(m) Wn ~ e(IC(m)l)wm . A VW-tree will be called globally-safe ifmaxtEL(T) e(vt)wt/Vt ~ millmEN(T) e(vt-1)wt/(vt-1) where N(T) denotes the set of internal nodes of T. Lemma 5.2 If T is a locally well-formed and globally safe VW-tree, then T is a possible output of the VW growth procedure and therefore I(T) ~ e(ITI). Proof: Since T is locally well formed we can use T as a "template" for making nondeterministic choices in the VW growth procedure. This process is guaranteed to produce T provided that the growth procedure is never forced to visit a node corresponding to a leaf of T. But the global safety condition guarantees that any unfinished internal node of T has a weight as least as large as any leaf node of T. o We now give a way of mapping ?i-trees into VW-trees. More specifically, for any ?i-tree T we define VW(T) to be the result of assigning each node m in T the weight PmI(qm), each internal node a visitation count equal to its number of children, and each leaf node a visitation count equal to 1. We now have the following lemmas. Lemma 5.3 If T is grown by a I-boosting run of the multi-branch procedure then VW(T) is locally well-formed. Proof: Note that the children of an internal node m are derived by selecting a branching function h for the node m. Since the run is I-boosting we have ~(St, h)/g(IRhi) 2': II(qt). Therefore ~(St, h) = (I(tit) - 1St (n)) 2': I(tit)(l e(IRhl)). This implies that Ist(Th ) ~ e(IRhDI(qt). Multiplying by Pt and trans0 forming the result into weights in the tree VW(T) gives the desired result. The following lemma now suffices for theorem 4.1. Lemma 5.4 If T is grown by a I-boosting run of the multi-branch procedure then VW(T) is globally safe. Proof: First note that the following is an invariant of a I-boosting run of the multi-branch procedure. max Wt < min Wt tEL(VW(T)) - mEN(VW(T)) 306 Y. Mansour and D. MeAl/ester The proof is a simple induction on ,-boosting tree growth using the fact that the procedure always expands a leaf node of maximal weight. We must now show that for every internal node m and every leaf ? we have that Wi ~ e(k -1)w m /(k -1) where k is the number of children of m. Note that if k = 2 then this reduces to Wi ~ Wm which follows from the above invariant. So we can assume without loss of generality that k > 2. Also, since e( k) / k < e( k - 1) / (k - 1), it suffices to show that Wi ~ e(k)wm/k. Let m be an internal node with k > 2 children and let T' be the tree at the time m was selected for expansion. Let Wi be the maximum weight of a leaf in the final tree T. By the definition of the acceptability condition, in the last s/2 iterations we are performing only binary branching. Each binary expansion reduces the index by at least , times the weight of the selected node. Since the sequence of nodes selected in the multi-branch procedure has non-increasing weights, we have that in any iteration the weight of the selected node is at least Wi . Since there are at least s/2 binary expansions after the expansion of m, each of which reduces I by at least ,Wi, we have that s,wd2 ~ I(T') so Wi ~ 2I(T')/(/s). The acceptability condition can be written as 2/(/s) ~ e(k)/(kIT'1) which now yields WI ~ I(T')e(k)/(kIT'I). But we have that I(T')/IT'I ~ Wm which now yields WI ~ e(k)wm/k as desired. 0 References [1] Leo Breiman, Jerome H. Friedman, Richard A. Olshen, and Charles J. Stone. Classification and Regression Trees. Wadsworth International Group, 1984. [2] Tom Dietterich, Michael Kearns and Yishay Mansour. Applying the Weak Learning Framework to understand and improve C4.5. In Proc. of Machine Learning, 96-104, 1996. [3] Yoav Freund. Boosting a weak learning algorithm by majority. Information and Computation, 121(2):256-285, 1995. [4] Yoav Freund and Robert E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. In Computational Learning Theory: Second European Conference, EuroCOLT '95, pages 23-37. SpringerVerlag, 1995. [5] Yoav Freund and Robert E. Schapire. Experiments with a new boosting algorithm. In Machine Learning: Proceedings of the Thirteenth International Conference, pages 148-156, 1996. [6] Michael Kearns and Yishay Mansour. On the boosting ability of top-down decision tree learning. In Proceedings of the Twenty-Eighth ACM Symposium on the Theory of Computing, pages 459-468,1996. [7] J. Ross Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann, 1993. [8] Robert E. Schapire. The strength of weak learnability. Machine Learning, 5(2):197-227, 1990.
1659 |@word seems:2 reduction:2 initial:2 att:1 selecting:4 existing:1 current:1 com:1 assigning:2 must:5 written:1 designed:1 greedy:1 leaf:33 selected:8 ith:1 argm:1 short:1 provides:1 boosting:24 node:27 ih1:1 constructed:1 symposium:1 prove:6 fitting:2 nondeterministic:1 roughly:2 isi:1 growing:1 multi:15 globally:3 eurocolt:1 lll:1 increasing:1 provided:2 bounded:1 qw:4 spends:1 nj:1 guarantee:1 quantitative:1 every:5 ptl:1 ti:3 expands:1 growth:26 qm:1 safety:1 mistake:1 plus:1 range:1 practical:1 practice:7 implement:1 procedure:24 confidence:1 get:2 applying:1 splitting:1 immediately:1 ltel:2 justification:3 pli:1 yishay:3 construction:2 target:8 pt:2 play:1 hypothesis:20 associate:1 element:2 satisfying:1 labeled:1 observed:1 ft:1 trade:2 argmaxl:1 tit:2 learner:8 acceptability:2 various:1 grown:2 leo:1 forced:1 effective:1 whose:3 quite:1 spend:2 say:2 florham:1 ability:1 think:1 final:2 sequence:1 twoway:1 inn:1 maximal:1 inserting:1 achieve:5 exploiting:1 produce:3 llet:1 measured:2 qt:6 progress:3 implies:6 quantify:1 differ:1 safe:3 attribute:5 vc:1 mcallester:2 sand:1 require:2 suffices:2 generalization:2 preliminary:1 pl:2 hold:2 sufficiently:1 ic:2 mapping:2 achieves:1 proc:1 iw:2 visited:1 ross:1 wl:2 tool:1 weighted:1 always:1 reaching:2 avoid:1 breiman:1 derived:1 improvement:3 ave:1 entire:1 relation:1 jq:2 selects:5 classification:1 denoted:1 fairly:2 wadsworth:1 equal:2 construct:3 never:1 argmaxt:1 park:2 richard:1 replaced:1 consisting:2 friedman:1 tree:88 desired:2 theoretical:3 instance:1 boolean:2 yoav:3 introducing:1 subset:2 learnability:1 considerably:1 st:6 fundamental:1 international:2 off:2 vm:2 eel:1 michael:2 quickly:1 w1:1 again:1 satisfied:2 ester:2 til:5 return:4 potential:1 satisfy:1 depends:1 try:1 view:1 lab:1 root:2 wm:6 start:1 option:5 maintains:1 formed:4 accuracy:1 kaufmann:1 yield:2 ofthe:1 weak:24 produced:1 multiplying:1 definition:3 proof:10 rational:2 proved:2 popular:2 adaboost:1 tom:1 done:1 generality:1 jerome:1 hand:1 replacing:1 quality:1 grows:2 dietterich:1 requiring:1 true:2 concept:1 hence:4 assigned:1 i2:7 ll:1 game:4 branching:30 essence:3 noted:2 steady:2 maintained:1 mutli:1 stone:1 presenting:1 tt:3 theoretic:1 performs:1 spending:1 charles:1 empirically:2 analog:1 he:3 meal:2 pmi:1 pl1:1 binary:9 success:1 vt:20 yi:3 exploited:1 morgan:1 kit:2 paradigm:1 ii:4 branch:23 reduces:3 stem:1 long:1 visit:1 basic:1 regression:1 iteration:9 achieved:1 thirteenth:1 wd2:1 cart:2 subject:2 member:1 seem:2 integer:1 vw:16 easy:2 wn:1 variety:1 fit:1 reassign:1 useful:1 amount:1 locally:4 schapire:3 sl:7 exist:2 popularity:1 write:1 prd:1 ist:1 begining:1 group:1 visitation:5 key:1 four:2 fraction:3 wand:1 run:8 extends:1 decision:19 acceptable:2 comparable:1 handicap:1 guaranteed:2 replaces:1 nontrivial:1 strength:1 constraint:1 min:2 performing:1 according:1 smaller:1 slightly:1 describes:1 wi:12 tw:2 making:1 hl:1 invariant:4 remains:4 count:5 fail:1 end:4 ffl:2 away:1 top:7 denotes:2 quinlan:1 exploit:2 lel:2 classical:3 ink:2 move:3 objective:1 question:1 quantity:3 already:1 strategy:2 majority:1 trivial:1 induction:3 assuming:1 index:7 optionally:1 ql:5 difficult:1 olshen:1 robert:3 statement:1 expense:1 motivates:1 unknown:1 perform:1 twenty:1 iti:22 finite:2 immediate:2 defining:1 wtll:1 mansour:13 david:1 introduced:1 pair:3 required:3 c4:5 adversary:10 usually:1 below:1 eighth:1 oft:3 program:1 max:1 improve:1 imply:1 categorical:1 naive:1 prior:1 freund:3 loss:1 expect:1 men:1 proportional:3 consistent:1 unfinished:1 land:1 course:2 repeat:1 last:1 allow:2 understand:1 template:1 taking:1 lrh:1 curve:1 dimension:1 c5:4 avoided:1 pruning:1 keep:1 global:1 continuous:1 iterative:1 tel:1 expansion:4 european:1 constructing:2 main:5 linearly:1 rh:4 incrementing:1 child:6 allowed:1 tl:6 fails:1 down:7 theorem:12 naively:1 exists:4 adding:1 nec:1 budget:5 entropy:1 forming:1 applies:1 satisfies:2 acm:1 itl:1 viewed:2 hard:1 springerverlag:1 specifically:1 reducing:2 wt:13 kearns:8 lemma:10 called:3 total:2 experimental:1 select:1 internal:11
721
166
712 A PROGRAMMABLE ANALOG NEURAL COMPUTER AND SIMULATOR Paul Mueller*, Jan Vander Spiegel, David Blackman*, Timothy Chiu, Thomas Clare, Joseph Dao, Christopher Donham, Tzu-pu Hsieh, Marc Loinaz *Dept.of Biochem. Biophys., Dept. of Electrical Engineering. University of Pennsylvania, Philadelphia Pa. ABSTRACT This report describes the design of a programmable general purpose analog neural computer and simulator. It is intended primarily for real-world real-time computations such as analysis of visual or acoustical patterns, robotics and the development of special purpose neural nets. The machine is scalable and composed of interconnected modules containing arrays of neurons, modifiable synapses and switches. It runs entirely in analog mode but connection architecture, synaptic gains and time constants as well as neuron parameters are set digitally. Each neuron has a limited number of inputs and can be connected to any but not all other neurons. For the determination of synaptic gains and the implementation of learning algorithms the neuron outputs are multiplexed, AID converted and stored in digital memory. Even at moderate size of 1()3 to IDS neurons computational speed is expected to exceed that of any current digital computer. OVERVIEW The machine described in this paper is intended to serve as a general purpose programmable neuron analog computer and simulator. Its architecture is loosely based on the cerebral cortex in the sense that there are separate neurons, axons and synapses and that each neuron can receive only a limited number of inputs. However, in contrast to the biological system, the connections can be modified by external control permitting exploration of different architectures in addition to adjustment of synaptic weights and neuron parameters. The general architecture of the computer is shown in Fig. 1. Themachinecontains large numbers of the following separate elements: neurons, synapses, routing switches and connection lines. Arrays of these elements are fabricated on VLSI chips which are mounted on planar chip carriers each of which forms a separate module. These modules are connected directly to neighboring modules. Neuron arrays are arranged in rows and columns and are surrounded by synaptic and axon arrays. A Programmable Analog Neural Computer and Simulator The machine runs entirely in analog mode. However, connection architectures, synaptic gains and neuron parameters such as thresholds and time constants are set by a digital computer. For determining synaptic weights in a learning mode, time segments of the outputs from neurons are multiplexed, digitized and stored in digital memory. The modular design allows expansion to any degree and at moderate to large size. to lOS neurons. operational speed would exceed that of any currently available i.e. digital computer. 103 I I I I ? I I I I ? I SWITCHES LINES D SYNAPSES NEURONS Figure 1. Layout and general architecture. The machine is composed of different modules shown here as squares. Each module contains on a VLSI chip an array of components (neurons. synapses or switches) and their control circuits. Our prototype design calls for 50 neuron modules for a total of 800 neurons each having 64 synapses. The insert shows the direction of data flow through the modules. Outputs from each neuron leave north and south and are routed through the switch modules east and west and into the synapse modules from north and south. They can also bypass the synapse modules north and south. Input to the neurons through the synapses is from east and west. Power and digital control lines run north and south. THE NEURON MODULES Each neuron chip contains 16 neurons, an analog multiplexer and control logic. (See Figs. 2 & 3.) Input-output relations of the neurons are idealized versions of a typical biological neuron. Each unit has an adjustable threshold (bias), an adjustable minimum output value at threshold and a maximum output (See Fig. 4). Output time constants are selected on the switch chips. The neuron is based on an earlier design which used discrete components (Mueller and Lazzaro, 1986). 713 714 Mueller, et al Inputs to each neuron come from synapse chips east and west (SIR, SIL), outputs (NO) go to switch chips north and south. Each neuron has a second input that sets the minimum output at threshold which is common for all neurons on the chip and selected through a separate synapse line. The threshold is set from one of the synapses connected to a fixed voltage. An analog multiplexer provides neuron output to a common line, OM, which connects to an AID converter. ~6 NJ IS CK 01\ ORI PHI2 ANALOG I\LL T I PLEXER CRO SILl SIL 2 SIR SIR I SILlS SIL I6 SIR IS SIR I6 ?I ?I NO. 2 NJ2 Figure 2. Block diagram of the neuron chip containing 16 neurons . .. . -- _ _ _ _~_ _ _ _ _ _ ~F _ _ _ _~ _ _ ? .... b-.. ~,r ~ "'1<1'" -. ,-~:""",,-- !l'r.:. 11,"" ~~ ........ ~ ~:..._ nil');~ Figure 3. Photograph of a test chip containing 5 neurons. A more recent version has only one output sign. A Programmable Analog Neural Computer and Simulator 4 o o 5 SUM OF INPUTS/VOLTS Figure 4. Transfer characteristic obtained from a neuron on the chip shown in Fig.3. Each unit has an adjustable threshold, Vt which was set here to 1.5V, a linear transfer region above threshold, an adjustable minimum output at threshold E x set to 1V and a maximum output, Emo. THE SYNAPSE MODULES Each synapse chip contains a 32 * 16 array of synapses. The synaptic gain of each synapse is set by serial input from the computer and is s tored at each synapse. Dynamic range of the synapse gains covers the range from 0 to 10 with 5 bit resolution, asixth bit determines the sign. The gains are implemented by current mirrors which scale the neuron output after it has been converted from a voltage to a current. The modifIable synapse designs reported in the literature use either analog or digital signals to set the gains (Schwartz, et. al., 1989, Raffel, et.al, 1987, Alspector and Allen, 1987). We chose the latter method because of its greater reproducibility and because direct analog setting of the gains from the neuron outputs would require a prior knowledge of and commitment to a particular learning algorithm. Layout and perfonnance of the synapse module are shown in Figs. 5-7. As seen in Fig. 7a, the synaptic transfer function is linear from 0 to 4 V. The use of current mirrors pennits arbitrary scaling of the synaptic gains (weights) with trade off between range and resolution limited to 5 bits. Our current design calls for a minimum gain of 1/32 and a maximum of 10. The lower end of the dynamic range is detennined by the number of possible inputs per neuron which when active should not drive the neuron output to its limit, whereas the high gain values are needed in situations where a single or very few synapses must be effective such as in the copying of activity from one neuron to another or for veto inhibition. The digital nature of the synaptic gain control does not allow straight forward implementation of a logarithmic gain scale. Fig. 7b. shows two possible relations between digital code and synaptic gain. In one case the total gain is the sum of 5 individual gains each controlled by one bit. This leads inevitably to jumps in the gain curve. In a second case a linear 3 bit gain is multiplied by four different constants 715 716 Mueller, et al controlled by the 4th and 5th bit. This scheme affords a better approximation to a logarithmic scale. So far we have implemented only the first scheme. Although the resolution of an individual synapse is limited to 5 bits, several synapses driven by one neuron can be combined through switching, permitting greater resolution and dynamic range. o o o NI32 DATA Figure S. Diagram of the synapse module. Each synapse gain is set by a 5 bit word stored in local memory. The memory is implemented as a quasi dynamic shift register that reads the gain data during the programming phase. Voltage to current converters transform the neuron output (N!) into a current. I Conv are current mirrors that scale the currents with 5 bit resolution. The weighted currents are summed on a common line to the neuron input (SO). Figure 6. Photograph of a synapse test chip. A Programmable Analog Neural Computer and Simulator AJOT""'T'--------.,..., B 1 iEm Weight-10 o ~u ~ ~ :J t 10 o .. o~---- o 1 Welght-1/32 ~~~~~~ 2 J Input (Volts) -J+--+-----~~~-----~-+-----+_~ o 4 8 12 18 20 24 28 J2 Code Figure 7a. Synapse transfer characteristics for three different settings. The data were obtained from the chip shown in Fig. 6. b. Digital code vs. synaptic gain, squares are current design, triangles use a two bit exponent. THE SWITCH MODULES The switch modules serve to route the signals between neurons and allow changes to the connection architecture. Each module contains a 32*32 cross point array of analog switches which are set by serial digital input. There is also a set of serial switches that can disconnect input and output lines. In addition to switches the modules contain circuits which control the time constants of the synapse transfer function (see Figs. 8 & 9). The switch perfonnance is summarized in Table 1. UIN UIN Lilli RIlIl Lilli RIll L3B R30 L31 R3t UN+RIN DIN RIN I BIT f'E.fl.Gty L I N - 0 - - RIN 2 BIT f'E.1lfJ/.Y/ LOGIC 0lIl111 0lIl1 031'1 031 Figure 8. Diagram of switching fabric. Squares and circles represent switch cells which connect the horizontal and vertical connectors or cut the conductors. The units labeled T represent adjustable time constants. 717 718 Mueller, et al TABLE 1. Switch Chip Performance Process On resistance Off resistance 3uCMOS <3 KOhm > 1 TOhm Input capacitance Array download time Memory/switch size < IpF 2us 75u x 90u ADJUSTMENT OF SYNAPTIC TIME CONSTANTS For the analysis or generation of temporal patterns as they occur in motion or speech, adjustable time constants of synaptic transfer must be available (Mueller, 1988). Low pass filtering of the input signal to the synapse with 4 bit control of the time constant over a range of 5 to 500 ms is sufficient to deal with real world data. By com bining the low passed input with a direct input of opposite sign, both originating from the same neuron, the typical "ON" and "OFF" responses which serve as measures of time after beginning and end of events and are common in biological systems can be obtained. Several designs are being considered for implementing the variable low pass filter. Since not all synapses need to have this feature, the circuit will be placed on only a limited number of lines on the switch chip. PACKAGING All chips are mounted on identical quad surface mount carriers. Input and output lines are arranged at right angles with identical leads on opposite sides. The chip carriers are mounted on boards. SOFTWARE CONTROL AND OPERATION Connections, synaptic gains and time constants are set from the central computer either manually or from libraries containing connection architectures for specific tasks. Eventually we envision developing a macro language that would generate subsystems and link them into a larger architecture. Examples are feature specific receptor fields, temporal pattern analyzers, or circuits for motion control. The connection routing is done under graphic control or through routing routines as they are used in circuit board design. A Programmable Analog Neural Computer and Simulator The primary areas of application include real-world real-time or compressed time pattern analysis, robotics, the design of dedicated neural circuits and the exploration of different learning algorithms. Input to the machine can come from sensory transducer arrays such as an electronic retina, cochlea (Mead, 1989) or tactile sensors. For other computational tasks, input is provided by the central digital computer through activation of selected neuron populations via threshold control. It might seem that the limited number of inputs per neuron restricts the computations performed by anyone neuron. However the results obtained by one neuron can be copied through a unity gain synapse to another neuron which receives the appropriate additional inputs. In performance mode the machine could exceed by orders of magnitude the computational speed of any currently available digital computer. A rough estimate of attainable speed can be made as follows: A network with 103 neurons each receiving 100 inputs with synaptic transfer time constants ranging from 1 ms to 1 s, can be described by 103 simultaneous differential equations. Assuming an average step length of 10 us and 10 iterations per step, real time numerical solutions of this system on a digital machine would require approximately 1011 FLOPS. Microsecond time constants and the computation of threshold non-linearities would require a computational speed equivalent to > 1012 FLOPS on a digital computer and this seems a reasonable estimate of the computational power of our machine. Furthermore, in contrast to digital mUltiprocessors, computational power would scale linearly with the number of neurons and connections. Acknowledgements Supported by grants from ONR (NOOOI4-89-J-1249) and NSF (EET 166685). References Alspector, J., Allen, R.B. A neuromorphic VLSI learning system. Advanced research in VLSI. Proceedings of the 1987 Stanford Conference. (1987). Mead, C. Analog VLSI and Neural Systems. Addison Wesley, Reading, Ma (1989). Mueller, P. Computation of temporal Pattern Primitives in a Neural Net for Speech Recognition. InternationalNeuralNetworkSociety. FirstAnnualMeeting,Boston Ma.? (1988). Mueller, P., Lazzaro, J. A Machine for Neural Computation of Acoustical Patterns. AlP Conference Proceedings. 151:321-326, (1986). Raffel, J.I., Mann, J.R., Berger, R., Soares, A.M., Gilbert, S., A Generic Architecture for Wafer-Scale neuromorphic Systems. IEEE First International Conference on Neural Networks. San Diego, CA. (1987). Schwartz, D., Howard, R., Hubbard, W., A Programmable Analog Neural Network Chip, J. of Solid State Circuits, (to be published). 719
166 |@word version:2 seems:1 donham:1 bining:1 hsieh:1 attainable:1 solid:1 contains:4 envision:1 current:11 com:1 activation:1 must:2 numerical:1 v:1 selected:3 beginning:1 provides:1 direct:2 differential:1 transducer:1 expected:1 alspector:2 simulator:7 quad:1 conv:1 provided:1 linearity:1 circuit:7 pennits:1 fabricated:1 nj:1 temporal:3 schwartz:2 control:11 unit:3 grant:1 carrier:3 engineering:1 local:1 limit:1 switching:2 receptor:1 id:1 mount:1 mead:2 approximately:1 might:1 chose:1 limited:6 sill:2 range:6 block:1 jan:1 area:1 word:1 subsystem:1 gilbert:1 equivalent:1 raffel:2 layout:2 go:1 primitive:1 resolution:5 array:9 population:1 diego:1 programming:1 pa:1 element:2 recognition:1 cut:1 labeled:1 module:19 electrical:1 region:1 connected:3 cro:1 trade:1 digitally:1 dynamic:4 segment:1 serve:3 rin:3 triangle:1 connector:1 chip:19 fabric:1 effective:1 modular:1 larger:1 stanford:1 compressed:1 spiegel:1 transform:1 uin:2 net:2 interconnected:1 commitment:1 macro:1 neighboring:1 j2:1 reproducibility:1 detennined:1 los:1 leave:1 implemented:3 come:2 direction:1 lfj:1 filter:1 exploration:2 routing:3 alp:1 implementing:1 mann:1 require:3 biological:3 insert:1 considered:1 purpose:3 currently:2 hubbard:1 weighted:1 rough:1 sensor:1 modified:1 ck:1 voltage:3 blackman:1 contrast:2 sense:1 mueller:8 multiprocessor:1 vlsi:5 relation:2 originating:1 quasi:1 exponent:1 development:1 special:1 summed:1 field:1 having:1 manually:1 identical:2 dao:1 report:1 primarily:1 few:1 retina:1 composed:2 individual:2 intended:2 connects:1 phase:1 iem:1 multiplexed:2 perfonnance:2 loosely:1 circle:1 column:1 earlier:1 cover:1 neuromorphic:2 graphic:1 stored:3 reported:1 connect:1 combined:1 international:1 off:3 receiving:1 central:2 tzu:1 containing:4 external:1 clare:1 multiplexer:2 converted:2 summarized:1 north:5 disconnect:1 register:1 idealized:1 performed:1 ori:1 om:1 square:3 characteristic:2 drive:1 straight:1 published:1 simultaneous:1 synapsis:12 synaptic:16 gain:23 noooi4:1 knowledge:1 routine:1 wesley:1 planar:1 response:1 synapse:19 arranged:2 done:1 furthermore:1 l31:1 receives:1 horizontal:1 christopher:1 mode:4 ipf:1 contain:1 din:1 read:1 volt:2 deal:1 ll:1 during:1 m:2 allen:2 motion:2 dedicated:1 ranging:1 common:4 overview:1 cerebral:1 analog:17 i6:2 analyzer:1 language:1 cortex:1 surface:1 inhibition:1 pu:1 biochem:1 soares:1 recent:1 moderate:2 driven:1 route:1 onr:1 vt:1 seen:1 minimum:4 greater:2 additional:1 signal:3 determination:1 cross:1 serial:3 permitting:2 controlled:2 packaging:1 scalable:1 cochlea:1 represent:2 iteration:1 robotics:2 cell:1 receive:1 whereas:1 addition:2 diagram:3 south:5 veto:1 flow:1 nj2:1 seem:1 call:2 exceed:3 switch:17 pennsylvania:1 architecture:10 converter:2 opposite:2 prototype:1 shift:1 passed:1 routed:1 tactile:1 resistance:2 speech:2 lazzaro:2 programmable:8 generate:1 affords:1 restricts:1 nsf:1 sign:3 per:3 modifiable:2 discrete:1 wafer:1 four:1 threshold:10 sum:2 run:3 angle:1 reasonable:1 electronic:1 scaling:1 bit:13 entirely:2 fl:1 copied:1 activity:1 occur:1 software:1 speed:5 anyone:1 developing:1 describes:1 unity:1 joseph:1 equation:1 emo:1 eventually:1 needed:1 addison:1 end:2 available:3 operation:1 multiplied:1 appropriate:1 generic:1 thomas:1 include:1 capacitance:1 primary:1 separate:4 link:1 acoustical:2 lil1:1 assuming:1 code:3 length:1 copying:1 berger:1 design:10 implementation:2 adjustable:6 vertical:1 neuron:52 howard:1 inevitably:1 situation:1 flop:2 digitized:1 arbitrary:1 vander:1 download:1 david:1 connection:9 pattern:6 reading:1 memory:5 power:3 event:1 advanced:1 scheme:2 library:1 philadelphia:1 prior:1 literature:1 acknowledgement:1 determining:1 sir:5 generation:1 mounted:3 filtering:1 sil:3 digital:16 degree:1 sufficient:1 bypass:1 surrounded:1 row:1 placed:1 supported:1 bias:1 allow:2 side:1 curve:1 world:3 sensory:1 forward:1 made:1 jump:1 san:1 far:1 eet:1 logic:2 active:1 un:1 table:2 nature:1 transfer:7 ca:1 operational:1 expansion:1 marc:1 linearly:1 paul:1 fig:9 west:3 board:2 aid:2 axon:2 specific:2 mirror:3 magnitude:1 biophys:1 boston:1 logarithmic:2 timothy:1 photograph:2 visual:1 adjustment:2 determines:1 ma:2 lilli:2 microsecond:1 change:1 typical:2 rill:1 conductor:1 total:2 nil:1 pas:2 east:3 chiu:1 latter:1 dept:2
722
1,660
Reconstruction of Sequential Data with Probabilistic Models and Continuity Constraints Miguel A. Carreira-Perpifian Dept. of Computer Science, University of Sheffield, UK miguel@dcs.shefac.uk Abstract We consider the problem of reconstructing a temporal discrete sequence of multidimensional real vectors when part of the data is missing, under the assumption that the sequence was generated by a continuous process. A particular case of this problem is multivariate regression, which is very difficult when the underlying mapping is one-to-many. We propose an algorithm based on a joint probability model of the variables of interest, implemented using a nonlinear latent variable model. Each point in the sequence is potentially reconstructed as any of the modes of the conditional distribution of the missing variables given the present variables (computed using an exhaustive mode search in a Gaussian mixture). Mode selection is determined by a dynamic programming search that minimises a geometric measure of the reconstructed sequence, derived from continuity constraints. We illustrate the algorithm with a toy example and apply it to a real-world inverse problem, the acoustic-toarticulatory mapping. The results show that the algorithm outperforms conditional mean imputation and multilayer perceptrons. 1 Definition of the problem Consider a mobile point following a continuous trajectory in a subset of ]RD. Imagine that it is possible to obtain a finite number of measurements of the position of the point. Suppose that these measurements are corrupted by noise and that sometimes part of, or all, the variables are missing. The problem considered here is to reconstruct the sequence from the part of it which is observed. In the particular case where the present variables and the missing ones are the same for every point, the problem is one of multivariate regression. If the pattern of missing variables is more general, the problem is one of missing data reconstruction. Consider the problem of regression. If the present variables uniquely identify the missing ones at every point of the data set, the problem can be adequately solved by a universal function approximator, such as a multilayer perceptron. In a probabilistic framework, the conditional mean of the missing variables given the present ones will minimise the average squared reconstruction error [3]. However, if the underlying mapping is one-to-many, there will be regions in the space for which the present variables do not identify uniquely the missing ones. In this case, the conditional mean mapping will fail, since it will give a compromise value-an average of the correct ones. Inverse problems, where the inverse Probabilistic Sequential Data Reconstruction 415 of a mapping is one-to-many, are of this type. They include the acoustic-to-articulatory mapping in speech [15], where different vocal tract shapes may produce the same acoustic signal, or the robot arm problem [2], where different configurations of the joint angles may place the hand in the same position. In some situations, data reconstruction is a means to some other objective, such as classification or inference. Here, we deal solely with data reconstruction of temporally continuous sequences according to the squared error. Our algorithm does not apply for data sets that either lack continuity (e.g. discrete variables) or have lost it (e.g. due to undersampling or shuffling). We follow a statistical learning approach: we attempt to reconstruct the sequence by learning the mapping from a training set drawn from the probability distribution of the data, rather than by solving a physical model of the system. Our algorithm can be described briefly as follows. First, a joint density model of the data is learned in an unsupervised way from a sample of the datal . Then, pointwise reconstruction is achieved by computing all the modes of the conditional distribution of the missing variables given the present ones at the current point. In principle, any of these modes is potentially a plausible reconstruction. When reconstructing a sequence, we repeat this mode search for every point in the sequence, and then find the combination of modes that minimises a geometric sequence measure, using dynamic programming. The sequence measure is derived from local continuity constraints, e.g. the curve length. The algorithm is detailed in ?2 to ?4. We illustrate it with a 2D toy problem in ?5 and apply it to an acoustic-to-articulatory-like problem in ?6. ?7 discusses the results and compares the approach with previous work. Our notation is as follows. We represent the observed variables in vector form as t = (tl' ... , t D) E ~D. A data set (possibly a temporal sequence) is represented as {t n } ~=l . Groups of variables are represented by sets of indices I, J E {I, ... , D}, so that if I = {I, 7, 3}, then tI = (tlt7t3). 2 Joint generative modelling using latent variables Our starting point is a joint probability model of the observed variables p( t). From it, we can compute conditional distributions of the form p( t..71 tI) and, by picking representative points, derive a (multivalued) mapping tI ~ t..7. Thus, contrarily to other approaches, e.g. [6], we adopt multiple pointwise imputation. In ?4 we show how to obtain a single reconstructed sequence of points. Although density estimation requires more parameters than mapping approximation, it has a fundamental advantage [6]: the density model represents the relation between any variables, which allows to choose any missing/present variable combination. A mapping approximator treats asymmetrically some variables as inputs (present) and the rest as outputs (missing) and can't easily deal with other relations. The existence of functional relationships (even one-to-many) between the observed variables indicates that the data must span a low-dimensional manifold in the data space. This suggests the use of latent variable models for modelling the joint density. However, it is possible to use other kinds of density models. In latent variable modelling the assumption is that the observed high-dimensional data t is generated from an underlying low-dimensional process defined by a small number L of latent variables x = (Xl, ... , xL) [1] . The latent variables are mapped by a fixed I In our examples we only use complete training data (i.e., with no missing data), but it is perfectly possible to estimate a probability model with incomplete training data by using an EM algorithm [6]. M 416 A. Carreira-Perpiful.n transformation into a D-dimensional data space and noise is added there. A particular model is specified by three parametric elements: a prior distribution in latent space p(x), a smooth mapping f from latent space to data space and a noise model in data space p(tlx). Marginalising the joint probability density function p(t, x) over the latent space gives the distribution in data space, p(t). Given an observed sample in data space {t n };;=l' a parameter estimate can be found by maximising the log-likelihood, typically using an EM algorithm. We consider the following latent variable models, both of which allow easy computation of conditional distributions of the form p( tJ ItI ): Factor analysis [1], in which the mapping is linear, the prior in latent space is unit Gaussian and the noise model is diagonal Gaussian. The density in data space is then Gaussian with a constrained covariance matrix. We use it as a baseline for comparison with more sophisticated models. The generative topographic mapping (GTM) [4] is a nonlinear latent variable model, where the mapping is a generalised linear model, the prior in latent space is discrete uniform and the noise model is isotropic Gaussian. The density in data space is then a constrained mixture of isotropic Gaussians. In latent variable models that sample the latent space prior distribution (like GTM), the mixture centroids in data space (associated to the latent space samples) are not trainable parameters. We can then improve the density model at a higher computational cost with no generalisation loss by increasing the number of mixture components. Note that the number of components required will depend exponentially on the intrinsic dimensionality of the data (ideally coincident with that of the latent space, L) and not on the observed one, D. 3 Exhaustive mode finding Given a conditional distribution p(tJltI), we consider all its modes as plausible predictions for tJ. This requires an exhaustive mode search in the space of t J . For Gaussian mixtures, we do this by using a maximisation algorithm starting from each centroid2 , such as a fixed-point iteration or gradient ascent combined with quadratic optimisation [5]. In the particular case where all variables are missing, rather than performing a mode search, we return as predictions all the component centroids. It is also possible to obtain error bars at each mode by locally approximating the density function by a normal distribution. However, if the dimensionality of tJ is high, the error bars become very wide due to the curse of the dimensionality. An advantage of multiple pointwise imputation is the easy incorporation of extra constraints on the missing variables. Such constraints might include keeping only those modes that lie in an interval dependent on the present variables [8] or discarding low-probability (spurious) modes-which speeds up the reconstruction algorithm and may make it more robust. A faster way to generate representative points of p(tJltI) is simply to draw a fixed number of samples from it-which may also give robustness to poor density models. However, in practice this resulted in a higher reconstruction error. 4 Continuity constraints and dynamic programming (D.P) search Application of the exhaustive mode search to the conditional distribution at every point of the sequence produces one or more candidate reconstructions per point. To select a 2 Actually, given a value of tz, most centroids have negligible posterior probability and can be removed from the mixture with practically no loss of accuracy. Thus, a large number of mixture components may be used without deteriorating excessively the computational efficiency. 417 Probabilistic Sequential Data Reconstrnction trajectory factor an. Average squared reconstruction error mean dpmode ..., N Mi ssing pattern h tl 0 tl or t2 -2 10% 50% 90% -4 Factor analysis 3.8902 4.3226 4.2020 1.0983 6.2914 21.4942 MLP" 0.2046 2.5126 - - GTM mean dpmode cmode 0.2044 2.4224 1.2963 0.3970 4.6530 20.7877 0.2168 0 .0522 0.1305 0.0253 0. 1176 2 .2261 0.2168 0.0522 0. 1305 0.0251 0.0771 0.0643 aThe MLP cannot be applied to varying patterns of missing data. -6 -6 -4 -2 tl Table 1: Trajectory reconstruction for a 2D problem. The table gives the average squared reconstruction error when t2 is missing (row 1), tl is missing (row 2), exactly one variable per point is missing at random (row 3) or a percentage of the values are missing at random (rows 4-6). The graph shows the reconstructed trajectory when tl is missing: factor analysis (straight, dotted line), mean (thick, dashed), dpmode (superimposed on the trajectory). single reconstructed sequence, we define a local continuity constraint: consecutive points in time should also lie nearby in data space. That is, if 8 is some suitable distance in JR.D, 8 (tn, tn+ 1) should be small. Then we define a global geometric measure ~ for a sequence {t n };;=1 as ~ ({t n };;=I) ~f ~ '2:.::118 (tn, tn+t). We take 8 as the Euclidean distance, so becomes simply the length of the sequence (considered as a polygonal line). Finding the sequence of modes with minimal ~ is efficiently achieved by dynamic programming. 5 Results with a toy problem To illustrate the algorithm, we generated a 2D data set from the curve (tl, t2) = (x, x + 3 sin(x)) for x E [-211',211'], with normal isotropic noise (standard deviation 0.2) added. Thus, the mapping tl -+ t2 is one-to-one but the inverse one, t2 -+ tl, is multivalued. One-dimensional factor analysis (6 parameters) and GTM models (21 parameters) were estimated from a 1000-point sample, as well as two 48-hidden-unit multilayer perceptrons (98 parameters), one for each mapping. For GTM we tried several strategies to select points from the conditional distribution: mean (the conditional mean), dpmode (the mode selected by dynamic programming) and cmode (the closest mode to the actual value of the missing variable). The cmode, unknown in practice, is used here to compute a lower bound on the performance of any mode-based strategy. Other strategies, such as picking the global mode, a random mode or using a local (greedy) search instead of dynamic programming, gave worse results than the dpmode. Table 1 shows the results for reconstructing a IOO-point trajectory. The nonlinear nature of the problem causes factor analysis to break down in all cases. For the one-to-one mapping case (t2 missing) all the other methods perform well and recover the original trajectory, with mean attaining the lowest error, as predicted by the theory3. For the one-to-many case (tl missing, see fig .), both the MLP and the mean are unable to track more than one branch of the mapping, but the dpmode still recovers the original mapping. For random missing 3 A combined strategy could retain the optimality of the mean in the one-to-one case and the advantage of the modes in the one-to-many case, by choosing the conditional mean (rather than the mode) when the conditional distribution is unimodal, and all the modes otherwise. M 418 Missing pattern PLP EPG 10% 50% blocks Factor analysis 0.9165 3.7177 0.2046 1.1285 0.1950 A. Carreira-Perpinan GTM mean dpmode cmode 0.6217 2.3729 0.0947 0.7540 0.1669 0.6250 2.0613 0.0903 0.6527 0.1005 0.4587 1.0538 0.0841 0.6023 0.0925 Table 2: Average squared reconstruction error for an utterance. The last row corresponds to a missing pattern of square blocks totalling 10% of the utterance. patterns4 , the dpmode is able to cope well with high amounts of missing data. The consistently low error of the cmode shows that the modes contain important information about the possible options to predict the missing values. The performance of the dpmode, close to that of the cmode even for large amounts of missing data, shows that application of the continuity constraint allows to recover that information. 6 Results with real speech data We report a preliminary experiment using acoustic and e1ectropalatographic (EPG) data5 for the utterance "Put your hat on the hatrack and your coat in the cupboard" (speaker FG) from the ACCOR database [10]. 12th-order perceptual linear prediction coefficients [7] plus the log-energy were computed at 200 Hz from its acoustic waveform. The EPG data consisted of 62-bit frames sampled at 200 Hz, which we consider as 62-dimensional vectors of real numbers. No further preprocessing of the data was carried out. Thus, the resulting sequence consisted of over 600 75-dimensional real vectors. We constructed a training set by picking, in random order, 80% of these vectors. The whole utterance was used for the reconstruction test. We trained two density models: a 9-dimensional factor analysis (825 parameters) and a two-dimensional 6 GTM (3676 parameters) with a 20 x 20 grid (resulting in a mixture of 400 isotropic Gaussians in the 75-dimensional data space). Table 2 confirms again that the linear method (factor analysis) fares worst (despite its use of a latent space of dimension L = 9). The dpmode attains almost always a lower error than the conditional mean, with up to a 40% improvement (the larger the higher the amount of missing data). When a shuffled version of the utterance (thus having lost its continuity) was reconstructed, the error of the dpmode was consistently higher than that of the mean, indicating that the application of the continuity constraint was responsible for the error decrease. 7 Discussion Using a joint probability model allows flexible construction of predictive distributions for the missing data: varying patterns of missing data and multiple pointwise imputations are possible, as opposed to standard function approximators. We have shown that the modes of the conditional distribution of the missing variables given the present ones are potentially 4Note that the nature of the missing pattern (missing at random, missing completely at random, etc. [9]) does not matter for reconstruction-although it does for estimation. 5 An EPG datum is the (binary) contact pattern between the tongue and the palate at selected locations in the latter. Note that it is an incomplete articulatory representation of speech. 6 A latent space of 2 dimensions is clearly too low for this data, but the computational complexity of GTM prevents the use of a higher one. Still, its nonlinear character compensates partly for this. Probabilistic Sequential Data Reconstruction 419 plausible reconstructions of the missing values, and that the application of local continuity constraints-when they hold-can help to recover the actually plausible ones. Previous work The key aspects of our approach are the use of a joint density model (learnt in an unsupervised way), the exhaustive mode search, the definition of a geometric trajectory measure derived from continuity constraints and its implementation by dynamic programming. Several of these ideas have been applied earlier in the literature, which we review briefly. The use of the joint density model for prediction is the basis of the statistical technique of multiple imputation [9]. Here, several versions of the complete data set are generated from the appropriate conditional distributions, analysed by standard complete-data methods and the results combined to produce inferences that incorporate missing-data uncertainty. Ghahramani and Jordan [6] also proposed the use of the joint density model to generate a single estimate of the missing variables and applied it to a classification problem. Conditional distributions have been approximated by MLPs rather than by density estimation [16], but this lacks flexibility to varying patterns of missing data and requires an extra model of the input variables distribution (unless assumed uniform). Rohwer and van der Rest [12] introduce a cost function with a description length interpretation whose minimum is approximated by the densest mode of a distribution. A neural network trained with this cost function can learn one branch of a multivariate mapping, but is unable to select other branches which may be correct at a given time. Continuity constraints implemented via dynamic programming have been used for the acoustic-to-articulatory mapping problem [15]. Reasonable results (better than using an MLP to approximate the mapping) can be obtained using a large codebook of acoustic and articulatory vectors. Rahim et al. [11] achieve similar quality with much less computational requirements using an assembly of MLPs, each one trained in a different area of the acoustic-articulatory space, to locally approximate the mapping. However, clustering the space is heuristic (with no guarantee that the mapping is one-to-one in each region) and training the assembly is difficult. It also lacks flexibility to varying missingness patterns. A number of trajectory measures have been used in the robot arm problem literature [2] and minimised by dynamic programming, such as the energy, torque, acceleration, jerk, etc. Temporal modelling It is important to remark that our approach does not attempt to model the temporal evolution of the system. The joint probability model is estimated statically. The temporal aspect of the data appears indirectly and a posteriori through the application of the continuity constraints to select a trajectory? In this respect, our approach differs from that of dynamical systems or from models based in Markovian assumptions, such as hidden Markov models or other trajectory models [13, 14]. However, the fact that the duration or speed of the trajectory plays no role in the algorithm may make it invariant to time warping (e.g. robust to fast/slow speech styles). Choice of density model The fact that the modes are a key aspect of our approach make it sensitive to the density model. With finite mixtures, spurious modes can appear as ripple superimposed on the density function in regions where the mixture components are sparsely distributed and have little interaction. Such modes can lead the DP search to a wrong trajectory. Possible solutions are to improve the density model (perhaps by increasing the number of components, see ?2, or by regularisation), to smooth the conditional distribution or to look for bumps (regions of high probability mass) instead of modes. 7However, the method may be derived by assuming a distribution over the whole sequence with a normal, Markovian dependence between adjacent frames. M. 420 A. Carreira-Perpifuin Computational cost The DP search has complexity O(N M2), where M is an average of the number of modes per sequence point and N the number of points in the sequence. In our experiments M is usually small and the DP search is fast even for long sequences. The bottleneck of the reconstruction part of the algorithm is obtaining the modes of the conditional distribution for every point in the sequence when there are many missing variables. Further work We envisage more thorough experiments using data from the Wisconsin X-ray microbeam database and comparing with recurrent MLPs or an MLP committee, which may be more suitable for multi valued mappings. Extensions of our algorithm include different geometric measures (e.g. curvature-based rather than length-based), different strategies for multiple pointwise imputation (e.g. bump searching) or multidimensional constraints (e.g. temporal and spatial). Other practical applications include audiovisual mappings for speech, hippocampal place cell reconstruction and wind vector retrieval from scatterometer data. Acknowledgments We thank Steve Renals for useful conversations and for comments about this paper. References [1] D. J. Bartholomew. Latent Variable Models and Factor Analysis. Charles Griffin & Company Ltd., London, 1987. [2] N. Bernstein. The Coordination and Regulation 0/ Movements. Pergamon, Oxford, 1967. [3] C. M. Bishop. Neural Networks/or Pattern Recognition. Oxford University Press, 1995. [4] C. M. Bishop, M. Svensen, and C. K. I. Williams. GTM: The generative topographic mapping. Neural Computation, 10(1):215-234, Jan. 1998. [5] M. A. Carreira-Perpifian. Mode-finding in Gaussian mixtures. Technical Report CS-99-03, Dept. of Computer Science, University of Sheffield, UK, Mar. 1999. Available online at http://vvv.dcs.shef.ac.uk/-miguel/papers/cs-99-03.html. [6] Z. Ghahramani and M. I. Jordan. Supervised learning from incomplete data via an EM approach. In NIPS 6, pages 120-127,1994. [7] H. Hermansky. Perceptual linear predictive (PLP) analysis of speech. 1. Acoustic Soc. Amer., 87(4):1738-1752, Apr. 1990. [8] L. Josifovski, M. Cooke, P. Green, and A. Vizinho. State based imputation of missing data for robust speech recognition and speech enhancement. In Proc. Eurospeech 99. pages 2837-2840, 1999. [9] R. 1. A. Little and D. B. Rubin. Statistical Analysis with Missing Data. John Wiley & Sons, New York, London, Sydney, 1987. [10] A. Marchal and W. J. Hardcastle. ACCOR: Instrumentation and database for the cross-language study of coarticulation. Language and Speech, 36(2, 3): 137-153, 1993. [11] M. G. Rahim, C. C. Goodyear, W. B. Kleijn, J. Schroeter, and M. M. Sondhi . On the use of neural networks in articulatory speech synthesis. 1. Acoustic Soc. Amer., 93(2): 1109-1121, Feb. 1993. [12] R. Rohwer and J. C. van der Rest. Minimum description length, regularization, and multi modal data. Neural Computation, 8(3):595-609, Apr. 1996. [13] S. Roweis. Constrained hidden Markov models. In NIPS 12 (this volume), 2000. [14] L. K. Saul and M. G . Rahim. Markov processes on curves for automatic speech recognition. In NIPS 11, pages 751-757, 1999. [15] 1. Schroeter and M. M. Sondhi. Techniques for estimating vocal-tract shapes from the speech signal. IEEE Trans. Speech and Audio Process., 2(1): 133-150, Jan. 1994. [16] V. Tresp, R. Neuneier, and S. Ahmad. Efficient methods for dealing with missing data in supervised learning. In NiPS 7, pages 689-696, 1995.
1660 |@word version:2 briefly:2 confirms:1 tried:1 covariance:1 configuration:1 outperforms:1 current:1 comparing:1 deteriorating:1 neuneier:1 analysed:1 must:1 john:1 shape:2 generative:3 selected:2 greedy:1 isotropic:4 codebook:1 location:1 constructed:1 become:1 ray:1 introduce:1 scatterometer:1 multi:2 torque:1 audiovisual:1 company:1 actual:1 curse:1 little:2 increasing:2 becomes:1 estimating:1 underlying:3 notation:1 mass:1 lowest:1 kind:1 finding:3 transformation:1 guarantee:1 temporal:6 thorough:1 every:5 multidimensional:2 ti:3 rahim:3 exactly:1 wrong:1 uk:4 unit:2 appear:1 generalised:1 negligible:1 local:4 treat:1 despite:1 oxford:2 solely:1 might:1 plus:1 suggests:1 josifovski:1 practical:1 responsible:1 acknowledgment:1 lost:2 maximisation:1 practice:2 block:2 differs:1 jan:2 area:1 universal:1 epg:4 vocal:2 cannot:1 close:1 selection:1 put:1 missing:45 williams:1 starting:2 duration:1 m2:1 searching:1 imagine:1 suppose:1 construction:1 play:1 densest:1 programming:9 element:1 approximated:2 recognition:3 sparsely:1 database:3 observed:7 role:1 solved:1 worst:1 region:4 ssing:1 removed:1 decrease:1 movement:1 ahmad:1 complexity:2 ideally:1 dynamic:9 trained:3 depend:1 solving:1 compromise:1 predictive:2 efficiency:1 completely:1 basis:1 easily:1 joint:12 sondhi:2 represented:2 gtm:9 fast:2 london:2 choosing:1 exhaustive:5 whose:1 heuristic:1 larger:1 plausible:4 valued:1 reconstruct:2 otherwise:1 compensates:1 topographic:2 kleijn:1 envisage:1 online:1 sequence:24 advantage:3 reconstruction:21 propose:1 interaction:1 renals:1 flexibility:2 achieve:1 roweis:1 description:2 enhancement:1 requirement:1 ripple:1 produce:3 tract:2 help:1 illustrate:3 derive:1 recurrent:1 svensen:1 miguel:3 minimises:2 ac:1 soc:2 implemented:2 predicted:1 c:2 sydney:1 waveform:1 thick:1 coarticulation:1 correct:2 preliminary:1 extension:1 hold:1 practically:1 considered:2 normal:3 mapping:27 predict:1 bump:2 adopt:1 consecutive:1 estimation:3 proc:1 coordination:1 sensitive:1 clearly:1 gaussian:7 always:1 rather:5 mobile:1 varying:4 derived:4 improvement:1 consistently:2 modelling:4 indicates:1 likelihood:1 superimposed:2 centroid:3 baseline:1 attains:1 posteriori:1 inference:2 dependent:1 typically:1 hidden:3 spurious:2 relation:2 classification:2 flexible:1 html:1 constrained:3 spatial:1 having:1 represents:1 look:1 unsupervised:2 hermansky:1 t2:6 report:2 resulted:1 attempt:2 tlx:1 interest:1 mlp:5 mixture:11 tj:3 articulatory:7 coat:1 unless:1 incomplete:3 euclidean:1 minimal:1 tongue:1 earlier:1 markovian:2 cost:4 deviation:1 subset:1 uniform:2 too:1 eurospeech:1 corrupted:1 learnt:1 combined:3 density:20 fundamental:1 retain:1 probabilistic:5 minimised:1 picking:3 synthesis:1 squared:5 again:1 opposed:1 choose:1 possibly:1 worse:1 tz:1 style:1 return:1 toy:3 attaining:1 coefficient:1 matter:1 break:1 wind:1 recover:3 option:1 mlps:3 square:1 accuracy:1 efficiently:1 identify:2 trajectory:13 straight:1 definition:2 rohwer:2 palate:1 energy:2 associated:1 mi:1 recovers:1 sampled:1 multivalued:2 conversation:1 dimensionality:3 sophisticated:1 actually:2 appears:1 steve:1 higher:5 supervised:2 follow:1 modal:1 amer:2 mar:1 marginalising:1 hand:1 nonlinear:4 lack:3 continuity:13 mode:35 quality:1 perhaps:1 excessively:1 contain:1 consisted:2 adequately:1 evolution:1 regularization:1 shuffled:1 deal:2 adjacent:1 sin:1 uniquely:2 plp:2 speaker:1 hippocampal:1 complete:3 tn:4 charles:1 functional:1 physical:1 exponentially:1 volume:1 fare:1 interpretation:1 measurement:2 shuffling:1 rd:1 automatic:1 grid:1 bartholomew:1 language:2 robot:2 etc:2 feb:1 curvature:1 multivariate:3 posterior:1 closest:1 instrumentation:1 binary:1 approximators:1 der:2 minimum:2 signal:2 dashed:1 branch:3 multiple:5 unimodal:1 smooth:2 technical:1 faster:1 cross:1 long:1 retrieval:1 prediction:4 regression:3 sheffield:2 multilayer:3 optimisation:1 iteration:1 sometimes:1 represent:1 achieved:2 cell:1 shef:1 interval:1 extra:2 rest:3 contrarily:1 ascent:1 comment:1 hz:2 vvv:1 jordan:2 bernstein:1 easy:2 jerk:1 gave:1 perfectly:1 idea:1 minimise:1 bottleneck:1 ltd:1 speech:13 york:1 cause:1 remark:1 useful:1 detailed:1 amount:3 locally:2 generate:2 http:1 percentage:1 dotted:1 estimated:2 per:3 track:1 discrete:3 group:1 datal:1 key:2 drawn:1 imputation:7 undersampling:1 graph:1 schroeter:2 missingness:1 inverse:4 angle:1 uncertainty:1 place:2 almost:1 reasonable:1 draw:1 griffin:1 data5:1 bit:1 bound:1 datum:1 quadratic:1 constraint:14 incorporation:1 athe:1 your:2 nearby:1 aspect:3 speed:2 span:1 optimality:1 performing:1 statically:1 according:1 combination:2 poor:1 jr:1 reconstructing:3 em:3 character:1 son:1 invariant:1 discus:1 fail:1 committee:1 available:1 gaussians:2 apply:3 appropriate:1 indirectly:1 robustness:1 hat:1 existence:1 original:2 microbeam:1 clustering:1 include:4 assembly:2 ghahramani:2 approximating:1 vizinho:1 contact:1 warping:1 objective:1 pergamon:1 added:2 parametric:1 strategy:5 dependence:1 diagonal:1 gradient:1 dp:3 distance:2 unable:2 mapped:1 thank:1 manifold:1 maximising:1 assuming:1 length:5 pointwise:5 index:1 relationship:1 difficult:2 regulation:1 potentially:3 implementation:1 unknown:1 perform:1 markov:3 iti:1 finite:2 coincident:1 situation:1 dc:2 frame:2 required:1 specified:1 acoustic:11 learned:1 nip:4 trans:1 able:1 bar:2 dynamical:1 pattern:11 usually:1 green:1 suitable:2 arm:2 improve:2 temporally:1 carried:1 utterance:5 tresp:1 prior:4 geometric:5 literature:2 review:1 regularisation:1 wisconsin:1 loss:2 approximator:2 rubin:1 principle:1 cooke:1 row:5 totalling:1 repeat:1 last:1 keeping:1 allow:1 perceptron:1 wide:1 saul:1 fg:1 van:2 distributed:1 curve:3 dimension:2 world:1 preprocessing:1 cope:1 reconstructed:6 approximate:2 dealing:1 global:2 assumed:1 continuous:3 latent:20 search:12 table:5 nature:2 learn:1 robust:3 obtaining:1 ioo:1 cupboard:1 apr:2 whole:2 noise:6 fig:1 representative:2 tl:10 slow:1 wiley:1 position:2 xl:2 lie:2 candidate:1 perpinan:1 perceptual:2 down:1 discarding:1 bishop:2 intrinsic:1 polygonal:1 sequential:4 simply:2 prevents:1 corresponds:1 conditional:19 acceleration:1 carreira:5 determined:1 generalisation:1 asymmetrically:1 partly:1 perceptrons:2 indicating:1 select:4 latter:1 incorporate:1 dept:2 audio:1 trainable:1
723
1,661
Inference for the Generalization Error Claude Nadeau CIRANO 2020, University, Montreal, Qc, Canada, H3A 2A5 jcnadeau@altavista.net Yoshua Bengio CIRANO and Dept. IRO Universite de Montreal Montreal, Qc, Canada, H3C 3J7 bengioy@iro.umontreal.ca Abstract In order to to compare learning algorithms, experimental results reported in the machine learning litterature often use statistical tests of significance. Unfortunately, most of these tests do not take into account the variability due to the choice of training set. We perform a theoretical investigation of the variance of the cross-validation estimate of the generalization error that takes into account the variability due to the choice of training sets. This allows us to propose two new ways to estimate this variance. We show, via simulations, that these new statistics perform well relative to the statistics considered by Dietterich (Dietterich, 1998). 1 Introduction When applying a learning algorithm (or comparing several algorithms), one is typically interested in estimating its generalization error. Its point estimation is rather trivial through cross-validation. Providing a variance estimate of that estimation, so that hypothesis testing and/or confidence intervals are possible, is more difficult, especially, as pointed out in (Hinton et aI., 1995), if one wants to take into account the variability due to the choice of the training sets (Breiman, 1996). A notable effort in that direction is Dietterich's work (Dietterich, 1998). Careful investigation of the variance to be estimated allows us to provide new variance estimates, which tum out to perform well. Let us first layout the framework in which we shall work. We assume that data are available in the form Zjl = {Z1, ... , Zn}. For example, in the case of supervised learning, Zi = (Xi,}Ii) E Z ~ RP+q, where p and q denote the dimensions of the X/s (inputs) and the }Ii 's (outputs). We also assume that the Zi'S are independent with Zi rv P(Z) . Let ?(D; Z), where D represents a subset of size nl ::; n taken from Zjl, be a function Znl X Z -t R For instance, this function could be the loss incurred by the decision that a learning algorithm trained on D makes on a new example Z. We are interested in estimating nJ.l. == E[?(Zjl; Zn+1)] where Zn+1 rv P(Z) is independent of Zjl. Subscript n stands for the size of the training set (Zjl here). The above expectation is taken over Zjl and Zn+1, meaning that we are interested in the performance of an algorithm rather than the performance of the specific decision function it yields on the data at hand. According to Dietterich's taxonomy (Dietterich, 1998), we deal with problems of type 5 through 8, (evaluating learning algorithms) rather then type 1 through 4 (evaluating decision functions). We call nJ.l. the generalization error even though it can also represent an error difference: ? Generalization error We may take ?(D; Z) = ?(D; (X, Y)) = Q(F(D)(X), Y), (1) C. Nadeau and Y. Bengio 308 where F(D) (F(D) : ]RP ~ ]Rq) is the decision function obtained when training an algorithm on D, and Q is a loss function measuring the inaccuracy of a decision. For instance, we could have Q(f), y) = I[f) 1= y], where I[ ] is the indicator function, for classification problems and Q(f), y) =11 f) - y 11 2 , where is II . II is the Euclidean norm, for "regression" problems. In that case nJ.L is what most people call the generalization error. ? Comparison of generalization errors Sometimes, we are not interested in the performance of algorithms per se, but instead in how two algorithms compare with each other. In that case we may want to consider .cCDi Z) = .c(Di (X, Y)) = Q(FA(D)CX), Y) - Q(FB(D)(X), Y), (2) where FA(D) and FB(D) are decision functions obtained when training two algorithms (A and B) on D , and Q is a loss function. In this case nJ.L would be a difference of generalization errors as outlined in the previous example. The generalization error is often estimated via some form of cross-validation. Since there are various versions of the latter, we layout the specific form we use in this paper. ? Let Sj be a random set of nl distinct integers from {I, ... , n }(nl < n). Here nl represents the size of the training set and we shall let n2 = n - nl be the size of the test set. ? Let SI, ... SJ be independent such random sets, and let Sj = {I, ... , n} \ Sj denote the complement of Sj. ? Let Z Sj = {Zi Ii E Sj} be the training set obtained by subsampling Zr according to the random index set Sj. The corresponding test set is ZSj = {Zili E Sj}. ? Let L(j, i) = .c(Zs;; Zi). According to (1), this could be the error an algorithm trained on the training set ZSj makes on example Zi. According to (2), this could be the difference of such errors for two different algorithms. ,i'k k ? Let (1,j = 2:~=1 L(j, i{) where i{, ... are randomly and independently drawn from Sj. Here we draw K examples from the test set ZS'j with replacement and compute the average error committed. The notation does not convey the fact that {1,j depends on K, nl and n2 . ? Let {1,j = limK ..... oo (1,j = ';2 2:iES~ L(j, i) denote what {1,j becomes as K increases J without bounds. Indeed, when sampling infinitely often from ZS'j' each Zi (i E Sj) is yielding the usual "average test error". The use of K is chosen with relative frequency .l.., n2 just a mathematical device to make the test examples sampled independently from Sj. Then the cross-validation estimate of the generalization error considered in this paper is ~K J I '""' ~ nl J.LJ - J L.J J.Lj. n2 _ j=1 We note that this an unbiased estimator of nlJ.L = E[.c{Zfl, Zn+r)] (not the same as nJ.L). This paper is about the estimation of the variance of ~~ {1,~. We first study theoretically this variance in section 2, leading to two new variance estimators developped in section 3. Section 4 shows part of a simulation study we performed to see how the proposed statistics behave compared to statistics already in use. 2 Analysis of Var[ ~~itr] Here we study Var[ ~~ {1,~]. This is important to understand why some inference procedures about nl J.L presently in use are inadequate, as we shall underline in section 4. This investigation also enables us to develop estimators of Var[ ~~ {1,~] in section 3. Before we proceed, we state the following useful lemma, proved in (Nadeau and Bengio, 1999). Inference for the Generalization Error 309 Lemma 1 Let U1, ... , Uk be random variables with common mean (3, common variance 6 and Cov[Ui , Uj] = "I, Vi '# j. Let1r = be the correlation between Ui and Uj (i '# j). J Let U = k- 1 2::=1 Ui and 8b = k~1 2::=1 (Ui U)2 be the sample mean and sample variance respectively. Then E[8b] = 6 - "I and Var[U] "I + (6~'Y) = 6 (11" + lk1l') . - = To study Var[ ~i j1,~] we need to define the following covariances. ? Let lio = liO(nl) = Var[L(j, i)] when i is randomly drawn from 8J. ? Let lil = lil (nl, n2) = Cov[L(j, i), L(j, i')] for i and i' randomly and independently drawn from 8j. ? Let li2 = liZ(nl, n2) = Cov[L(j, i), L(j', i')], with j independently drawn from 8j and 8jl respectively. '# j', i and i' randomly and ? Let li3 = li3(nl) = Cov[L(j, i), L(j, i')] for i, i' E 8j and i same as lil. In fact, it may be shown that ") L('), z")] -C O[L( V ) , z, lil lio + (nz nz - 1) li 3 -_ '# i'. This is not the . li3 + lio - nz li3 . (3) nz Let us look at the mean and variance of j1,j and ~i j1,~. Concerning expectations, we obviously have E[j1,j] = n1f.? and thus E[ ~ij1,~] = n1f.?. From Lemma 1, we have Var[j1,j] lil + O'?KO'I which implies Var[j1,j] = Var[ lim j1,j] = lim Var[j1,j] = lil. = K-too It can also be shown that Cov[j1,j, j1,j'] = TT var [n2 ~K] _ n1f.?J -liz+ TT [~] K-too liZ, var f.?j J j '# j', and therefore (using Lemma 1) liZ _ lil -liZ+ + 0'0-0'1 K J - liZ . (4) We shall often encounter liO, lil, liZ, li3 in the future, so some knowledge about those quantities is valuable. Here's what we can say about them. Proposition 1 For given nl and n2, we have 0 ~ liz ~ lil ~ lio and 0 ~ li3 ~ lil. Proof See (Nadeau and Bengio, 1999). A natural question about the estimator ~i j1,~ is how nl, nz, K and J affect its variance. Proposition 2 The variance of ~i j1,~ is non-increasing in J, K and nz. Proof See (Nadeau and Bengio, 1999). Clearly, increasing K leads to smaller variance because the noise introduced by sampling with replacement from the test set disappears when this is done over and over again. Also, averaging over many trainltest (increasing J) improves the estimation of nl f.?. Finally, all things equal elsewhere (nl fixed among other things), the larger the size of the test sets, the better the estimation of nl f.?. The behavior of Var[ ~i j1,~] with respect to nl is unclear, but we conjecture that in most situations it should decrease in nl. Our argument goes like this. The variability in ~i j1,~ comes from two sources: sampling decision rules (training process) and sampling testing examples. Holding n2, J and K fixed freezes the second source of variation as it solely depends on those three quantities, not nl. The problem to solve becomes: how does nl affect the first source of variation? It is not unreasonable to say that the decision function yielded by a learning algorithm is less variable when the training set is large. We conclude that the first source of variation, and thus the total variation (that is Var[ ~ij1,~]) is decreasing in nl. We advocate the use of the estimator (5) C. Nadeau and Y Bengio 310 as it is easier to compute and has smaller variance than ~~it} (J, nl, n2 held constant). Var[ n2 11 00 ] = lim Var[ nl,-J K-+oo where P -- ~ 111 n2 rl.K] nl,-J = (72 + (71 - (72 - J - (7 1 (p + -1 -Jp)- ' (6) oo r/OO] -- Corr[ll. '-j , '-j' . 3 Estimation of Var[ ~~JtJ] We are interested in estimating ~~(7J == Var[ ~~ it:f] where ~~ it:f is as defined in (5). We provide two different estimators of Var[ ~~ it:f]. The first is simple but may have a positive or negative bias for the actual variance. The second is meant to be conservative, that is, if our conjecture of the previous section is correct, its expected value exceeds the actual variance. 1st Method: Corrected Resampled t- Test. Let us recall that ~~ it:f = jj2 be the sample variance of the itj's. According to Lemma 1, E[jj 21=(71(1-p)= I-p P+ !=?(71 J ( I-P) p+~ = (p+!=?) ~ 1. + I-p Var[ ~~ iL:f]. J (J + G) jj2 is an unbiased estimator of so that (71 J'Ef=1 itj . Let Var[ ~~it:f] l+--L ' J I-p (7) The only problem is that p = p(nl,n2) = :~t~:::~~, the correlation between the itj's, is unknown and difficult to estimate. We use a naive surrogate for p as follows. Let us recall that iLj = 'EiES~, ?(ZSj; Zi). For the purpose of building our estimator, let us make the :2 approximation that ?(ZSj; Zi) depends only on Zi and nl. Then it is not hard to show (see (Nadeau and Bengio, 1999)) that the correlation between the itj's becomes nl~n2' There- Var[~~iL:fl is (J + l~~o) jj2 where Po = po(nl,n2) = nl~n2' that is (J + ~ ) jj2. This will tend to overestimate or underestimate Var[ ~~ iL:f] according to whether Po > p or Po < p. Note that this first method basically does not require fore our first estimator of any more computations than that already performed to estimate generalization error by cross-validation. 2nd Method: Conservative Z. Our second method aims at overestimating Var[ ~~ iL:f] which will lead to conservative inference, that is tests of hypothesis with actual size less than the nominal size. This is important because techniques currently in use have the opposite defect, that is they tend to be liberal (tests with actual size exceeding the nominal size), which is typically regarded as less desirable than conservative tests. Estimating ~~ (7J unbiasedly is not trivial as hinted above. However we may estimate uJ be the unbiased unbiasedly nn?1 (7J = Var[ nn?1 it:fl where n~ = L!!2 J - n2 < nl. Let n? n1 estimator, developed below, of the above variance. We argued in the previous section that Var[ ~~ it:fl ~ Var[ ~~ iL:fl. Therefore ~;uJ will tend to overestimate ~~(7J, that is E[ n2a-2] = n2(72 > n2(72 n; J n; J - nl J' Here's how we may estimate ~? (7J without bias. For simplicity, assume that n is even. 1 We have to randomly split our data into two distinct data sets, Dl and D1, of size ~ each. Let iL(1) be the statistic of interest ( ~; iL:f) computed on D 1 . This involves, among other things, drawing J train/test subsets from DI . Let iL(l) be the statistic computed on D 1? Then iL(l) and iL(l) are independent since Dl and Dl are independent data sets, so A it(I )+it(1))2 it(!)+it(I))2 I(A AC)2' b' ed . that (/-L(l) 2 + (AC J.L(I) 2 = 2" /-L(l) - /-L(l) IS an un las estImate of ~?(7J. This splitting process may be repeated M times. This yields Dm and D~, with Zr 1 Inference for the Generalization Error Dm U D~ 311 = zf, Dm n D~ = 0 for m = 1, ... , M. Each split yields a pair (it(m) , it(m?) that is such that ~(it(m) - it(m?)2 is unbiased for ~~U}. This allows us to use the following unbiased estimator of ~? U}: 1 ~2 _ n2 n~ U J - M 1 ""' (~ ~ c )2 2M L..J J-t(m) - J-t(m) . (8) m=1 Note that, according to Lemma 1, Var[ ~~oj] = t Var[(it(m) - it(m?)2] (r + IMr) with r = Corr[(it(i) - it(i?)2, (it(j) - it(j?)2] for i i- j. Simulations suggest that r is usually close to 0, so that the above variance decreases roughly like for M up to 20, say. The second method is therefore a bit more computation intensive, since requires to perform cross-validation M times, but it is expected to be conservative. k 4 Simulation study We consider five different test statistics for the hypothesis Ho : niJ-t = J-to. The first three are methods already in use in the machine learning community, the last two are the new methods we put forward. They all have the following form reject Ho if Iit ~J-to I > c. (9) Table 1 describes what they are 1. We performed a simulation study to investigate the size (probability of rejecting the null hypothesis when it is true) and the power (probability of rejecting the null hypothesis when it is false) of the five test statistics shown in Table 1. We consider the problem of estimating generalization errors in the Letter Recognition classification problem (available from www. ics. uci . edu/pub/machine-learning-databases). The learning algorithms are 1. Classification tree We used the function tree in Splus version 4.5 for Windows. The default arguments were used and no pruning was performed. The function predict with option type="class" was used to retrieve the decision function of the tree: FA (Zs)(X). Here the classification loss function LAU,i) = I[FA(Zsj)(X i ) i- Yi ] is equal to 1 whenever this algorithm misclassifies example i when the training set is Sj; otherwise it is O. 2. First nearest neighbor We apply the first nearest neighbor rule with a distorted distance metric to pun down the performance of this algorithm to the level of the classification tree (as in (Dietterich, 1998?. We have LBU, i) equal to 1 whenever this algorithm misclassifies example i when the training set is Sj; otherwise it is O. In addition to inference about the generalization errors ni J-tA and ni J-tB associated with those two algorithms, we also consider inference about niJ-tA-B = niJ-tA - niJ-tB = E[LA-B(j,i)] whereLA_B(j,i) = LAU,i) - LB(j,i). We sample, without replacement, 300 examples from the 20000 examples available in the Letter Recognition data base. Repeating this 500 times, we obtain 500 sets of data of the form {ZI,"" Z300}. Once a data set = {ZI,'" Z300} has been generated, we may zloO lWhen comparing two classifiers, (Nadeau and Bengio, 1999) show that the t-test is closely related to McNemar's test described in (Dietterich, 1998). The 5 x 2 cv procedure was developed in (Dietterich, 1998) with solely the comparison of classifiers in mind but may trivially be extended to other problems as shown in (Nadeau and Bengio, 1999). C. Nadeau and Y. Bengio 312 c II II Name n2 AOO nl/-Ll n2 AOO ~2 SV(L(I, i)) yO-:.l t n2 - 1,1-ar/2 t J - 1,1-ar/2 see (Dietterich, 1998) t S ,1-ar/2 1: conservative Z n/2/-Ll n2 AOO 2: corr. resampled t n2 AOO nl /-L J t-test (McNemar) resampled t nl/-LJ n(2 AOO Dietterich's 5 x 2 cv nl/-LJ n2 A2 n'UJ 1 Zl-ar/2 2 (!.J + ~) nl 0- tJ-l,1-ar/2 n2IT3+~ITO-IT3) ITn -IT3 I+J~ >1 >1 " ? ~~IT? < 1 n,u J l+JE l+J~ Table 1: Description of five test statistics in relation to the rejection criteria shown in (9). Zp and h,p refer to the quantile p of the N(O, 1) and Student tk distribution respectively. 0- 2 is as defined above (7) and SV (L(I, i)) is the sample variance of the L(I, i)'s involved in ~i {l{'. The ~t~~l ratio (which comes from proper application of Lemma 1, except for Dietterich's 5 x 2 cv and the Conservative Z) indicates if a test will tend to be conservative (ratio less than 1) or liberal (ratio greater than 1). perform hypothesis testing based on the statistics shown in Table 1. A difficulty arises 300 here), those methods don't aim at inference for the same however. For a given n (n generalization error. For instance, Dietterich's 5 x 2 cv test aims at n/2/-L, while the others aim at nl/-L where nl would usually be different for different methods (e.g. nl = 23n for the t test statistic, and nl = ~~ for the resampled t test statistic, for instance). In order to compare the different techniques, for a given n, we shall always aim at n/2/-L, i.e. use nl = ?-. However, for statistics involving ~ip.r with J > 1, normal usage would call for nl to be 5 or 10 times larger than n2, not nl = n2 = ?-. Therefore, for those statistics, we = p.r l~ so that ~ 5. To obtain ~~;o we simply throw out 40% also use nl = ?- and n2 of the data. For the conservative Z, we do the variance calculation as we would normally = = for instance) to obtain ~i2-n2a-J = ;~~~a-J. However, in the numerator we n/2AOO d n2 AOO n/lOAoo' d f n2 AOO I' db compute b oth n/2/-LJ an n/2/-LJ n/2 /-LJ mstea 0 n-n2/-LJ' as exp rune a ove. do (n2 = l~ Note that the rationale that led to the conservative Z statistics is maintained, that is ;~~~a-J 2A OO] . bot h TT E [n/lOA2] overestimates var [n/lOAOO] n/2 /-LJ an d TT var [n/ n/2/-LJ: 2n/s u J > _ TT var [n/lOAOO] n/2 /-LJ > [n/2 00] var n/2/-LJ . TT A Figure 1 shows the estimated power of different statistics when we are interested in /-LA and /-LA-B. We estimate powers by computing the proportion of rejections of Ho . We see that tests based on the t-test or resampled t-test are liberal, they reject the null hypothesis with probability greater than the prescribed a = 0.1, when the null hypothesis is true. The other tests appear to have sizes that are either not significantly larger the 10% or barely so. Note that Dietterich's 5 x 2cv is not very powerful (note that its curve has the lowest power on the extreme values of muo). To make a fair comparison of power between two curves, one should mentally align the size (bottom of the curve) of these two curves. Indeed, even the resampled t-test and the conservative Z that throw out 40% of the data are more powerful. That is of course due to the fact that the 5 x 2 cv method uses J = 1 instead of J = 15. This is just a glimpse of a much larger simulation study. When studying the corrected resampled t-test and the conservative Z in their natural habitat (nl = ~9 and n2 = l~)' we see that they are usually either right on the money in term of size, or slIghtly conservative. Their powers appear equivalent. The simulations were performed with J up to 25 and M up to 20. We found that taking J greater than 15 did not improve much the power of the Inference for the Generalization Error 313 Figure 1: Powers of the tests about Ho : /-LA = /-Lo (left panel) and Ho : /-LA-B = /-Lo (right panel) at level a = 0.1 for varying /-Lo. The dotted vertical lines correspond to the 95% confidence interval for the actual/-LA or /-LA-B. therefore that is where the actual size of the tests may be read. The solid horizontal line displays the nominal size of the tests. i.e. 10%. Estimated probabilities of rejection laying above the dotted horizontal line are significatively greater than 10% (at significance level 5%). Solid curves either correspond to the resampled t-test or the corrected resampled t-test. The resampled t-test is the one that has ridiculously high size. Curves with circled points are the versions of the ordinary and corrected resampled t-test and conservative Z with 40% of the data thrown away. Where it matters J = 15. M = 10 were used. statistics. Taking M = 20 instead of M = 10 does not lead to any noticeable difference in the distribution of the conservative Z. Taking M = 5 makes the statistic slightly less conservative. See (Nadeau and Bengio. 1999) for further details. 5 Conclusion This paper addresses a very important practical issue in the empirical validation of new machine learning algorithms: how to decide whether one algorithm is significantly better than another one. We argue that it is important to take into account the variability due to the choice of training set. (Dietterich. 1998) had already proposed a statistic for this purpose. We have constructed two new variance estimates of the cross-validation estimator of the generalization error. These enable one to construct tests of hypothesis and confidence intervals that are seldom liberal. Furthermore. tests based on these have powers that are unmatched by any known techniques with comparable size. One of them (corrected resampled t-test) can be computed without any additional cost to the usual K-fold crossvalidation estimates. The other one (conservative Z) requires M times more computation. where we found sufficiently good values of M to be between 5 and 10. References Breiman, L. (1996). Heuristics of instability and stabilization in model selection. Annals of Statistics, 24 (6):2350-2383. Dietterich, T. (1998). Approximate statistical tests for comparing supervised classification learning algorithms. Neural Computation, 10 (7):1895-1924. Hinton. G., Neal. R., Tibshirani, R., and DELVE team members (1995). Assessing learning procedures using DELVE. Technical report, University of Toronto, Department of Computer Science. Nadeau, C. and Bengio, Y. (1999). Inference for the generalisation error. Technical Report in preparation, CIRANO.
1661 |@word version:3 norm:1 proportion:1 underline:1 nd:1 simulation:7 covariance:1 solid:2 pub:1 comparing:3 si:1 ij1:2 j1:14 aoo:8 enables:1 device:1 toronto:1 liberal:4 pun:1 unbiasedly:2 five:3 mathematical:1 constructed:1 advocate:1 theoretically:1 li3:6 expected:2 indeed:2 roughly:1 behavior:1 decreasing:1 actual:6 window:1 increasing:3 becomes:3 estimating:5 notation:1 panel:2 null:4 what:4 lowest:1 z:4 developed:2 nj:5 classifier:2 uk:1 zl:1 normally:1 appear:2 overestimate:3 before:1 positive:1 subscript:1 solely:2 nz:6 delve:2 practical:1 testing:3 developped:1 procedure:3 empirical:1 reject:2 significantly:2 confidence:3 suggest:1 close:1 selection:1 put:1 applying:1 instability:1 www:1 equivalent:1 layout:2 go:1 independently:4 qc:2 simplicity:1 splitting:1 rune:1 estimator:12 rule:2 regarded:1 retrieve:1 variation:4 annals:1 nominal:3 us:1 hypothesis:9 recognition:2 database:1 bottom:1 decrease:2 valuable:1 rq:1 ui:4 trained:2 ove:1 po:4 iit:1 various:1 train:1 distinct:2 heuristic:1 larger:4 solve:1 say:3 drawing:1 otherwise:2 cov:5 statistic:20 n2a:2 h3c:1 ip:1 obviously:1 claude:1 net:1 propose:1 uci:1 description:1 crossvalidation:1 zp:1 assessing:1 tk:1 oo:5 develop:1 montreal:3 ac:2 nearest:2 noticeable:1 throw:2 involves:1 implies:1 come:2 direction:1 closely:1 correct:1 stabilization:1 enable:1 require:1 argued:1 generalization:18 investigation:3 proposition:2 hinted:1 sufficiently:1 considered:2 liz:8 ic:1 normal:1 exp:1 predict:1 a2:1 purpose:2 estimation:6 currently:1 clearly:1 j7:1 always:1 aim:5 rather:3 breiman:2 varying:1 yo:1 indicates:1 inference:10 nn:2 typically:2 lj:12 relation:1 interested:6 issue:1 classification:6 among:2 misclassifies:2 equal:3 once:1 construct:1 sampling:4 represents:2 look:1 future:1 yoshua:1 others:1 overestimating:1 report:2 randomly:5 replacement:3 n1:1 thrown:1 interest:1 a5:1 investigate:1 extreme:1 nl:46 yielding:1 tj:1 oth:1 held:1 glimpse:1 tree:4 euclidean:1 theoretical:1 nij:4 instance:5 ar:5 measuring:1 zn:5 ordinary:1 cost:1 subset:2 inadequate:1 too:2 reported:1 sv:2 st:1 itj:4 ilj:1 again:1 unmatched:1 leading:1 li:1 account:4 de:1 student:1 matter:1 notable:1 depends:3 vi:1 performed:5 jtj:1 option:1 il:10 ni:2 variance:23 yield:3 correspond:2 rejecting:2 basically:1 fore:1 whenever:2 ed:1 underestimate:1 frequency:1 involved:1 jj2:4 dm:3 universite:1 proof:2 di:2 associated:1 sampled:1 proved:1 recall:2 lim:3 knowledge:1 improves:1 tum:1 ta:3 supervised:2 done:1 though:1 furthermore:1 just:2 correlation:3 hand:1 horizontal:2 usage:1 dietterich:16 name:1 building:1 true:2 unbiased:5 read:1 i2:1 neal:1 deal:1 ll:3 numerator:1 maintained:1 criterion:1 tt:6 meaning:1 ef:1 umontreal:1 common:2 mentally:1 rl:1 jp:1 jl:1 refer:1 freeze:1 ai:1 cv:6 seldom:1 outlined:1 trivially:1 pointed:1 had:1 money:1 base:1 align:1 mcnemar:2 yi:1 greater:4 additional:1 itn:1 ii:7 rv:2 desirable:1 exceeds:1 technical:2 h3a:1 calculation:1 cross:7 concerning:1 y:1 involving:1 regression:1 ko:1 expectation:2 metric:1 sometimes:1 represent:1 addition:1 want:2 interval:3 source:4 limk:1 tend:4 db:1 thing:3 member:1 call:3 integer:1 bengio:12 split:2 affect:2 zi:12 zili:1 opposite:1 itr:1 intensive:1 whether:2 effort:1 proceed:1 jj:1 useful:1 se:1 repeating:1 it3:2 dotted:2 bot:1 estimated:4 per:1 tibshirani:1 li2:1 shall:5 splus:1 drawn:4 defect:1 letter:2 powerful:2 distorted:1 decide:1 draw:1 decision:9 comparable:1 bit:1 bound:1 resampled:12 fl:4 display:1 fold:1 yielded:1 u1:1 argument:2 prescribed:1 conjecture:2 department:1 according:7 smaller:2 describes:1 slightly:2 presently:1 lau:2 taken:2 mind:1 studying:1 available:3 unreasonable:1 apply:1 away:1 encounter:1 ho:5 rp:2 subsampling:1 quantile:1 especially:1 uj:5 already:4 quantity:2 question:1 fa:4 usual:2 surrogate:1 unclear:1 distance:1 lio:6 argue:1 trivial:2 iro:2 barely:1 laying:1 index:1 providing:1 ratio:3 difficult:2 unfortunately:1 taxonomy:1 holding:1 negative:1 lil:10 proper:1 litterature:1 perform:5 unknown:1 zf:1 vertical:1 imr:1 behave:1 situation:1 hinton:2 variability:5 committed:1 extended:1 team:1 lb:1 community:1 canada:2 zjl:6 introduced:1 complement:1 pair:1 z1:1 nlj:1 inaccuracy:1 address:1 below:1 usually:3 tb:2 oj:1 power:9 natural:2 difficulty:1 indicator:1 zr:2 improve:1 habitat:1 disappears:1 nadeau:12 naive:1 circled:1 relative:2 loss:4 rationale:1 var:33 validation:8 incurred:1 lo:3 elsewhere:1 course:1 last:1 bias:2 understand:1 neighbor:2 taking:3 curve:6 dimension:1 default:1 stand:1 evaluating:2 fb:2 forward:1 cirano:3 sj:14 pruning:1 approximate:1 conclude:1 xi:1 don:1 un:1 why:1 table:4 ca:1 did:1 significance:2 noise:1 n2:34 repeated:1 fair:1 convey:1 je:1 exceeding:1 bengioy:1 ito:1 down:1 specific:2 dl:3 false:1 corr:3 easier:1 rejection:3 cx:1 led:1 simply:1 infinitely:1 zsj:5 careful:1 hard:1 generalisation:1 except:1 corrected:5 averaging:1 lemma:7 conservative:17 total:1 experimental:1 la:8 people:1 latter:1 arises:1 meant:1 preparation:1 dept:1 d1:1
724
1,662
Application of Blind Separation of Sources to Optical Recording of Brain Activity Holger Schoner, Martin Stetter, Ingo Schie61 Department of Computer Science Technical University of Berlin Germany {hjsch,moatl,ingos}@cs.tu-berlin.de John E. W. Mayhew University of Sheffield, UK j. e.mayhew@sheffield.ac.uk Jennifer S. Lund, Niall McLoughlin Klaus Obermayer Institute of Ophthalmology University College London, UK {j.lund,n.mcloughlin}@ucl.ac.uk Department of Computer Science, Technical University of Berlin, Germany oby@cs.tu-berlin.de Abstract In the analysis of data recorded by optical imaging from intrinsic signals (measurement of changes of light reflectance from cortical tissue) the removal of noise and artifacts such as blood vessel patterns is a serious problem. Often bandpass filtering is used, but the underlying assumption that a spatial frequency exists, which separates the mapping component from other components (especially the global signal), is questionable. Here we propose alternative ways of processing optical imaging data, using blind source separation techniques based on the spatial decorre1ation of the data. We first perform benchmarks on artificial data in order to select the way of processing, which is most robust with respect to sensor noise. We then apply it to recordings of optical imaging experiments from macaque primary visual cortex. We show that our BSS technique is able to extract ocular dominance and orientation preference maps from single condition stacks, for data, where standard post-processing procedures fail. Artifacts, especially blood vessel patterns, can often be completely removed from the maps. In summary, our method for blind source separation using extended spatial decorrelation is a superior technique for the analysis of optical recording data. 1 Introduction One approach in the attempt of comprehending how the human brain works is the analysis of neural activation patterns in the brain for different stimuli presented to a sensory system. An example is the extraction of ocular dominance or orientation preference maps from recordings of activity of neurons in the primary visual cortex of mammals. A common technique for extracting such maps is optical imaging (01) of intrinsic signals. Currently this is the imaging technique with the highest spatial resolution (~ 100 J1m) for mapping of the cortex. This method is explained e.g. in [1], for similar methods using voltage sensitive dyes see [2, 3] . 01 uses changes in light reflection to estimate spatial patterns of stimulus 950 H. Sch6ner. M Stetter. I. Schiej3l, J E. Mayhew, J Lund, N. Mcloughlin and K. Obermayer answers. The overall change recorded by a CCD or video camera is the total signal. The part of the total signal due to local neural activity is called the mapping component and it derives from changes in deoxyhemoglobin absorption and light scattering properties of the tissue. Another component of the total signal is a "global" component, which is also correlated with stimulus presentation, but has a much coarser spatial re~olution . It derives its part from changes in the blood volume with the time. Other components are blood vessel artifacts, the vasomotor signal (slow oscillations of neural activity), and ongoing activity (spontaneous, stimulus-uncorrelated activity). Problematic for the extraction of activity maps are especially blood vessel artifacts and sensor noise, such as photon shot noise. A procedure often used for extracting the activity maps from the recordings is bandpass filtering, after preprocessing by temporal , spatial , and trial averaging. Lowpass filtering is unproblematic, as the spatial resolution of the mapping signal is limited by the scattering properties of the brain tissue, hence everything above a limiting frequency must be noise. The motivation for highpass filtering, on the other hand, is questionable as there is no specific spatial frequency separating local neural activity patterns and the global signal [4]). A different approach, Blind Source Separation (BSS), models the components of the recorded image frames as independent sources, and the observations (recorded image frames) as noisy linear mixtures of the unknown sources. After performing the BSS the mapping component should ideally be concentrated in one estimated source, the global signal in another, and blood vessel artifacts, etc. in further ones. Previous work ([5]) has shown that BSS algorithms, which are based on higher order statistics ([6, 7, 8]), fail for optical imaging data, because of the high signal to noise ratio. In this work we suggest and investigate versions of the M&S algorithm [9, 10], which are robust against sensor noise, and we analyze their performance on artificial as well as real optical recording data. In section 2 we describe an improved algorithm, which we later compare to other methods in section 3. There an artificial data set is used for the analysis of noise robustness, and benchmark results are presented. Then, in section 4, it is shown that the newly developed algorithm is very well able to separate the different components of the optical imaging data, for ocular dominance as well as orientation preference data from monkey striate cortex. Finally, section 5 provides conclusions and perspectives for future work. 2 Second order blind source separation Let m be the number of mixtures and r the sample index, i.e. a vector specifying a pixel in the recorded images. The observation vectors y(r) = (Y1(r) , ... ,Ym?')f are assumed to be linear mixtures of m unknown sources s(r) = (Sl (r) , . . . ,Sm (r)) with A being the m x m mixing matrix and n describing the sensor noise: y(r) = As(r) + n (1) The goal of BSS is to obtain optimal source estimates s(r) under the assumption that the original sources are independent. In the noiseless case W = A -1 would be the optimal demixing matrix. In the noisy case, however, W also has to compensate for the added W . A . s(r) + W . n. BSS algorithms are generally only able to noise: s(r) == Wy(r) recover the original sources up to a permutation and scaling. = Extended Spatial Decorrelation (ESD) uses the second order statistics of the observations to find the source estimates. If sources are statistical independent all source crosscorrelations Ci(,~) (D.r) = (si (r)Sj(r+ D.r))r = ~ LSi (r)Sj(r+ D.r) r , i =F j (2) Application of BSS to Optical Recording of Brain Activity 951 must vanish for all shifts ~r, while the autocorrelations (i = j) of the sources remain (the variances). Note that this implies that the sources must be spatially smooth. Motivated by [to] we propose to optimize the cost function, which is the sum of the squared cross-correlations of the estimated sources over a set of shifts {~r}, E(W) = LL ((WC(~r)WT)i,jr (3) 6r i~j = L L \Si(r)Sj(r + ~r))~ , 6r i t j with respect to the demixing matrix W. The matrix Ci,j(~r) = (Yi (r)Yj(r + ~r))r denotes the mixture cross-correlations for a shift ~r. This cost function is minimized using the Polak Ribiere Conjugate Gradient technique, where the line search is substituted by a dynamic step width adaptation ([11]). To keep the demixing matrix W from converging to the zero matrix, we introduce a constraint which keeps the diagonal elements of T = W-l (in the noiseless case and for non-sphered data T is an estimate of the mixing matrix, with possible permutations) at a value of 1.0. Convergence properties are improved by sphering the data (transforming their correlation matrix for shift zero to an identity matrix) prior to decorrelating the mixtures. Note that use of multiple shifts ~r allows to use more information about the auto- and cross-correlation structure of the mixtures for the separation process. Two shifts provide just enough constraints for a unique solution ([to]). Multiple shifts, and the redundancy they introduce, additionally allow to cancel out part of the noise by approximate simultaneous diagonalization of the corresponding cross correlation matrices. In the presence of sensor noise, added after mixing, the standard sphering technique is problematic. When calculating the zero-shift cross-correlation matrix the variance of the noise contaminates the result, and sphering using a shifted cross-correlation matrix, is recommended ([12]). For spatially white sensor noise and sources with reasonable auto correlations this technique is more appropriate. In the following we denote the standard algorithm by dpaO, and the variant using noise robust sphering by dpa1. 3 Benchmarks for artificial data The artificial data set used here, whose sources are approximately uncorrelated for all shifts, is shown in the left part of figure 1. The mixtures were produced by generating a random mixing matrix (in this case with condition number 3.73), applying it to the sources, and finally adding white noise of different variances. In order to measure the performance on the artificial data set we measure a reconstruction error (RE) between the estimated and the correct sources via (see [l3]): RE(W) = od(L ?(r)sT(r)) , r od(C) 1 ~ 1 =N ~ N _ I 1 ( L J IC? ?1 maXk'I~i ,kl - ) 1 (4) The correlation between the real and the estimated sources (the argument to "od"), should be close to a permutation matrix, if the separation is successful. If the maxima of two rows are in the same column, the separation is labeled unsuccessful. Otherwise, the normalized absolute sum of non-permutation (cross-correlation) elements is computed and returned as the reconstruction error. We now compare the method based on optimization of (3) by gradient descent with the following variants of second order blind source separation: (1) standard spatial decorrelation 952 H. Schaner; M Stetter; I. Schiej3l, J. E. Mayhew, J. Lund, N. McLoughlin and K. Obermayer -~-.-- opt 0.5 g ~0.4 ? gOJ 0.5 mean . ~. <; JiO.4 cor -. '. l:l ~ hI''I, c: 0 '' ?. 80.2 " ': " H, ~02 0:: '. 10 15 20 Signal to Noise Ratio (dB) " ~ 0.1 25 00 - jacO jacl dpaO dpal % ~ , ~ :. , '!i ___ 5 , 1 b 0.1 o0 .__ . ... gO.3 -" 0:: -..., 5 -- 10 ----15 20 25 Signal to Noi se Ratio (dB) Figure 1: The set of three approximately uncorrelated source images of the artificial data set (left) . The two plots (middle, right) show the reconstruction error versus signal to noise ratio for different separation algorithms. In the right plot jac1 and dpa1 are very close together. using the optimal single shift yielding the smallest reconstruction error (opt). (2) Spatial decorrelation using the shift selected by .6.rcor = argmax{.D.r} norm (C(.6.r) - diag (C(.6.r))) norm (diag (C(.6.r)))?' (5) where "diag" sets all off-diagonal elements of its argument matrix to zero, and "norm" computes the largest singular value of its argument matrix (cor). .6.rcor is the shift for which the cross correlations are largest, i.e. whose signal to noise ratio (SNR) should be best. (3) Standard spatial decorrelation using the average reconstruction error for all successful shifts in a 61 x 61 square around the zero shift (mean). (4) A multi-shift algorithm ([12]), using several elementary rotations (Jacobi method) to build an orthogonal demixing matrix, which optimizes the cost function (3). The variants using standard sphering and noise robust sphering are denoted by dacO) and dac1). cor, opt, and mean use two shifts for their computation; but as one of those is always the zero-shift, there is only one shift to choose and they are called single-shift algorithms here. Figure 1 gives two plots which show the reconstruction error (4) versus the SNR (measured in dB) for single shift (middle) and multi-shift (right) algorithms. The error bars indicate twice the standard error of the mean (2x SEM), for 10 runs with the same mixing matrix, but newly generated noise of the given noise level. In each of these runs, the best result of three was selected for the gradient descent method. This is because, contrary to the other algorithms, the gradient descent algorithm depends on the initial choice of the demixing matrix. All multi-shift algorithms (all except opt and mean), used 8 shifts (?r, ?r), (?r, 0), and (0, ?r) for each r E {I, 3, 5, 10,20, 30}, so 48 all together. Several points are noticeable in the plots. (i) The cor algorithm is generally closer to the optimum than to the average successful shift. (ii) A comparison between the two plots shows that the multi-shift algorithms (right plot) are able to perform much better than even the optimal single-shift method. For low to medium noise levels this is even the case when using the standard sphering method combined with the gradient descent algorithm. (iii) The advantage of the noise robust sphering method, compared to the standard sphering, is obvious: the reconstruction error stays very low for all evaluated noise levels, for both the jac1 and dpa1 algoritlnns. (iv) The gradient descent technique is more robust than the Jacobi method For the standard sphering its performance is much better than that of the Jacobi method. Figure 1 shows results which were produced using a single mixing matrix. However, our simulations show that the algorithms compare qualitatively similar when using mixing ma- 953 Application of BSS to Optical Recording ofBrain Activity t = 1 sec. t = 2 sec. t = 3 sec. t = 4 sec. t = 5 sec. t = 6 sec. t = 7 sec. Figure 2: Optical imaging stacks. The top stack is a single condition stack from ocular dominance experiments, the lower one a difference stack from orientation preference experiments (images for 90? gratings subtracted from those for 0? gratings). The stimulus was present during recording images 2-7 in each row. Two large blood vessels in the top and left regions of the raw images were masked out prior to the analysis. trices with condition numbers between 2 and 10. The noise robust versions of the multishift algorithms generally yield the best separation results of all evaluated algorithms. 4 Application to optical imaging We now apply extended spatial decorrelation to the analysis of optical imaging data. The data consists of recordings from the primary visual cortex of macaque monkeys. Each trial lasted 8 seconds, which were recorded with frame rates of 15 frames per second. A visual stimulus (a drifting bar grating of varying orientation) was presented between seconds 2 and 8. Trials were separated by a recovery period of 15 seconds without stimulation. The cortex was illuminated at a wavelength of 633 nm. One pixel corresponds to about 15 J.Lm on the cortex; the image stacks used for further processing, consisting of 256 x 256 pixels, covered an area of cortex of approximately 3.7 mm 2 . Blocks of 15 consecutive frames were averaged, and averaging over 8 trials using the same visual stimulus further improved the SNR. First frame analysis (subtraction of the first, blank, frame from the others) was then applied to the resulting stack of 8 frames, followed by lowpass filtering with 14 cycles/mm. Figure 2 shows the resulting image stacks for an ocular dominance and an orientation preference experiment. One observes strong blood vessel artifacts (particularly in the top row of images), which are superimposed to the patchy mapping component that pops up over time. Figure 3 shows results obtained by the application of extended spatial decorrelation (using dpaO). Only those estimated sources containing patterns different from white noise are shown. Backprojection of the estimated sources onto the original image stack yields the amplitude time series of the estimated sources, which is very useful in selecting the mapping component: it can be present in the recordings only after the stimulus onset (starting at t = 2 sec.). The middle part shows four estimated sources for the ocular dominance single condition stack. The mapping component (first image) is separated from the global component (second image) and blood vessel artifacts (second to fourth) quite well. The time course of the mapping component is plausible as well: calculation of a plausibility index (sum of squared differences between the normalized time series and a step function, which is 0 before and 1 after the stimulus onset) gives 0.5 for the mapping component and 2.31 for the next best one. Results for the gradient descent algorithm are similar for this data set, regardless of the sphering technique used. The Jacobi method also gives similar results, but a small blood vessel artifact is remaining in the resulting map. The cor algorithm usually gives much worse separation results. In the right part of figure 3 two es- 954 H Schaner, M Stetter, I SchieJ3l, 1. E. Mayhew, 1. Lund, N McLoughlin and K. Obermayer Figure 3: Left: Summation technique for ocular dominance (aD) experiment (upper) and orientation preference (OP) experiment (lower). Middle, Right: dpaO algorithm applied to the same aD single condition (middle) and OP (right) stacks. The images show the 4 (aD) and 2 (OP) estimated components, which are visually different from white noise. In the bottom row the respective time courses of the estimated sources are given. timated sources (those different from white noise) for the orientation preference difference stack can be seen. Here the proposed algorithm (dpaO) again works very well (plausibility index is 0.56 for mapping component, compared to 3.04 for the best other component). It generally has to be applied a few times (usually around 3 times) to select the best separation result Uudging by visual quality of the separation and the time courses of the estimated sources), because of its dependence on parameter initialization; in return it yields the best results of all algorithms used, especially when compared to the traditional summation technique. The similar results when using standard and noise robust sphering, and the small differences between the gradient descent and the Jacobi algorithms indicate, that not sensor noise is the limiting factor for the quality of the extracted maps. Instead it seems that, assuming a linear mixing model, no better results can be obtained from the used image stacks. It will remain for further research to analyze, how appropriate the linear mixing model is, and whether the underlying biophysical components are sufficiently uncorrelated . In the meantime the maps obtained by the ESD algorithm are superior to those obtained using conventional techniques like summation of the image stack. 5 Conclusion The results presented in the previous sections show the advantages of the proposed algorithm: In the comparison with other spatial decorrelation algorithms the benefit in using multiple shifts compared to only two shifts is demonstrated. The robustness against sensor noise is improved, and in addition, the selection of multiple shifts is less critical than selecting a single shift, as the resulting multi-shift system of equations contains more redundancy. In comparison with the Jacobi method, which is restricted to find only orthogonal demixing matrices, the greater tolerance of demixing by a gradient descent technique concerning noise and incorrect sphering are demonstrated. The application of second order blind separation of sources to optical imaging data shows that these techniques represent an important alternative to the conventional approach, bandpass filtering followed by summation of the image stack, for extraction of neural activity maps. Vessel artifacts can be separated from the mapping component better than using classical approaches. The spatial decorrelation algorithms are very well adapted to the optical imaging task, because of their use of spatial smoothness properties of the mapping and other biophysical components. An important field for future research concerning BSS algorithms is the incorporation of prior knowledge about sources and the mixing process, e.g. that the mixing has to be causal: the mapping signal cannot occur before the stimulus is presented. Assumptions Application ofBSS to Optical Recording of Brain Activity 955 about the time course of signals could also be helpful, as well as knowledge about their spatial statistics. Smearing and scattering limit the resolution of recordings of biological components, and, depending on the wavelength of the light used for illumination, the mapping component constitutes only a certain percentage of the changes in total light reflections. Acknowledgments This work has been supported by the Wellcome Trust (050080IZJ97). References [I] T. Bonhoeffer and A. Grinvald. Optical imaging based on intrinsic signals: The methodology. In A. Toga and J. C. Maziotta, editors, Brain mapping: The methods, pages 55-97, San Diego, CA, 1996. Academic Press, Inc. [2] G. G. Blasdel and G. Salama. Voltage-sensitive dyes reveal a modular organization in monkey striate cortex. Nature, 321 :579-585, 1986. [3] G. G. Blasdel. Differential imaging of ocular dominance and orientation selectivity in monkey striate cortex. 1. Neurosci., 12:3115-3138, 1992. [4] M. Stetter, T. Otto, T. Mueller, F. Sengpiel, M . Huebener, T. Bonhoeffer, and K. Obermayer. Temporal and spatial analysis of intrinsic signals from cat visual cortex. Soc. Neurosci. Abstr., 23:455,1997. [5] I. SchieGl, M. Stetter, J. E. W. Mayhew, S. Askew, N. McLoughlin, J. B. Levitt, J. S. Lund, and K. Obermayer. Blind separation of spatial signal patterns from optical imaging records. In J .-F. Cardoso, C. Jutten, and P. Loubaton, editors, Proceedings of the lCA99 workshop, volume I, pages 179-184, 1999. [6] A. J. Bell and T. J. Sejnowski. An information-maximization approach to blind separation and blind deconvolution. Neural Computation, 7: 1129-1159, 1995. [7] S. Amari . Neural learning in structured parameter spaces - natural riemannian gradient. In M. C. Mozer, M. I. Jordan, and T. Petsche, editors, Advances in Neural Information Processing Systems, volume 9, 1996. [8] A. Hyvtlrinen and E. Oja. A fast fixed point algorithm for independent component analysis. Neural Comput., 9:1483-1492,1997. [9] J. C. Platt and F. Faggin. Networks for the separation of sources that are superimposed and delayed. In 1. E. Moody, S. 1. Hanson, and R. P. Lippmann, editors, Advances in Neurallnformation Processing Systems, volume 4, pages 730--737, 1991. [10] L. Molgedey and H. G. Schuster. Separation of a mixture of independent signals using time delayed correlations. Phys. Rev. Lett., 72:3634-3637, 1994. [II] S. M. Riiger. Stable dynamic parameter adaptation. In D. S. Touretzky, M. C. Mozer, and M. E. Hasselmo, editors , Advances in Neural Information Processing Systems., volume 8, pages 225231 . MIT Press Cambridge, MA, 1996. [12] K.-R. Miiller, Philips P, and A. Ziehe. Jadetd: Combining higher-order statistics and temporal information for Blind Source Separation (with noise). In J.-F. Cardoso, C. Jutten, and P. Loubaton, editors, Proceedings of the 1. lCA99 Workshop, Aussois, volume I, pages 87-92, 1999. [13] B.-U. Koehler and R. Orglmeister. Independent component analysis using autoregressive models. In 1.-F. Cardoso, C. Jutten, and P. Loubaton, editors, Proceedings of the lCA99 workshop, volume I, pages 359-363, 1999.
1662 |@word trial:4 middle:5 version:2 norm:3 seems:1 simulation:1 mammal:1 shot:1 initial:1 schoner:1 series:2 contains:1 selecting:2 blank:1 od:3 activation:1 si:2 must:3 john:1 plot:6 selected:2 record:1 provides:1 preference:7 differential:1 incorrect:1 consists:1 introduce:2 multi:5 brain:7 underlying:2 medium:1 monkey:4 developed:1 temporal:3 questionable:2 uk:4 platt:1 before:2 local:2 limit:1 approximately:3 twice:1 initialization:1 specifying:1 limited:1 trice:1 averaged:1 unique:1 camera:1 acknowledgment:1 yj:1 block:1 procedure:2 area:1 bell:1 suggest:1 onto:1 close:2 selection:1 cannot:1 applying:1 optimize:1 conventional:2 map:10 demonstrated:2 go:1 regardless:1 starting:1 resolution:3 recovery:1 limiting:2 spontaneous:1 diego:1 us:2 element:3 particularly:1 coarser:1 labeled:1 bottom:1 region:1 cycle:1 removed:1 highest:1 noi:1 observes:1 mozer:2 transforming:1 ideally:1 dynamic:2 molgedey:1 completely:1 lowpass:2 cat:1 separated:3 fast:1 describe:1 london:1 sejnowski:1 artificial:7 oby:1 klaus:1 whose:2 quite:1 modular:1 plausible:1 otherwise:1 amari:1 otto:1 statistic:4 polak:1 noisy:2 advantage:2 biophysical:2 ucl:1 propose:2 reconstruction:7 adaptation:2 tu:2 combining:1 mixing:11 convergence:1 abstr:1 optimum:1 generating:1 depending:1 ac:2 measured:1 op:3 noticeable:1 mayhew:6 strong:1 soc:1 grating:3 c:2 implies:1 indicate:2 correct:1 human:1 everything:1 opt:4 biological:1 absorption:1 elementary:1 summation:4 mm:2 around:2 sufficiently:1 ic:1 visually:1 mapping:16 blasdel:2 lm:1 consecutive:1 smallest:1 currently:1 sensitive:2 largest:2 hasselmo:1 mit:1 sensor:8 always:1 varying:1 voltage:2 ribiere:1 sengpiel:1 superimposed:2 lasted:1 helpful:1 mueller:1 salama:1 germany:2 pixel:3 overall:1 orientation:9 denoted:1 smearing:1 spatial:21 field:1 extraction:3 holger:1 cancel:1 constitutes:1 future:2 minimized:1 others:1 stimulus:10 serious:1 few:1 loubaton:3 oja:1 delayed:2 argmax:1 consisting:1 attempt:1 organization:1 investigate:1 mixture:8 yielding:1 light:5 closer:1 respective:1 orthogonal:2 iv:1 re:3 causal:1 column:1 patchy:1 maximization:1 cost:3 snr:3 masked:1 successful:3 answer:1 faggin:1 combined:1 st:1 stay:1 off:1 ym:1 together:2 moody:1 squared:2 again:1 recorded:6 nm:1 containing:1 choose:1 worse:1 return:1 photon:1 de:2 sec:8 inc:1 toga:1 ad:3 blind:11 depends:1 onset:2 later:1 analyze:2 recover:1 square:1 variance:3 yield:3 raw:1 produced:2 tissue:3 simultaneous:1 phys:1 touretzky:1 against:2 frequency:3 ocular:8 obvious:1 jacobi:6 riemannian:1 newly:2 knowledge:2 amplitude:1 scattering:3 higher:2 methodology:1 improved:4 decorrelating:1 evaluated:2 just:1 correlation:12 hand:1 trust:1 autocorrelations:1 jutten:3 quality:2 reveal:1 artifact:9 normalized:2 hence:1 spatially:2 white:5 ll:1 during:1 width:1 reflection:2 image:17 superior:2 common:1 rotation:1 stimulation:1 volume:7 measurement:1 cambridge:1 smoothness:1 l3:1 stable:1 cortex:11 etc:1 dye:2 perspective:1 optimizes:1 selectivity:1 certain:1 yi:1 seen:1 greater:1 mcloughlin:6 subtraction:1 period:1 recommended:1 signal:21 ii:2 multiple:4 smooth:1 technical:2 academic:1 esd:2 cross:8 compensate:1 calculation:1 plausibility:2 concerning:2 post:1 converging:1 variant:3 sheffield:2 noiseless:2 represent:1 addition:1 singular:1 source:36 recording:13 db:3 contrary:1 jordan:1 extracting:2 neurallnformation:1 presence:1 iii:1 enough:1 shift:31 whether:1 motivated:1 o0:1 returned:1 miiller:1 generally:4 useful:1 se:1 covered:1 cardoso:3 concentrated:1 crosscorrelations:1 sl:1 goj:1 lsi:1 problematic:2 percentage:1 shifted:1 estimated:11 per:1 aussois:1 dominance:8 redundancy:2 four:1 blood:10 imaging:15 sum:3 run:2 comprehending:1 fourth:1 reasonable:1 separation:20 oscillation:1 scaling:1 illuminated:1 hi:1 followed:2 activity:13 adapted:1 occur:1 constraint:2 incorporation:1 wc:1 argument:3 performing:1 optical:19 martin:1 sphering:13 department:2 jio:1 structured:1 conjugate:1 jr:1 ophthalmology:1 remain:2 rev:1 explained:1 restricted:1 wellcome:1 equation:1 jennifer:1 describing:1 fail:2 cor:5 apply:2 appropriate:2 petsche:1 subtracted:1 alternative:2 robustness:2 drifting:1 original:3 denotes:1 top:3 remaining:1 ccd:1 calculating:1 reflectance:1 especially:4 build:1 classical:1 backprojection:1 added:2 koehler:1 primary:3 dependence:1 striate:3 diagonal:2 traditional:1 obermayer:6 gradient:10 separate:2 berlin:4 separating:1 philip:1 assuming:1 index:3 ratio:5 unknown:2 perform:2 upper:1 neuron:1 observation:3 sm:1 ingo:1 benchmark:3 descent:8 maxk:1 extended:4 dpal:1 frame:8 y1:1 highpass:1 stack:15 kl:1 hanson:1 timated:1 pop:1 macaque:2 able:4 bar:2 wy:1 pattern:7 usually:2 lund:6 orglmeister:1 unsuccessful:1 video:1 critical:1 decorrelation:9 natural:1 meantime:1 extract:1 auto:2 prior:3 removal:1 deoxyhemoglobin:1 stetter:6 permutation:4 filtering:6 versus:2 editor:7 uncorrelated:4 row:4 course:4 summary:1 supported:1 allow:1 institute:1 absolute:1 benefit:1 tolerance:1 bs:9 cortical:1 lett:1 computes:1 sensory:1 autoregressive:1 qualitatively:1 preprocessing:1 san:1 sj:3 approximate:1 lippmann:1 keep:2 global:5 assumed:1 search:1 additionally:1 nature:1 robust:8 contaminates:1 ca:1 sem:1 vessel:10 substituted:1 diag:3 neurosci:2 motivation:1 noise:34 unproblematic:1 levitt:1 slow:1 bandpass:3 grinvald:1 comput:1 vanish:1 specific:1 derives:2 intrinsic:4 exists:1 demixing:7 workshop:3 adding:1 deconvolution:1 ci:2 diagonalization:1 illumination:1 bonhoeffer:2 wavelength:2 visual:7 corresponds:1 extracted:1 jaco:1 ma:2 goal:1 presentation:1 identity:1 change:6 except:1 averaging:2 wt:1 total:4 called:2 e:1 select:2 college:1 ziehe:1 sphered:1 ongoing:1 schuster:1 correlated:1
725
1,663
Model Selection for Support Vector Machines Olivier Chapelle*,t, Vladimir Vapnik* * AT&T Research Labs, Red Bank, NJ t LIP6, Paris, France { chapelle, vlad} @research.au.com Abstract New functionals for parameter (model) selection of Support Vector Machines are introduced based on the concepts of the span of support vectors and rescaling of the feature space. It is shown that using these functionals, one can both predict the best choice of parameters of the model and the relative quality of performance for any value of parameter. 1 Introduction Support Vector Machines (SVMs) implement the following idea: they map input vectors into a high dimensional feature space, where a maximal margin hyperplane is constructed [6]. It was shown that when training data are separable, the error rate for SVMs can be characterized by (1) where R is the radius ofthe smallest sphere containing the training data and M is the margin (the distance between the hyperplane and the closest training vector in feature space). This functional estimates the VC dimension of hyperplanes separating data with a given margin M. To perform the mapping and to calculate Rand M in the SVM technique. one uses a positive definite kernel K(x, x') which specifies an inner product in feature space. An example of such a kernel is the Radial Basis Function (RBF). K(x, x') = e-llx-x'II2/20'2. This kernel has a free parameter (7 and more generally, most kernels require some parameters to be set. When treating noisy data with SVMs. another parameter. penalizing the training errors. also needs to be set. The problem of choosing the values of these parameters which minimize the expectation of test error is called the model selection problem. It was shown that the parameter of the kernel that minimizes functional (1) provides a good choice for the model: the minimum for this functional coincides with the minimum of the test error [1]. However. the shapes of these curves can be different. In this article we introduce refined functionals that not only specify the best choice of parameters (both the parameter of the kernel and the parameter penalizing training error). but also produce curves which better reflect the actual error rate. 231 Model Selection for Support Vector Machines The paper is organized as follows. Section 2 describes the basics of SVMs, section 3 introduces a new functional based on the concept of the span of support vectors, section 4 considers the idea of rescaling data in feature space and section 5 discusses experiments of model selection with these functionals. 2 Support Vector Learning We introduce some standard notation for SVMs; for a complete description, see [6]. Let (Xi, Yih <i<l be a set of training examples, Xi E jRn which belong to a class labeled by Yi E {- f, f}. The decision function given by a SVM is : (2) where the coefficients a? are obtained by maximizing the following functional: l I t W(a) = Lai - i=l 2' L aiajYiYjK(Xi,Xj) (3) i,j=l under constraints t L aiYi = 0 and 0 ~ ai ~ C i = 1, ... , f. i=l C is a constant which controls the tradeoff between the complexity of the decision function and the number of training examples misclassified. SVM are linear maximal margin classifiers in a high-dimensional feature space where the data are mapped through a non-linear function <p(x) such that <P(Xi) . <p(Xj) = K(Xi,Xj). The points Xi with ai > 0 are called support vectors. We distinguish between those with o < ai < C and those with ai = C. We call them respectively support vectors of the first and second category. 3 Prediction using the span of support vectors The results introduced in this section are based on the leave-one-out cross-validation estimate. This procedure is usually used to estimate the probability of test error of a learning algorithm. 3.1 The leave-one-out procedure The leave-one-out procedure consists of removing from the training data one element, constructing the decision rule on the basis of the remaining training data and then testing the removed element. In this fashion one tests all f elements of the training data (using f different decision rules). Let us denote the number of errors in the leave-one-out procedure by ?(Xl' Yl, .. . , Xl, Yl) . It is known [6] that the the leave-one-out procedure gives an almost unbiased estimate of the probability of test error: the expectation of test error for the machine trained on f - 1 examples is equal to the expectation of 1?(Xl' Yl, ... , Xl, Yt). We now provide an analysis of the number of errors made by the leave-one-out procedure. For this purpose, we introduce a new concept, called the span of support vectors [7]. O. Chapelle and V. N. Vapnik 232 3.2 Span of support vectors Since the results presented in this section do not depend on the feature space, we will consider without any loss of generality, linear SVMs, i.e. K (Xi, Xj) = Xi . Xj. Suppose that 0? = (a?, .. ., a~) is the solution of the optimization problem (3). For any fixed support vector xp we define the set Ap as constrained linear combinations of the support vectors of the first category (Xi)i:;t:p : .t Ai = 1, 0 t=l , t#p ~ a? + Yiypa~Ai ~ c} . (4) Note that Ai can be less than O. We also define the quantity Sp, which we call the span of the support vector xp as the minimum distance between xp and this set (see figure 1) (5) t... .. AI 2= +inf t...3 = -inf ?? ? '' - - ' ~ .. 2,, Figure 1: Three support vectors with al = a2 = a3/2. The set Al is the semi-opened dashed line. It was shown in [7] that the set Ap is not empty and that Sp = d(xp, Ap) ~ Dsv, where D sv is the diameter of the smallest sphere containing the support vectors. Intuitively, the smaller Sp = d(xp, Ap) is, the less likely the leave-one-out procedure is to make an error on the vector xp' Formally, the following theorem holds : Theorem 1 [7J If in the leave-one-out procedure a support vector xp corresponding to o < a p < C is recognized incorrectly, then the following inequality holds aO > p - 1 Sp max(D, 1/.JC)? This theorem implies that in the separable case (C = (0), the number of errors made by the leave-one-out procedure is bounded as follows: ?(Xl' Yl, .'" Xl, Yl) ~ 2: p a~ maxp SpD = maxp SpD / M2 , because 2: a~ = 1/M2 [6]. This is already an improvement compared to functional (I), since Sp ~ Dsv. But depending on the geometry of the support vectors the value of the span Sp can be much less than the diameter D sv of the support vectors and can even be equal to zero. We can go further under the assumption that the set of support vectors does not change during the leave-one-out procedure, which leads us to the following theorem: Model Selection for Support Vector Machines 233 Theorem 2 If the sets of support vectors of first and second categories remain the same during the leave-one-out procedure. then for any support vector xp. the following equality holds: yp[fO(xp) - fP(x p)] = o~S; where fO and fP are the decisionfunction (2) given by the SVM trained respectively on the whole training set and after the point xp has been removed. The proof of the theorem follows the one of Theorem 1 in [7]. The assumption that the set of support vectors does not change during the leave-one-out procedure is obviously not satisfied in most cases. Nevertheless, the proportion of points which violate this assumption is usually small compared to the number of support vectors. In this case, Theorem 2 provides a good approximation of the result of the leave-one procedure, as pointed out by the experiments (see Section 5.1, figure 2). As already noticed in [1], the larger op is, the more "important" in the decision function the support vector xp is. Thus, it is not surprising that removing a point xp causes a change in the decision function proportional to its Lagrange multiplier op . The same kind of result as Theorem 2 has also been derived in [2], where for SVMs without threshold, the following inequality has been derived: yp(f?(xp) - fP(xp)) ~ o~K(xp,xp). The span Sp takes into account the geometry of the support vectors in order to get a precise notion of how "important" is a given point. The previous theorem enables us to compute the number of errors made by the leave-oneout procedure: Corollary 1 Under the assumption of Theorem 2, the test error prediction given by the leave-one-out procedure is (6) Note that points which are not support vectors are correctly classified by the leave-one-out procedure. Therefore t/. defines the number of errors of the leave-one-out procedure on the entire training set. Under the assumption in Theorem 2, the box constraints in the definition of Ap (4) can be removed. Moreover, if we consider only hyperplanes passing through the origin, the constraint E Ai = 1 can also be removed. Therefore, under those assumptions, the computation of the span Sp is an unconstrained minimization of a quadratic form and can be done analytically. For support vectors of the first category, this leads to the closed form S~ = l/(K Mpp, where Ksv is the matrix of dot products between support vectors of the first category. A similar result has also been obtained in [3] . s In Section 5, we use the span-rule (6) for model selection in both separable and nonseparable cases. 4 Rescaling As we already mentioned, functional (1) bounds the VC dimension of a linear margin classifier. This bound is tight when the data almost "fills" the surface of the sphere enclosing the training data, but when the data lie on a flat ellipsoid, this bound is poor since the radius of the sphere takes into account only the components with the largest deviations. The idea we present here is to make a rescaling of our data in feature space such that the radius of the sphere stays constant but the margin increases, and then apply this bound to our rescaled data and hyperplane. 0. Chapelle and V. N. Vapnik 234 Let us first consider linear SVMs, i.e. without any mapping in a high dimensional space. The rescaling can be achieved by computing the covariance matrix of our data and rescaling according to its eigenvalues. Suppose our data are centered and let ('PI' ... ,'Pn) be the normalized eigenvectors of the covariance matrix of our data. We can then compute the smallest enclosing box containing our data, centered at the origin and whose edges are parallels to ('PI' ... , 'Pn)' This box is an approximation of the smallest enclosing ellipsoid. The length of the edge in the direction 'P k is J-Lk = maxi IXi . 'P k I. The rescaling consists of the following diagonal transformation: D :x --t Dx = LJ-Lk(X' 'Pk) 'Pk' k = Let us consider Xi = D-I xi and w Dw. The decision function is not changed under this transformation since w . Xi = W . xi and the data Xi fill a box of side length 1. Thus, in functional (l), we replace R2 by 1 and 1/M2 by w2 . Since we rescaled our data in a box, we actually estimated the radius of the enclosing ball using the foo-norm instead of the classical f 2 -norm. Further theoretical works needs to be done to justify this change of norm. In the non-linear case, note that even if we map our data in a high dimensional feature space, they lie in the linear subspace spanned by these data. Thus, if the number of training data f is not too large, we can work in this subspace of dimension at most f. For this purpose, one can use the tools of kernel PCA [5] : if A is the matrix of normalized eigenvectors of the Gram matrix Kij = K (Xi, Xj) and (>'d the eigenvalues, the dot product Xi . 'P k is replaced by v'XkAik and W? 'Pk becomes v'XkL:i AikYiO'i. Thus, we can still achieve the diagonal transformation A and finally functional (1) becomes L k >.~ max Ark ~ (2: A ik YiO'i)2 . i 5 Experiments To check these new methods, we performed two series of experiments. One concerns the choice of (7, the width of the RBF kernel, on a linearly separable database, the postal database. This dataset consists of 7291 handwritten digit of size 16x16 with a test set of 2007 examples. Following [4], we split the training set in 23 subsets of 317 training examples. Our task consists of separating digit 0 to 4 from 5 to 9. Error bars in figures 2a and 3 are standard deviations over the 23 trials. In another experiment, we try to choose the optimal value of C in a noisy database, the breast-cancer database! . The dataset has been split randomly 100 times into a training set containing 200 examples and a test set containing 77 examples. Section 5.1 describes experiments of model selection using the span-rule (6), both in the separable case and in the non-separable one, while Section 5.2 shows VC bounds for model selection in the separable case both with and without rescaling. 5.1 Model selection using the span-rule In this section, we use the prediction of test error derived from the span-rule (6) for model selection. Figure 2a shows the test error and the prediction given by the span for different values of the width (7 of the RBF kernel on the postal database. Figure 2b plots the same functions for different values of C on the breast-cancer database. We can see that the method predicts the correct value of the minimum. Moreover, the prediction is very accurate and the curves are almost identical. I Available from http; I Ihorn. first. gmd. del "'raetsch/da ta/breast-cancer 235 Model Selection for Support Vector Machines 40,-----~---~r=_=_-""T==es=:'t=er=ro=r= ' l l - 35 Span prediction 30 25 g20 UJ 15 i" ", 10 5 ~6 -4 -2 0 Log sigma 2 4 (a) choice of (T in the postal database 6 o 2 4 6 8 10 12 Loge (b) choice of C in the breast-cancer database Figure 2: Test error and its prediction using the span-rule (6). The computation of the span-rule (6) involves computing the span Sp (5) for every support Yp!(xp)/a~, rather vector. Note, however, that we are interested in the inequality than the exact value of the span Sp. Thus, while minimizing Sp = d(xp, Ap), if we find a point x* E Ap such that d(xp, x*)2 ::; Yp! (xp ) / a~, we can stop the minimization because this point will be correctly classified by the leave-one-out procedure. S; ::; It turned out in the experiments that the time required to compute the span was not prohibitive, since it is was about the same than the training time. There is a noteworthy extension in the application of the span concept. If we denote by one hyperparameter of the kernel and if the derivative 8K(;~'Xi) is computable, then it e is possible to compute analytically 8 ~ aiS~~y;fO(x;) , which is the derivative of an upper bound of the number of errors made by the leave-one-out procedure (see Theorem 2). This provides us a more powerful technique in model selection. Indeed, our initial approach was to choose the value of the width (T of the RBF kernel according to the minimum of the span-rule. In our case, there was only hyperparamter so it was possible to try different values of (T. But, if we have several hyperparameters, for example one (T per component, _~ (Xk- X j,)2 2<T~ , it is not possible to do an exhaustive search on all the possible values of of the hyperparameters. Nevertheless, the previous remark enables us to find their optimal value by a classical gradient descent approach. K(x, x') = e k Preliminary results seem to show that using this approach with the previously mentioned kernel improve the test error significantely. 5.2 VC dimension with rescaling In this section, we perform model selection on the postal database using functional (1) and its rescaled version. Figure 3a shows the values of the classical bound R2 / M2 for different values of (T. This bound predicts the correct value for the minimum, but does not reflect the actual test error. This is easily understandable since for large values of (T, the data in input space tend to be mapped in a very flat ellipsoid in feature space, a fact which is not taken into account [4]. Figure 3b shows that by performing a rescaling of our data, we manage to have a much tighter bound and this curve reflects the actual test error, given in figure 2a. 0. Chape/le and V. N. Vapnik 236 120 18000'---~--~--~--r=~==~=.=~ 16000 - VC Dimension with rescali 100 14000 12000 80 E E 10000 ~ 8000 ~ 60 > 6000 40 '6 4000 20 2000 ~L-==~===c~~ ....., -4 -2 ____~__~__~ 0 Log sigma 2 (a) without rescaling 4 6 ~ -4 -2 0 Log sigma 2 4 6 (b) with rescaling Figure 3: Bound on the VC dimension for different values of ~ on the postal database. The shape of the curve with rescaling is very similar to the test error on figure 2. 6 Conclusion In this paper, we introduced two new techniques of model selection for SVMs. One is based on the span, the other is based on rescaling of the data in feature space. We demonstrated that using these techniques, one can both predict optimal values for the parameters of the model and evaluate relative performances for different values of the parameters. These functionals can also lead to new learning techniques as they establish that generalization ability is not only due to margin. Acknowledgments The authors would like to thank Jason Weston and Patrick Haffner for helpfull discussions and comments. References [1] C. J. C. Burges. A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, 2(2):121-167, 1998. [2] T. S. Jaakkola and D. Haussler. Probabilistic kernel regression models. In Proceedings of the J999 Conference on AI and Statistics, 1999. [3] M. Opper and O. Winther. Gaussian process classification and SVM: Mean field results and leave-one-out estimator. In Advances in Large Margin Classifiers. MIT Press, 1999. to appear. [4] B. SchOlkopf, J. Shawe-Taylor, A. 1. Smola, and R. C. Williamson. Kernel-dependent Support Vector error bounds. In Ninth International Conference on Artificial Neural Networks, pp. 304 309 [5] B. SchOlkopf, A. Smola, and K.-R. Muller. Kernel principal component analysis. In Artificial Neural Networks -ICANN'97, pages 583 - 588, Berlin, 1997. Springer Lecture Notes in Computer Science, Vol. 1327. [6] V. Vapnik. Statistical Learning Theory. Wiley, New York, 1998. [7] V. Vapnik and O. Chapelle. Bounds on error expectation for SVM. Neural Computation, 1999. Submitted.
1663 |@word trial:1 version:1 proportion:1 norm:3 covariance:2 yih:1 initial:1 series:1 com:1 surprising:1 dx:1 shape:2 enables:2 treating:1 plot:1 prohibitive:1 xk:1 provides:3 postal:5 hyperplanes:2 constructed:1 ik:1 scholkopf:2 consists:4 introduce:3 indeed:1 nonseparable:1 actual:3 becomes:2 notation:1 bounded:1 moreover:2 kind:1 minimizes:1 transformation:3 nj:1 every:1 ro:1 classifier:3 control:1 appear:1 positive:1 ap:7 noteworthy:1 au:1 acknowledgment:1 testing:1 implement:1 definite:1 digit:2 procedure:19 yio:1 radial:1 get:1 selection:15 hyperparamter:1 map:2 demonstrated:1 yt:1 maximizing:1 go:1 m2:4 rule:9 haussler:1 estimator:1 fill:2 spanned:1 dw:1 notion:1 suppose:2 exact:1 olivier:1 us:1 ixi:1 origin:2 element:3 recognition:1 ark:1 predicts:2 labeled:1 database:10 calculate:1 removed:4 rescaled:3 mentioned:2 complexity:1 trained:2 depend:1 tight:1 basis:2 easily:1 artificial:2 choosing:1 refined:1 exhaustive:1 whose:1 larger:1 maxp:2 ability:1 statistic:1 noisy:2 obviously:1 eigenvalue:2 g20:1 maximal:2 product:3 turned:1 achieve:1 description:1 empty:1 produce:1 leave:20 depending:1 op:2 involves:1 implies:1 direction:1 radius:4 correct:2 opened:1 vc:6 centered:2 require:1 ao:1 generalization:1 preliminary:1 tighter:1 extension:1 hold:3 mapping:2 predict:2 smallest:4 a2:1 purpose:2 lip6:1 largest:1 tool:1 reflects:1 minimization:2 mit:1 gaussian:1 rather:1 pn:2 jaakkola:1 dsv:2 corollary:1 derived:3 improvement:1 check:1 dependent:1 entire:1 lj:1 misclassified:1 france:1 interested:1 classification:1 constrained:1 equal:2 field:1 identical:1 randomly:1 replaced:1 geometry:2 mining:1 introduces:1 accurate:1 edge:2 taylor:1 loge:1 theoretical:1 kij:1 deviation:2 subset:1 too:1 sv:2 winther:1 international:1 stay:1 probabilistic:1 yl:5 reflect:2 satisfied:1 manage:1 containing:5 choose:2 derivative:2 rescaling:14 yp:4 account:3 coefficient:1 jc:1 performed:1 try:2 jason:1 lab:1 closed:1 red:1 parallel:1 minimize:1 ofthe:1 handwritten:1 classified:2 submitted:1 fo:3 definition:1 pp:1 proof:1 stop:1 dataset:2 vlad:1 knowledge:1 organized:1 actually:1 ta:1 specify:1 rand:1 done:2 box:5 generality:1 smola:2 del:1 defines:1 quality:1 concept:4 unbiased:1 multiplier:1 normalized:2 equality:1 analytically:2 during:3 width:3 coincides:1 complete:1 xkl:1 aiyi:1 functional:10 belong:1 raetsch:1 llx:1 ai:11 unconstrained:1 pointed:1 shawe:1 dot:2 chapelle:5 surface:1 patrick:1 closest:1 inf:2 inequality:3 yi:1 muller:1 minimum:6 recognized:1 dashed:1 semi:1 violate:1 characterized:1 cross:1 sphere:5 lai:1 prediction:7 basic:1 regression:1 breast:4 expectation:4 kernel:15 achieved:1 w2:1 comment:1 tend:1 seem:1 call:2 split:2 xj:6 spd:2 inner:1 idea:3 haffner:1 tradeoff:1 mpp:1 computable:1 pca:1 passing:1 cause:1 york:1 remark:1 generally:1 eigenvectors:2 jrn:1 svms:9 category:5 diameter:2 gmd:1 http:1 specifies:1 tutorial:1 estimated:1 correctly:2 per:1 hyperparameter:1 vol:1 nevertheless:2 threshold:1 penalizing:2 powerful:1 almost:3 decision:7 bound:12 distinguish:1 quadratic:1 constraint:3 flat:2 span:23 performing:1 separable:7 according:2 combination:1 poor:1 ball:1 describes:2 smaller:1 remain:1 intuitively:1 taken:1 previously:1 discus:1 available:1 apply:1 remaining:1 uj:1 establish:1 classical:3 noticed:1 already:3 quantity:1 aiajyiyjk:1 diagonal:2 gradient:1 subspace:2 distance:2 thank:1 mapped:2 separating:2 berlin:1 considers:1 length:2 ellipsoid:3 minimizing:1 vladimir:1 sigma:3 oneout:1 enclosing:4 understandable:1 perform:2 upper:1 descent:1 incorrectly:1 precise:1 ninth:1 introduced:3 paris:1 required:1 bar:1 usually:2 pattern:1 fp:3 max:2 improve:1 lk:2 ii2:1 discovery:1 relative:2 loss:1 lecture:1 proportional:1 validation:1 xp:20 article:1 bank:1 pi:2 cancer:4 changed:1 free:1 side:1 burges:1 curve:5 dimension:6 opper:1 gram:1 author:1 made:4 functionals:5 xi:17 search:1 williamson:1 constructing:1 da:1 sp:11 pk:3 icann:1 linearly:1 whole:1 hyperparameters:2 fashion:1 x16:1 wiley:1 foo:1 xl:6 lie:2 theorem:13 removing:2 er:1 maxi:1 r2:2 svm:6 a3:1 concern:1 vapnik:6 margin:8 likely:1 lagrange:1 springer:1 weston:1 rbf:4 replace:1 change:4 hyperplane:3 ksv:1 justify:1 principal:1 called:3 e:1 formally:1 support:35 evaluate:1
726
1,664
Approximate Planning in Large POMDPs via Reusable Trajectories Michael Kearns AT&T Labs mkearns@research.att.com Yishay Mansour Tel Aviv University mansour@math.tau.ac.il AndrewY. Ng UC Berkeley ang@cs.berkeley.edu Abstract We consider the problem of reliably choosing a near-best strategy from a restricted class of strategies TI in a partially observable Markov decision process (POMDP). We assume we are given the ability to simulate the POMDP, and study what might be called the sample complexity that is, the amount of data one must generate in the POMDP in order to choose a good strategy. We prove upper bounds on the sample complexity showing that, even for infinitely large and arbitrarily complex POMDPs, the amount of data needed can be finite, and depends only linearly on the complexity of the restricted strategy class TI, and exponentially on the horizon time. This latter dependence can be eased in a variety of ways, including the application of gradient and local search algorithms. Our measure of complexity generalizes the classical supervised learning notion of VC dimension to the settings of reinforcement learning and planning. 1 Introduction Much recent attention has been focused on partially observable Markov decision processes (POMDPs) which have exponentially or even infinitely large state spaces. For such domains, a number of interesting basic issues arise. As the state space becomes large, the classical way of specifying a POMDP by tables of transition probabilities clearly becomes infeasible. To intelligently discuss the problem of planning - that is, computing a good strategy 1 in a given POMDP - compact or implicit representations of both POMDPs, and of strategies in POMDPs, must be developed. Examples include factored next-state distributions [2, 3, 7], and strategies derived from function approximation schemes [8]. The trend towards such compact representations, as well as algorithms for planning and learning using them, is reminiscent of supervised learning, where researchers have long emphasized parametric models (such as decision trees and neural networks) that can capture only limited structure, but which enjoy a number of computational and information-theoretic benefits. Motivated by these issues, we consider a setting were we are given a generative model, or lThroughout, we use the word strategy to mean any mapping from observable histories to actions, which generalizes the notion of policy in a fully observable MDP. M Kearns. Y. Mansour and A. Y. Ng 1002 simulator, for a POMDP, and wish to find a good strategy 7r from some restricted class of strategies II . A generative model is a "black box" that allows us to generate experience (trajectories) from different states of our choosing. Generative models are an abstract notion of compact POMDP representations, in the sense that the compact representations typically considered (such as factored next-state distributions) already provide efficient generative models. Here we are imagining that the strategy class II is given by some compact representation or by some natural limitation on strategies (such as bounded memory). Thus, the view we are adopting is that even though the world (POMDP) may be extremely complex, we assume that we can at least simulate or sample experience in the world (via the generative model), and we try to use this experience to choose a strategy from some "simple" class II. We study the following question: How many calls to a generative model are needed to have enough data to choose a near-best strategy in the given class? This is analogous to the question of sample complexity in supervised learning - but harder. The added difficulty lies in the reuse of data. In supervised learning, every sample (x, f(x)) provides feedback about every hypothesis function h(x) (namely, how close h(x) is to f(x)) . If h is restricted to lie in some hypothesis class 1i, this reuse permits sample complexity bounds that are far smaller than the size of 1i. For instance, only O(log(I1il)) samples are needed to choose a near-best model from a finite class 1i. If 1i is infinite, then sample sizes are obtained that depend only on some measure of the complexity of1i (such as VC dimension [9]), but which have no dependence on the complexity of the target function or the size of the input domain. In the POMDP setting, we would like analogous sample complexity bounds in terms of the "complexity" of the strategy class II - bounds that have no dependence on the size or complexity of the POMDP. But unlike the supervised learning setting, experience "reuse" is not immediate in POMDPs. To see this, consider the "straw man" algorithm that, starting with some 7r E II, uses the generative model to generate many trajectories under 7r, and thus forms a Monte Carlo estimate of V 7r (so). It is not clear that these trajectories under 7r are of much use in evaluating a different 7r' E II, since 7r and 7r' may quickly disagree on which actions to take. The naive Monte Carlo method thus gives 0(1111) bounds on the "sample complexity," rather than O(log(IIII)), for the finite case. In this paper, we shall describe the trajectory tree method of generating "reusable" trajectories, which requires generating only a (relatively) small number of trajectories - a number that is independent of the state-space size of the POMDP, depends only linearly on a general measure of the complexity of the strategy class II, and depends exponentially on the horizon time. This latter dependence can be eased via gradient algorithms such as Williams' REINFORCE [10] and Baird and Moore's more recent YAPS [1], and by local search techniques. Our measure of strategy class complexity generalizes the notion of VC dimension in supervised learning to the settings of reinforcement learning and planning, and we give bounds that recover for these settings the most powerful analogous results in supervised learning - bounds for arbitrary, infinite strategy classes that depend only on the dimension of the class rather than the size of the state space. 2 Preliminaries We begin with some standard definitions. A Markov decision process (MDP) is a tuple (S, So, A , {P (,1 s, a)}, R), where: S is a (possibly infinite) state set; So E S is a start state; A {al' . .. ,ad are actions; PC Is, a) gives the next-state distribution upon taking action a from state s; and the reward function R(s, a) gives the corresponding rewards. We assume for simplicity that rewards are deterministic, and further that they are bounded = Approximate Planning in Large POMDPs via Reusable Trajectories 1003 in absolute value by Rmax. A partially observable Markov decision process (POMDP) consists of an underlying MOP and observation distributions Q(ols) for each state s, where 0 is the random observation made at s. We have adopted the common assumption of a fixed start state,2 because once we limit the class of strategies we entertain, there may not be a single "best" strategy in the classdifferent start states may have different best strategies in II. We also assume that we are given a POMOP M in the form of a generative model for M that, when given as input any state-action pair (s, a), will output a state S' drawn according to P(?ls, a), an observation o drawn according to Q(?ls), and the reward R(s, a). This gives us the ability to sample the POMOP M in a random-access way. This definition may initially seem unreasonably generous: the generative model is giving us a fully observable simulation of a partially observable process. However, the key point is that we must still find a strategy that performs well in the partially observable setting. As a concrete example, in designing an elevator control system, we may have access to a simulator that generates random rider arrival times, and keeps track of the waiting time of each rider, the number of riders waiting at every floor at every time of day, and so on. However helpful this information might be in designing the controller, this controller must only use information about which floors currently have had their call button pushed (the observables). In any case, readers uncomfortable with the power provided by our generative models are referred to Section 5, where we briefly describe results requiring only an extremely weak form of partially observable simulation. At any time t, the agent will have seen some sequence of observations, 00,??., Ot, and will have chosen actions and received rewards for each of the t time We write its observable history as h steps prior to the current one. (( 00, ao, TO), ... , (Ot-l , at-I, Tt-l ), (Ot, _, _)). Such observable histories, also called trajectories, are the inputs to strategies. More formally, a strategy 7r is any (stochastic) mapping from observable histories to actions. (For example, this includes approaches which use the observable history to track the belief state [5].) A strategy class II is any set of strategies. We will restrict our attention to the case of discounted return,3 and we let, E [0,1) be the discount factor. We define the t::-horizon time to be HE = 10gl'(t::(1 - ,)/2Rmax ). Note that returns beyond the first HE-steps can contribute at most t::/2 to the total discounted return. Also, let Vmax = Rmax/(l - ,) bound the value function. Finally, for a POMDP M and a strategy class II, we define opt(M, II) = SUP7rEII V7r (so) to be the best expected return achievable from So using II. Our problem is thus the following: Given a generative model for a POMOP M and a strategy class II, how many calls to the generative model must we make, in order to have enough data to choose a 7r E II whose performance V7r(so) approaches opt(M, II)? Also, which calls should we make to the generative model to achieve this? 3 The Trajectory Tree Method We now describe how we can use a generative model to create "reusable" trajectories. For ease of exposition, we assume there are only two actions al and a2, but our results generalize easily to any finite number of actions. (See the full paper [6].) 2 An equivalent definition is to assume a fixed distribution D over start states, since So can be a "dummy" state whose next-state distribution under any action is D. 3The results in this paper can be extended without difficulty to the undiscounted finite-horizon setting [6]. M. Keams, Y. Mansour and A. Y. Ng 1004 A trajectory tree is a binary tree in which each node is labeled by a state and observation pair, and has a child for each of the two actions. Additionally, each link to a child is labeled by a reward, and the tree's depth will be H~, so it will have about 2H e nodes. (In Section 4, we will discuss settings where this exponential dependence on H~ can be eased.) Each trajectory tree is built as follows: The root is labeled by So and the observation there, 00 ' Its two children are then created by calling the generative model on (so, ad and (so, a2), which gives us the two next-states reached (say s~ and s~ respectively), the two observations made (say o~ and o~), and the two rewards received (r~ R(so, ad and r~ = R(so, a2). Then (s~ , aD and (s~, o~) label the root's aI-child and a2-child, and the links to these children are labeled r~ and r~. Recursively, we generate two children and rewards this way for each node down to depth H~ . = Now for any deterministic strategy tr and any trajectory tree T, tr defines a path through tr starts at the root, and inductively, if tr is at some internal node in T, then we feed to tr the observable history along the path from the root to that node, and tr selects and moves to a child of the current node. This continues until a leaf node is reached, and we define R( tr , T) to be the discounted sum of returns along the path taken. In the case that tr is stochastic, tr defines a distribution on paths in T, and R(tr , T) is the expected return according to this distribution. (We will later also describe another method for treating stochastic strategies.) Hence, given m trajectory trees T 1 , ... , T m, a natural estimate for V7r(so) is V7r(so) = ,; 2:::1 R(tr, Ti). Note that each tree can be used to evaluate any strategy, much the way a single labeled example (x , f(x)) can be used to evaluate any hypothesis h(x) in supervised learning. Thus in this sense, trajectory trees are reusable. T: Our goal now is to establish uniform convergence results that bound the error of the estimates V7r (so) as a function of the "sample size" (number of trees) m. Section 3.1 first treats the easier case of deterministic classes II; Section 3.2 extends the result to stochastic classes. 3.1 The Case of Deterministic II Let us begin by stating a result for the special case of finite classes of deterministic strategies, which will serve to demonstrate the kind of bound we seek. Theorem 3.1 Let II be any finite class of deterministic strategies for an arbitrary twoaction POMDP M. Let m trajectory trees be created using a generative modelfor M, and V7r(so) be the resulting estimates. lfm = 0 ((Vrnax /t)2(log(IIII) + log(1/8))), then with probability 1 - 8, IV7r (so) - V7r (so) I t holds simultaneously for alltr E II. :s Due to space limitations, detailed proofs of the results of this section are left to the full paper [6] , but we will try to convey the intuition behind the ideas. Observe that for any fixed deterministic tr, the estimates R( tr, Ti) that are generated by the m different trajectory trees Ti are independent. Moreover, each R(tr, T i ) is an unbiased estimate of the expected discounted H~ -step return of tr, which is in turn t/2-close to V7r(so). These observations, combined with a simple Chernoff and union bound argument, are sufficient to establish Theorem 3.1. Rather than developing this argument here, we instead move straight on to the harder case of infinite II. When addressing sample complexity in supervised learning, perhaps the most important insight is that even though a class 1i may be infinite, the number of possible behaviors of 1i on a finite set of points is often not exhaustive. More precisely, for boolean functions, we say that the set Xl, ... , Xd is shattered by 1i if every of the 2d possible labelings of Approximate Planning in Large POMDPs via Reusable Trajectories 1005 these points is realized by some h E 1i. The VC dimension of 1i is then defined as the size of the largest shattered set [9]. It is known that if the VC dimension of 1i is d, then the number <P d(m) of possible labelings induced by 1i on a set of m points is at most (em Jd)d, which is much less than 2 m for d ? m. This fact provides the key leverage exploited by the classical VC dimension results, and we will concentrate on replicating this leverage in our setting. If II is a (possibly infinite) set of deterministic strategies, then each strategy tr E II is simply a deterministic function mapping from the set of observable histories to the set {al' a2}, and is thus a boolean function on observable histories. We can therefore write VC(II) to denote the familiar VC dimension of the set of binary functions II. For example, if II is the set of all thresholded linear functions of the current vector of observations (a particular type of memoryless strategy), then VC(II) simply equals the number of parameters. We now show intuitively why a class II of bounded VC dimension d cannot induce exhaustive behavior on a set T l , ... ,Tm of trajectory trees for m ? d. Note that if trl, tr2 E II are such that their "reward labelings" (R(trl' T l ), ... ,R(trl' T m)) and (R( tr2, Tt), ... , R( tr2, T m)) differ, then R( trl, Ti) =f. R(tr2' T i ) for some 1 ::; i ::; m. But if trl and tr2 give different returns on T i , then they must choose different actions at some node in T i . In other words, every different reward labeling of the set of m trees yields a different (binary) labeling of the set of m . 2H ? observable histories in the trees. So, the number of different tree reward labelings can be at most <Pd(m? 2H<) ::; (em? 2H<Jd)d. By developing this argument carefully and applying classical uniform convergence techniques, we obtain the following theorem. (Full proof in [6].) Theorem 3.2 Let II be any class of deterministic strategies for an arbitrary two-action POMDP M, and let VC(II) denote its VC dimension. Let m trajectory trees be created using a generative model for M, and "\I7r (so) be the resulting estimates. If (1) then with probability 1 - 6, IV 7r (so) - "\I7r (so) I ::; ? holds simultaneously for alltr E II. 3.2 The Case of Stochastic II We now address the case of stochastic strategy classes. We describe an approach where we transform stochastic strategies into "equivalent" deterministic ones and operate on the deterministic versions, reducing the problem to the one handled in the previous section. The transformation is as follows: Given a class of stochastic strategies II, each with domain X (where X is the set of all observable histories), we first extend the domain to be X x [0,1]. Now for each stochastic strategy tr E II, define a corresponding deterministic transformed strategy tr' with domain X x [0,1], given by: tr'(h, r) = al if r ::; Pr[tr(h) = ad, and 7r'(h,r) = a2 otherwise (for any hEX, r E [0,1]). Let II' be the collection of these transformed deterministic strategies tr'. Since II' is just a set of deterministic boolean functions, its VC dimension is well-defined. We then define the pseudo-dimension of the original set of stochastic strategies II to be p VC(II) = VC(II').4 Having transformed the strategy class, we also need to transform the POMDP, by augmenting the state space S to be S x [0,1]. Informally, the transitions and rewards remain the same, except that after each state transition, we draw a new random variable r uniformly in [0,1], and independently of all previous events. States are now of the form (s, r), and we let r be an observed variable. Whenever in the original POMDP a stochastic strategy tr would 4This is equivalent to the conventional definition of the pseudo-dimension of IT [4], when it is viewed as a set of maps into real-valued action-probabilities. M Kearns, Y. Mansour and A. Y. Ng 1006 have been given a history h, in the transformed POMDP the corresponding deterministic transformed strategy 7r' is given (h, r), where r is the [0, l]-random variable at the current state. By the definition of 7r', it is easy to see that 7r' and 7r have exactly the same chance of choosing each action at any node (randomization over r). We are now back in the deterministic case, so Theorem 3.2 applies, with VC(II) replaced by pVC (II) = VC(II'), and we again have the desired uniform convergence result. 4 Algorithms for Approximate Planning Given a generative model for a POMDP, the preceding section's results immediately suggest a class of approximate planning algorithms: generate m trajectory trees T 1 , ... , T m, and search for a 7r E II that maximizes V7r (so) = (1/ m) L R( 7r, Ti). The following corollary to the uniform convergence results establishes the soundness of this approach. Corollary 4.1 Let II be a class of strategies in a POMDP M, and let the number m of trajectory trees be as given in Theorem 3.2. Let it = argmax 7r Err{V7r(so)} be the policy in II with the highest empirical return on the m trees. Then with probability 1 - 0, it is near-optimal within II: (2) V7T(SO) ~ opt(M, II) - 2?. If the suggested maximization is computationally infeasible, one can search for a local maximum 7r instead, and uniform convergence again assures us that V7r (so) is a trusted estimate of our true performance. Of course, even finding a local maximum can be expensive, since each trajectory tree is of size exponential in H{. However, in practice it may be possible to significantly reduce the cost of the search. Suppose we are using a class of (possibly transformed) deterministic strategies, and we perform a greedy local search over II to optimize V7r (so). Then at any time in the search, to evaluate the policy we are currently considering, we really need to look at only a single path of length Hf in each tree, corresponding to the path taken by the strategy being considered. Thus, we should build the trajectory trees lazily - that is, incrementally build each node of each tree only as it is needed to evaluate R( 7r, Ti) for the current strategy 7r. If there are parts of a tree that are reached only by poor policies, then a good search algorithm may never even build these parts of the tree. In any case, for a fixed number of trees, each step of the local search now takes time only linear in H f ? 5 There is a different approach that works directly on stochastic strategies (that is, without requiring the transformation to deterministic strategies). In this case each stochastic strategy 7r defines a distribution over all the paths in a trajectory tree, and thus calculating R( 7r, T) may in general require examining complete trees. However, we can view each trajectory tree as a small, deterministic POMDP by itself, with the children of each node in the tree being its successor nodes. So if II = {7re : E IRd} is a smoothly parameterized family of stochastic strategies, then algorithms such as William's REINFORCE [10] can be used to find an unbiased estimate of the gradient (d/ de) V7r 9 (so), which in turn can be used to e 5 See also (Ng and Jordan, in preparation) which, by assuming a much stronger model of a POMDP (a deterministic function 1 such that I(s, a, r) is distributed according to P('ls, a) when r is distributed Uniform[O,l]), gives an algorithm that enjoys uniform convergence bounds similar to those presented here, but with only a polynomial rather than exponential dependence on H,. The algorithm samples a number of vectors r(i) E [0, IjH., each of which, with I, defines an H,-step Monte Carlo evaluation trial for any policy 7r . The bound is on the number of such random vectors needed (rather than on the total number of calls to f). Approximate Planning in Large POMDPs via Reusable Trajectories 1007 perform stochastic gradient ascent to maximize V7r 8 (so). Moreover, for a fixed number of trees, these algorithms need only O(H?) time per gradient estimate; so combined with lazy tree construction, we again have a practical algorithm whose per-step complexity is only linear in the horizon time. This line of thought is further developed in the long version of the paper.6 5 The Random Trajectory Method Using a fully observable generative model of a POMDP, we have shown that the trajectory tree method gives uniformly good value estimates, with an amount of experience linear in VC(II), and exponential in H?. It turns out we can significantly weaken the generative model, yet still obtain essentially the same theoretical results. In this harder case, we assume a generative model that provides only partially observable histories generated by a truly random strategy (which takes each action with equal probability at every step, regardless of the history so far). Furthermore, these trajectories always begin at the designated start state, so there is no ability provided to "reset" the POMDP to any state other than so. (Indeed, underlying states may never be observed.) Our method for this harder case is called the Random Trajectory method. It seems to lead less readily to practical algorithms than the trajectory tree method, and its formal description and analysis, which is more difficult than for trajectory trees, are given in the long version of this paper [6]. As in Theorem 3.2, we prove that the amount of data needed is linear in VC(II), and exponential in the horizon time - that is, by averaging appropriately over the resulting ensemble of trajectories generated, this amount of data is sufficient to yield uniformly good estimates of the values for all strategies in II. References [1] L. Baird and A. W. Moore. Gradient descent for general Reinforcement Learning. In Advances in Neural Information Processing Systems 11, 1999. [2] C. Boutilier, T. Dean, and S. Hanks. Decision theoretic planning: Structural assumptions and computational leverage. Journal of Artificial Intelligence Research, 1999. [3] X. Boyen and D. Koller. Tractable inference for complex stochastic processes. In Proc. UAI, pages 33-42, 1998. [4] David Haussler. Decision theoretic generalizations of the PAC model for neural net and oter learning applications. Information and Computation, 100:78-150, 1992. [5] L. P. Kaelbling, M. L. Littman, and A. R. Cassandra. Planning and acting in partially observable stochastic domains. ArtifiCial Intelligence, 101, 1998. [6] M. Kearns, Y. Mansour, and A. Y. Ng. Approximate planning in large POMDPs via reusable trajectories. (long version), 1999. [7] D. Koller and R. Parr. Computing factored value functions for poliCies in structured MDPs. In Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence, 1999. [8] R. S. Sutton and A. G. Barto. Reinforcement Learning. MIT Press, 1998. [9] Y.N. Vapnik. Estimation of Dependences Based on Empirical Data. Springer-Verlag, 1982. [10] R. J. Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8:229-256, 1992. 6In the full paper, we also show how these algorithms can be extended to find in expected O( He) time an unbiased estimate of the gradient of the true value V 8 (so) for discounted infinite horizon problems (whereas most current algorithms either only converge asymptotically to an unbiased estimate of this gradient, or need an absorbing state and "proper" strategies). 7T
1664 |@word trial:1 version:4 briefly:1 achievable:1 stronger:1 polynomial:1 seems:1 simulation:2 seek:1 tr:22 harder:4 recursively:1 mkearns:1 att:1 err:1 current:6 com:1 yet:1 must:6 readily:1 reminiscent:1 treating:1 generative:21 leaf:1 greedy:1 intelligence:3 provides:3 math:1 contribute:1 node:12 along:2 prove:2 consists:1 indeed:1 expected:4 behavior:2 planning:13 simulator:2 discounted:5 considering:1 becomes:2 begin:3 provided:2 bounded:3 underlying:2 moreover:2 maximizes:1 v7t:1 what:1 kind:1 rmax:3 developed:2 finding:1 transformation:2 pseudo:2 berkeley:2 every:7 ti:8 xd:1 exactly:1 uncomfortable:1 control:1 twoaction:1 enjoy:1 local:6 treat:1 limit:1 pomop:3 sutton:1 path:7 might:2 black:1 specifying:1 ease:1 limited:1 practical:2 union:1 practice:1 empirical:2 significantly:2 thought:1 word:2 induce:1 suggest:1 cannot:1 close:2 applying:1 optimize:1 equivalent:3 deterministic:21 conventional:1 map:1 dean:1 williams:2 attention:2 starting:1 l:3 independently:1 pomdp:24 focused:1 regardless:1 simplicity:1 immediately:1 factored:3 insight:1 haussler:1 notion:4 analogous:3 yishay:1 target:1 suppose:1 construction:1 us:1 designing:2 hypothesis:3 trend:1 yap:1 expensive:1 continues:1 labeled:5 observed:2 capture:1 highest:1 intuition:1 pd:1 complexity:16 reward:12 littman:1 inductively:1 depend:2 serve:1 upon:1 observables:1 easily:1 joint:1 describe:5 monte:3 artificial:3 labeling:2 choosing:3 exhaustive:2 whose:3 valued:1 say:3 otherwise:1 ability:3 soundness:1 transform:2 itself:1 sequence:1 intelligently:1 net:1 reset:1 achieve:1 sixteenth:1 description:1 convergence:6 undiscounted:1 generating:2 ac:1 augmenting:1 stating:1 received:2 c:1 of1i:1 differ:1 concentrate:1 stochastic:17 vc:19 successor:1 require:1 ao:1 generalization:1 really:1 preliminary:1 randomization:1 opt:3 hold:2 considered:2 mapping:3 mop:1 parr:1 generous:1 a2:6 estimation:1 proc:1 label:1 currently:2 largest:1 create:1 establishes:1 trusted:1 mit:1 clearly:1 always:1 rather:5 barto:1 corollary:2 derived:1 sense:2 helpful:1 inference:1 shattered:2 typically:1 initially:1 koller:2 keams:1 transformed:6 labelings:4 selects:1 issue:2 special:1 uc:1 equal:2 once:1 never:2 having:1 ng:6 chernoff:1 look:1 connectionist:1 simultaneously:2 elevator:1 familiar:1 replaced:1 argmax:1 william:1 evaluation:1 truly:1 pc:1 behind:1 tuple:1 entertain:1 experience:5 tree:38 iv:1 desired:1 re:1 theoretical:1 weaken:1 instance:1 boolean:3 maximization:1 cost:1 kaelbling:1 addressing:1 uniform:7 examining:1 combined:2 international:1 straw:1 michael:1 tr2:5 quickly:1 concrete:1 again:3 choose:6 possibly:3 return:9 de:1 includes:1 baird:2 depends:3 ad:5 later:1 view:2 try:2 lab:1 root:4 reached:3 start:6 recover:1 hf:1 il:1 ensemble:1 v7r:14 yield:2 generalize:1 weak:1 carlo:3 trajectory:36 pomdps:10 researcher:1 straight:1 history:13 whenever:1 definition:5 proof:2 carefully:1 back:1 feed:1 supervised:9 day:1 box:1 though:2 hank:1 furthermore:1 just:1 implicit:1 until:1 incrementally:1 defines:4 perhaps:1 mdp:2 aviv:1 requiring:2 true:2 unbiased:4 hence:1 memoryless:1 moore:2 theoretic:3 tt:2 demonstrate:1 complete:1 performs:1 pvc:1 ols:1 common:1 absorbing:1 exponentially:3 extend:1 he:3 ai:1 replicating:1 had:1 access:2 recent:2 verlag:1 binary:3 arbitrarily:1 exploited:1 seen:1 floor:2 preceding:1 converge:1 maximize:1 ii:52 full:4 long:4 basic:1 controller:2 essentially:1 adopting:1 whereas:1 iiii:2 appropriately:1 ot:3 operate:1 unlike:1 ascent:1 induced:1 seem:1 jordan:1 call:5 structural:1 near:4 leverage:3 enough:2 easy:1 variety:1 restrict:1 reduce:1 idea:1 tm:1 motivated:1 handled:1 reuse:3 ird:1 action:16 boutilier:1 clear:1 detailed:1 informally:1 amount:5 discount:1 ang:1 generate:5 track:2 dummy:1 per:2 write:2 shall:1 waiting:2 reusable:8 key:2 drawn:2 modelfor:1 thresholded:1 button:1 asymptotically:1 sum:1 i7r:2 parameterized:1 powerful:1 extends:1 family:1 eased:3 reader:1 draw:1 decision:7 pushed:1 bound:13 precisely:1 calling:1 generates:1 simulate:2 argument:3 extremely:2 relatively:1 structured:1 developing:2 according:4 designated:1 poor:1 smaller:1 rider:3 em:2 remain:1 intuitively:1 restricted:4 pr:1 taken:2 computationally:1 assures:1 discus:2 turn:3 needed:6 tractable:1 adopted:1 generalizes:3 permit:1 observe:1 jd:2 original:2 unreasonably:1 include:1 calculating:1 giving:1 build:3 establish:2 classical:4 move:2 already:1 question:2 added:1 realized:1 strategy:58 parametric:1 dependence:7 gradient:9 link:2 reinforce:2 assuming:1 length:1 trl:5 difficult:1 reliably:1 proper:1 policy:6 perform:2 oter:1 upper:1 disagree:1 observation:9 markov:4 finite:8 descent:1 immediate:1 extended:2 mansour:6 arbitrary:3 lfm:1 david:1 namely:1 pair:2 address:1 beyond:1 suggested:1 boyen:1 built:1 including:1 tau:1 memory:1 belief:1 power:1 event:1 natural:2 difficulty:2 scheme:1 mdps:1 created:3 naive:1 prior:1 fully:3 interesting:1 limitation:2 agent:1 sufficient:2 course:1 gl:1 infeasible:2 hex:1 enjoys:1 formal:1 taking:1 absolute:1 benefit:1 distributed:2 feedback:1 dimension:13 depth:2 transition:3 world:2 evaluating:1 made:2 reinforcement:5 vmax:1 collection:1 far:2 approximate:7 observable:21 compact:5 keep:1 uai:1 search:9 why:1 table:1 additionally:1 lazily:1 tel:1 imagining:1 complex:3 domain:6 linearly:2 arise:1 arrival:1 child:9 convey:1 referred:1 wish:1 exponential:5 xl:1 lie:2 down:1 theorem:7 emphasized:1 showing:1 pac:1 vapnik:1 horizon:7 cassandra:1 easier:1 smoothly:1 simply:2 infinitely:2 lazy:1 partially:8 applies:1 springer:1 chance:1 andrewy:1 goal:1 viewed:1 exposition:1 towards:1 man:1 infinite:7 except:1 reducing:1 uniformly:3 averaging:1 acting:1 kearns:4 called:3 total:2 formally:1 internal:1 latter:2 preparation:1 evaluate:4
727
1,665
An Environment Model for N onstationary Reinforcement Learning Samuel P. M. Choi pmchoi~cs.ust.hk Dit-Yan Yeung Nevin L. Zhang dyyeung~cs.ust.hk lzhang~cs.ust.hk Department of Computer Science, Hong Kong University of Science and Technology Clear Water Bay, Kowloon, Hong Kong Abstract Reinforcement learning in nonstationary environments is generally regarded as an important and yet difficult problem. This paper partially addresses the problem by formalizing a subclass of nonstationary environments. The environment model, called hidden-mode Markov decision process (HM-MDP), assumes that environmental changes are always confined to a small number of hidden modes. A mode basically indexes a Markov decision process (MDP) and evolves with time according to a Markov chain. While HM-MDP is a special case of partially observable Markov decision processes (POMDP), modeling an HM-MDP environment via the more general POMDP model unnecessarily increases the problem complexity. A variant of the Baum-Welch algorithm is developed for model learning requiring less data and time. 1 Introduction Reinforcement Learning (RL) [7] is a learning paradigm based upon the framework of Markov decision process (MDP). Traditional RL research assumes that environment dynamics (i.e., MDP parameters) are always fixed (Le., stationary). This assumption, however, is not realistic in many real-world applications. In elevator control [3], for instance, the passenger arrival and departure rates can vary significantly over one day, and should not be modeled by a fixed MDP. Nonetheless, RL in nonstationary environments is regarded as a difficult problem. In fact, it is an impossible task if there is no regularity in the ways environment dynamics change. Hence, some degree of regularity must be assumed. Typically, nonstationary environments are presummed to change slowly enough such that online RL algorithms can be employed to keep track the changes. The online approach is memoryless in the sense that even if the environment ever revert to the previously learned dynamics, learning must still need to be started all over again. S. P. M Choi, D.-y' Yeung and N. L. Zhang 988 1.1 Our Proposed Model This paper proposes a formal model [1] for the nonstationary environments that repeats their dynamics in certain ways. Our model is inspired by the observations from the real-world nonstationary tasks with the following properties: Property 1. Environmental changes are confined to a small number of modes, which are stationary environments with distinct dynamics. The environment is in exactly one of these modes at any given time. This concept of modes seems to be applicable to many real-world tasks. In an elevator control problem, for example, the system might operate in a morning-rush-hour mode, an evening-rush-hour mode and a non-rush-hour mode. One can also imagine similar modes for other control tasks, such as traffic control and dynamic channel allocation [6]. Property 2. Unlike states, modes cannot be directly observed; the current mode can only be estimated according to the past state transitions. It is analogous to the elevator control example in that the passenger arrival rate and pattern can only be inferred through the occurrence of pick-up and drop-off requests. Property 3. Mode transitions are stochastic events and are independent of the control system's responses. In the elevator control problem, for instance, the events that change the current mode of the environment could be an emergency meeting in the administrative office, or a tea break for the staff on the 10th floor. Obviously, the elevator's response has no control over the occurrence of these events. Property 4. Mode transitions are relatively infrequent. In other words, a mode is more likely to retain for some time before switching to another one. If we consider the emergency meeting example, employees on different floors take time to arrive at the administrative office, and thus would generate a similar traffic pattern (drop-off requests on the same floor) for some period of time. Property 5. The number of states is often substantially larger than the number of modes. This is a common property for many real-world applications. In the elevator example, the state space comprises all possible combinations of elevator positions, pick-up and drop-off requests, and certainly would be huge. On the other hand, the mode space could be small. For instance, an elevator control system can simply have the three modes as described above to approximate the reality. Based on these properties, an environment model is proposed by introducing a mode variable to capture environmental changes. Each mode specifies an MDP and hence completely determines the current state transition function and reward function (property 1). A mode, however, is not directly observable (property 2), and evolves with time according to a Markov process (property 3). The model is therefore called hidden-mode model. Note that our model does not impose any constraint to satisfy properties 4 and 5. In other words, the hidden-mode model can work for environments without these two properties. Nevertheless, as will be shown later, these properties can improve learning in practice. 1.2 Related Work Our hidden-mode model is related to a non stationary model proposed by Dayan and Sejnowski [4]. Although our model is more restrictive in terms of representational power, it involves much fewer parameters and is thus easier to learn. Besides, other than the number of possible modes, we do not assume any other knowledge about 989 An Environment Model for Nonstationary Reinforcement Learning the way environment dynamics change. Dayan and Sejnowski, on the other hand, assume that one knows precisely how the environment dynamics change. The hidden-mode model can also be viewed as a special case of the hidden-state model, or partially observable Markov decision process (POMDP). As will be shown later, a hidden-mode model can always be represented by a hidden-state model through state augmentation. Nevertheless, modeling a hidden-mode environment via a hidden-state model will unnecessarily increase the problem complexity. In this paper, the conversion from the former to the latter is also briefly discussed. 1.3 Our Focus There are two approaches for RL. Model-based RL first acquires an environment model and then, from which, an optimal policy is derived. Model-free RL, on the contrary, learns an optimal policy directly through its interaction with the environment. This paper is concerned with the first part of the model-based approach, i.e., how a hidden-mode model can be learned from experience. We will address the policy learning problem in a separate paper. 2 Hidden-Mode Markov Decision Processes This section presents our hidden-mode model. Basically, a hidden-mode model is defined as a finite set of MDPs that share the same state space and action space, with possibly different transition functions and reward functions. The MDPs correspond to different modes in which a system operates. States are completely observable and their transitions are governed by an MDP. In contrast, modes are not directly observable and their transitions are controlled by a Markov chain. We refer to such a process as a hidden-mode Markov decision process (HM-MDP). An example of HM-MDP is shown in Figure l(a). Time ? Mode Action ... ... StaIC (a) A 3-mode, 4-state, I-action HM-MDP (b) The evolution of an HM-MDP. The arcs indicate dependencies between the variables Figure 1: An HM-MDP Formally, an HM-MDP is an 8-tuple (Q,S,A,X,Y,R,rr,'l'), where Q, S and A represent the sets of modes, states and actions respectively; the mode transition function X maps mode m to n with a fixed probability Xmn; the state transition function Y defines transition probability, Ym(8, a, s'), from state 8 to 8' given mode m and action a; the stochastic reward function R returns rewards with mean value rm (8, a); II and 'l1 denote the prior probabilities of the modes and of the states respectively. The evolution of modes and states over time is depicted in Figure 1 (b). S. P. M. Choi, D.-y' Yeung and N. L. Zhang 990 HM-MDP is a subclass of POMDP. In other words, the former can be reformulated as a special case of the latter. Specifically, one may take an ordered pair of any mode and observable state in the HM-MDP as a hidden state in the POMDP, and any observable state of the former as an observation of the latter. Suppose the observable states sand s' are in modes m and n respectively. These two HMMDP states together with their corresponding modes form two hidden states (m, s) and (n, s') for its POMDP counterpart. The transition probability from (m, s) to (n, s') is then simply the mode transition probability Xmn multiplied by the state transition probability Ym(s, a, s'). For an M-mode, N-state, K-action HM-MDP, the equivalent POMDP thus has N observations and M N hidden states. Since most state transition probabilities are collapsed into mode transition probabilities through parameter sharing, the number of parameters in an HM-MDP (N 2 M K + M2) is much less than that of its corresponding POMDP (M2 N 2K). 3 Learning a Hidden-Mode Model There are now two ways to learn a hidden-mode model. One may learn either an HM-MDP, or an equivalent POMDP instead. POMDP models can be learned via a variant of the Baum-Welch algorithm [2]. This POMDP Baum-Welch algorithm requires 8(M2 N 2T) time and 8(M2 N 2K) storage for learning an M-mode, Nstate, K-action HM-MDP, given T data items. A similar idea can be applied to the learning of an HM-MDP. Intuitively, one can estimate the model parameters based on the expected counts of the mode transitions, computed by a set of auxiliary variables. The major difference from the original algorithm is that consecutive state transitions, rather than the observations, are considered. Additional effort is thus needed for handling the boundary cases. This HM-MDP Baum-Welch algorithm is described in Figure 2. 4 Empirical Studies This section empirically examines the POMDP Baum-Welch 1 and HM-MDP BaumWelch algorithms. Experiments based on various randomly generated models and some real-world environments were conducted. The results are quite consistent. For illustration, a simple traffic control problem is presented. In this problem, one direction of a two-way traffic is blocked, and cars from two different directions (left and right) are forced to share the remaining road. To coordinate the traffic, two traffic lights equipped with sensors are set. The system then has two possible actions: either to signal cars from the left or cars from the right to pass. For simpliCity, we assume discrete time steps and uniform speed of the cars. The system has 8 possible states; they correspond to the combinations of whether there are cars waiting on the left and the right directions, and the stop signal position in the previous time step. There are 3 traffic modes. The first one has cars waiting on the left and the right directions with probabilities 0.3 and 0.1 respectively. In the second mode, these probabilities are reversed. For the last one, both probabilities are 0.3. In addition, the mode transition probability is 0.1. A cost of -1.0 results if lChrisman's algorithm also attempts to learn a minimal possible number of states. Our paper concerns only with learning the model parameters. 991 An Environment Model for Nonstationary Reinforcement Learning Given a collection of data and an initial model parameter vector repeat 0. 0=0 Compute forward variables (Xt. (Xl (i) = 1/;$1 (X2(i) = 1I"i 1/;$1 Yi(SI, al,S2) (Xt+l(j) = L:iEQ (Xt(i) Xii Yi(St,at,St+l) "Ii E Q "Ii E Q "Ii E Q Compute backward variables (3t . (3T(i) = 1 (3t(i) = LiEQXii Yi(St,at,St+I) (3t+I(j) (31(i) = L:iEQ 1I"j Yi(sl , al,s2) (32(j) "Ii E Q "Ii E Q "Ii E Q "I i , j E Q "Ii E Q Compute the new model parameter 0. _ .. _ L;-2 {. (i,i) Xl] - ~T . 8(a, b) = L....t=1 "'Yt (I) {01 a= b af.b 1Ti = "Yl (i) until maxi IOi - OJ I < to Figure 2: HM-MDP Baum-Welch Algorithm a car waits on either side. The experiments were run with the same initial model for data sets of various sizes. The algorithms iterated until the maximum change of the model parameters was less than a threshold of 0.0001. The experiment was repeated for 20 times with different random seeds in order to compute the median. Then the learned models were compared in their POMDP forms using the Kullback-Leibler (KL) distance [5], and the total CPU running time on a SUN Ultra I workstation was measured. Figure 3 (a) and (b) report the results. Generally speaking, both algorithms learn a more accurate environment model as the data size increases (Figure 3 (a)). This result is expected as both algorithms are statistically-based, and hence their performance relies largely on the data size. When the training data size is very small , both algorithms perform poorly. However, as the data size increases, HM-MDP Baum-Welch improves substantially faster than POMDP Baum-Welch. It is because an HM-MDP in general consists of fewer free s. P M 992 Choi. D.-y' Yeung and N. L. Zhang '0000 " /---.-.- ..-....-----.-.. ~...--.... --... ............ ...... .. -.----.".~ o ~~~~~~~~~~~~~~ !SOO 1000 1501) 2000 2&00 )000 :1500 oKIOO .&500 5000 o Wndow9tD (a) Error in transition function 10500L--'OOOJ...--'..."500-2000~~ .... ,,---:"_':::"--=_.,.,.......-:-""':':-:--,.... ...,,...,---:-!,OOO WIndowSiz. (b) Required learning time Figure 3: Empirical results on model learning parameters than its POMDP counterpart. HM-MDP Baum-Welch also runs much faster than POMDP Baum-Welch (Figure 3 (b)). It holds in general for the same reason discussed above. Note that computational time is not necessarily monotonically increasing with the data size. It is because the total computation depends not only on the data size, but also on the number of iterations executed. From our experiments, we noticed that the number of iterations tends to decrease as the data size increases. Larger models have also been tested. While HM-MDP Baum-Welch is able to learn models with several hundred states and a few modes, POMDP Baum-Welch was unable to complete the learning in a reasonable time. Additional experimental results can be found in [1]. 5 Discussions and Future Work The usefulness of a model depends on the validity of the assumptions made. We now discuss the assumptions of HM-MDP, and shed some light on its applicability to real-world nonstationary tasks. Some possible extensions are also discussed. Modeling a nonstationary environment as a number of distinct MDPs. MDP is a flexible framework that has been widely adopted in various applications. Modeling nonstationary environments by distinct MDPs is a natural extension to those tasks. Comparing to POMDP, our model is more comprehensive: each MDP naturally describes a mode of the environment. Moreover, this formulation facilitates the incorporation of prior knowledge into the model initialization step. States are directly observable while modes are not. While completely observable states are helpful to infer the current mode, it is also possible to extend the model to allow partially observable states. In this case, the extended model would be equivalent in representational power to a POMDP. This could be proved easily by showing the reformulation of the two models in both directions. An Environment Model for Nonstationary Reinforcement Learning 993 Mode changes are independent of the agent's responses. This property may not always hold for all real-world tasks. In some applications, the agent's actions might affect the state as well as the environment mode. In that case, an MDP should be used to govern the mode transition process. Mode transitions are relatively infrequent. This is a property that generally holds in many applications. Our model, however, is not limited by this condition. We have tried to apply our model-learning algorithms to problems in which this property does not hold. We find that our model still outperforms POMDP, although the required data size is typically larger for both models. Number of states is substantially larger than the number of modes. This is the key property that significantly reduces the number of parameters in HM-MDP compared to that in POMDP. In practice, introduction of a few modes is sufficient for boosting the system performance. More modes might only help little. Thus a trade-off between performance and response time must be decided. There are additional issues that need to be addressed. First, an efficient algorithm for policy learning is required. Although in principle it can be achieved indirectly via any POMDP algorithm, a more efficient algorithm based on the model-based approach is possible. We will address this issue in a separate paper. Next, the number of modes is currently assumed to be known. We are now investigating how to remove this limitation. Finally, the exploration-exploitation issue is currently ignored. In our future work, we will address this important issue and apply our model to real-world nonstationary tasks. References [1] S. P. M. Choi, D. Y. Yeung, and N. L. Zhang. Hidden-mode Markov decision processes. In IJCAI 99 Workshop on Neural, Symbolic, and Reinforcement Methods for Sequence Learnin9, 1999. [2] L. Chrisman. Reinforcement learning with perceptual aliasing: The perceptual distinctions approach. In AAAI-92, 1992. [3] R. H. Crites and A. G. Barto. Improving elevator performance using reinforcement learning. In D. Touretzky, M. Mozer, and M. Hasselmo, editors, Advances in Neural Information Processing Systems 8, 1996. [4] P. Dayan and T. J. Sejnowski. Exploration bonuses and dual control. Machine Learning, 25(1):5- 22, Oct. 1996. [5J S. Kullback. Information Theory and Statistics. Wiley, New York, NY, USA, 1959. [6] S. Singh and D. P. Bertsekas. Reinforcement learning for dynamic channel allocation in cellular telephone systems. In Advances in Neural Information Processing Systems 9, 1997. [7] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. The MIT Press, 1998.
1665 |@word kong:2 exploitation:1 briefly:1 seems:1 tried:1 pick:2 initial:2 past:1 outperforms:1 current:4 comparing:1 si:1 yet:1 ust:3 must:3 realistic:1 remove:1 drop:3 stationary:3 fewer:2 item:1 boosting:1 zhang:5 consists:1 expected:2 aliasing:1 inspired:1 cpu:1 little:1 equipped:1 increasing:1 moreover:1 formalizing:1 bonus:1 substantially:3 developed:1 subclass:2 ti:1 shed:1 exactly:1 rm:1 control:11 bertsekas:1 before:1 tends:1 switching:1 sutton:1 might:3 initialization:1 limited:1 statistically:1 decided:1 practice:2 empirical:2 yan:1 significantly:2 word:3 road:1 wait:1 symbolic:1 cannot:1 storage:1 collapsed:1 impossible:1 equivalent:3 map:1 yt:1 baum:12 pomdp:22 welch:12 simplicity:1 m2:4 examines:1 regarded:2 coordinate:1 analogous:1 imagine:1 suppose:1 infrequent:2 observed:1 capture:1 nevin:1 sun:1 decrease:1 trade:1 mozer:1 environment:30 govern:1 complexity:2 reward:4 dynamic:9 singh:1 upon:1 completely:3 easily:1 represented:1 various:3 revert:1 distinct:3 forced:1 sejnowski:3 quite:1 larger:4 widely:1 statistic:1 online:2 obviously:1 sequence:1 rr:1 interaction:1 ioi:1 poorly:1 representational:2 ijcai:1 regularity:2 help:1 measured:1 auxiliary:1 c:3 involves:1 indicate:1 direction:5 stochastic:2 exploration:2 sand:1 ultra:1 extension:2 hold:4 considered:1 seed:1 major:1 vary:1 consecutive:1 applicable:1 currently:2 hasselmo:1 mit:1 kowloon:1 always:4 sensor:1 rather:1 barto:2 office:2 derived:1 focus:1 hk:3 contrast:1 sense:1 helpful:1 dayan:3 typically:2 hidden:22 issue:4 dual:1 flexible:1 proposes:1 special:3 unnecessarily:2 future:2 report:1 few:2 randomly:1 comprehensive:1 elevator:9 attempt:1 huge:1 certainly:1 light:2 chain:2 accurate:1 tuple:1 experience:1 rush:3 minimal:1 instance:3 modeling:4 cost:1 introducing:1 applicability:1 uniform:1 hundred:1 usefulness:1 conducted:1 dependency:1 st:4 retain:1 off:4 yl:1 ym:2 together:1 again:1 augmentation:1 aaai:1 slowly:1 possibly:1 return:1 satisfy:1 depends:2 passenger:2 later:2 break:1 traffic:7 largely:1 correspond:2 iterated:1 basically:2 touretzky:1 sharing:1 nonetheless:1 naturally:1 workstation:1 stop:1 proved:1 knowledge:2 car:7 improves:1 day:1 response:4 ooo:1 formulation:1 until:2 hand:2 morning:1 defines:1 mode:70 mdp:34 usa:1 validity:1 requiring:1 concept:1 counterpart:2 former:3 hence:3 evolution:2 memoryless:1 leibler:1 acquires:1 samuel:1 hong:2 complete:1 l1:1 common:1 rl:7 empirically:1 discussed:3 extend:1 employee:1 refer:1 blocked:1 dyyeung:1 certain:1 meeting:2 yi:4 additional:3 staff:1 floor:3 employed:1 impose:1 paradigm:1 period:1 monotonically:1 signal:2 ii:8 infer:1 reduces:1 faster:2 af:1 controlled:1 variant:2 yeung:5 iteration:2 represent:1 confined:2 achieved:1 addition:1 addressed:1 median:1 operate:1 unlike:1 facilitates:1 contrary:1 nonstationary:13 enough:1 concerned:1 affect:1 idea:1 whether:1 effort:1 reformulated:1 speaking:1 york:1 action:9 ignored:1 generally:3 clear:1 dit:1 generate:1 specifies:1 sl:1 estimated:1 track:1 xii:1 discrete:1 tea:1 waiting:2 key:1 reformulation:1 nevertheless:2 threshold:1 backward:1 baumwelch:1 run:2 arrive:1 reasonable:1 decision:8 emergency:2 constraint:1 precisely:1 incorporation:1 x2:1 speed:1 relatively:2 department:1 according:3 request:3 combination:2 describes:1 evolves:2 intuitively:1 previously:1 discus:1 count:1 needed:1 know:1 adopted:1 multiplied:1 apply:2 indirectly:1 occurrence:2 original:1 assumes:2 remaining:1 running:1 restrictive:1 noticed:1 traditional:1 reversed:1 separate:2 distance:1 unable:1 cellular:1 water:1 reason:1 besides:1 index:1 modeled:1 illustration:1 difficult:2 executed:1 policy:4 perform:1 conversion:1 observation:4 markov:11 arc:1 finite:1 extended:1 ever:1 inferred:1 pair:1 required:3 kl:1 learned:4 distinction:1 chrisman:1 hour:3 address:4 xmn:2 able:1 pattern:2 departure:1 oj:1 soo:1 oooj:1 power:2 event:3 natural:1 improve:1 technology:1 mdps:4 started:1 hm:25 prior:2 limitation:1 allocation:2 degree:1 agent:2 sufficient:1 consistent:1 principle:1 editor:1 share:2 repeat:2 last:1 free:2 formal:1 side:1 allow:1 boundary:1 world:8 transition:21 forward:1 collection:1 reinforcement:11 made:1 approximate:1 observable:11 kullback:2 keep:1 investigating:1 assumed:2 evening:1 bay:1 reality:1 channel:2 learn:6 improving:1 necessarily:1 crites:1 s2:2 arrival:2 repeated:1 ny:1 wiley:1 position:2 comprises:1 xl:2 governed:1 perceptual:2 administrative:2 learns:1 choi:5 xt:3 showing:1 maxi:1 concern:1 workshop:1 easier:1 depicted:1 simply:2 likely:1 ordered:1 partially:4 environmental:3 determines:1 relies:1 oct:1 viewed:1 change:11 specifically:1 telephone:1 operates:1 called:2 total:2 pas:1 experimental:1 formally:1 latter:3 tested:1 handling:1
728
1,666
Rules and Similarity in Concept Learning Joshua B. Tenenbaum Department of Psychology Stanford University, Stanford, CA 94305 jbt@psych.stanford.edu Abstract This paper argues that two apparently distinct modes of generalizing concepts - abstracting rules and computing similarity to exemplars - should both be seen as special cases of a more general Bayesian learning framework. Bayes explains the specific workings of these two modes - which rules are abstracted, how similarity is measured - as well as why generalization should appear rule- or similarity-based in different situations. This analysis also suggests why the rules/similarity distinction, even if not computationally fundamental, may still be useful at the algorithmic level as part of a principled approximation to fully Bayesian learning. 1 Introduction In domains ranging from reasoning to language acquisition, a broad view is emerging of cognition as a hybrid of two distinct modes of computation, one based on applying abstract rules and the other based on assessing similarity to stored exemplars [7]. Much support for this view comes from the study of concepts and categorization. In generalizing concepts, people's judgments often seem to reflect both rule-based and similarity-based computations [9], and different brain systems are thought to be involved in each case [8]. Recent psychological models of classification typically incorporate some combination of rule-based and similarity-based modules [1,4]. In contrast to this currently popular modularity position, I will argue here that rules and similarity are best seen as two ends of a continuum of possible concept representations. In [11,12], I introduced a general theoretical framework to account for how people can learn concepts from just a few positive examples based on the principles of Bayesian inference. Here I explore how this framework provides a unifying explanation for these two apparently distinct modes of generalization. The Bayesian framework not only includes both rules and similarity as special cases but also addresses several questions that conventional modular accounts do not. People employ particular algorithms for selecting rules and measuring similarity. Why these algorithms as opposed to any others? People's generalizations appear to shift from similarity-like patterns to rule-like patterns in systematic ways, e.g., as the number of examples observed increases. Why these shifts? This short paper focuses on a simple learning game involving number concepts, in which both rule-like and similarity-like generalizations clearly emerge in the judgments of human subjects. Imagine that I have written some short computer programs which take as input a natural number and return as output either "yes" or "no" according to whether that number 60 J. B. Tenenbaum satisfies some simple concept. Some possible concepts might be "x is odd", "x is between 30 and 45", "x is a power of3", or"x is less than 10". For simplicity, we assume that only numbers under 100 are under consideration. The learner is shown a few randomly chosen positive examples - numbers that the program says "yes" to - and must then identify the other numbers that the program would accept. This task, admittedly artificial, nonetheless draws on people's rich knowledge of number while remaining amenable to theoretical analysis. Its structure is meant to parallel more natural tasks, such as word learning, that often require meaningful generalizations from only a few positive examples of a concept. Section 2 presents representative experimental data for this task. Section 3 describes a Bayesian model and contrasts its predictions with those of models based purely on rules or similarity. Section 4 summarizes and discusses the model's applicability to other domains. 2 The number concept game Eight subjects participated in an experimental study of number concept learning, under essentially the same instructions as those given above [11]. On each trial, subj ects were shown one or more random positive examples of a concept and asked to rate the probability that each of 30 test numbers would belong to the same concept as the examples observed. X denotes the set of examples observed on a particular trial, and n the number of examples. Trials were designed to fall into one of three classes. Figure la presents data for two representative trials of each class. Bar heights represent the average judged probabilities that particular test numbers fall under the concept given one or more positive examples X, marked by "*"s. Bars are shown only for those test numbers rated by subjects; missing bars do not denote zero probability of generalization, merely missing data. = = On class I trials, subjects saw only one example of each concept: e.g., X {16} and X {60}. To minimize bias, these trials preceded all others on which multiple examples were given. Given only one example, people gave most test numbers fairly similar probabilities of acceptance. Numbers that were intuitively more similar to the example received slightly higher ratings: e.g., for X = {16}, 8 was more acceptable than 9 or 6, and 17 more than 87; for X = {60}, 50 was more acceptable than 51, and 63 more than 43 . The remaining trials each presented four examples and occured in pseudorandom order. On class II trials, the examples were consistent with a simple mathematical rule: X = {16 , 8, 2, 64} or X {60, 80 , 10, 30}. Note that the obvious rules, "powers of two" and "multiples often", are in no way logically implied by the data. "Multiples offive" is a possibility in the second case, and "even numbers" or "all numbers under 80" are possibilities in both, not to mention other logically possible but psychologically implausible candidates, such as "all powers of two, except 32 or4". Nonetheless, subjects overwhelmingly followed an all-or-none pattern of generalization, with all test numbers rated near 0 or 1 according to whether they satisified the single intuitively "correct" rule. These preferred rules can be loosely characterized as the most specific rules (i.e., with smallest extension) that include all the examples and that also meet some criterion of psychological simplicity. = On class III trials, the examples satisified no simple mathematical rule but did have similar magnitudes: X {16, 23 , 19, 20} and X {60, 52, 57, 55} . Generalization now followed a similarity gradient along the dimension of magnitude. Probability ratings fell below 0.5 for numbers more than a characteristic distance beyond the largest or smallest observed examples - roughly the typical distance between neighboring examples ("'" 2 or 3). Logically, there is no reason why participants could not have generalized according to = = e 61 Rules and Similarity in Concept Learning various complex rules that happened to pick out the given examples, or according to very different values of~, yet all subjects displayed more or less the same similarity gradients. To summarize these data, generalization from a single example followed a weak similarity gradient based on both mathematical and magnitude properties of numbers. When several more examples were observed, generalization evolved into either an all-or-none pattern determined by the most specific simple rule, or, when no simple rule applied, a more articulated magnitude-based similarity gradient falling off with characteristic distance roughly equal to the typical separation between neighboring examples. Similar patterns were observed on several trials not shown (including one with a different value of and on two other experiments in quite different domains (described briefly in Section 4). e e) 3 The Bayesian model In [12], I introduced a Bayesian framework for concept learning in the context oflearning axis-parallel rectangles in a multidimensional feature space. Here I show that the same framework can be adapted to the more complex situation oflearning number concepts and can explain all of the phenomena of rules and similarity documented above. Formally, we observe n positive examples X = {x(1), ... , x(n)} of concept C and want to compute p(y E CIX), the probability that some new object y belongs to C given the observations X. Inductive leverage is provided by a hypothesis space 11. of possible concepts and a probabilistic model relating hypotheses h to data X. The hypothesis space. Elements ofll. correspond to subsets of the universe of objects that are psychologically plausible candidates for the extensions of concepts. Here the universe consists of numbers between 1 and 100, and the hypotheses correspond to subsets such as the even numbers, the numbers between 1 and 10, etc. The hypotheses can be thought of in terms of either rules or similarity, i.e., as potential rules to be abstracted or as features entering into a similarity computation, but Bayes does not distinguish these interpretations. Because we can capture only a fraction of the hypotheses people might bring to this task, we would like an objective way to focus on the most relevant parts of people's hypothesis space. One such method is additive clustering (ADCLUS) [6,10], which extracts a setoffeatures that best accounts for subjects' similarity judgments on a given set of objects. These features simply correspond to subsets of objects and are thus naturally identified with hypotheses for concept learning. Applications of ADCLUS to similarity judgments for the numbers 0-9 reveal two kinds of subsets [6,10]: numbers sharing a common mathematical property, such as {2, 4, 8} and {3, 6, 9}, and consecutive numbers of similar magnitude, such as {I, 2, 3, 4} and {2, 3, 4, 5, 6}. Applying ADCLUS to the full set of numbers from 1 to 100 is impractical, but we can construct an analogous hypothesis space for this domain based on the two kinds of hypotheses found in the ADCLUS solution for 0-9. One group of hypotheses captures salient mathematical properties: odd, even, square, cube, and prime numbers, multiples and powers of small numbers (~ 12), and sets of numbers ending in the same digit. A second group of hypotheses, representing the dimension of numerical magnitude, includes all intervals of consecutive numbers with endpoints between 1 and 100. Priors and likelihoods. The probabilistic model consists of a prior p( h) over 11. and a likelihood p( X Ih) for each hypothesis h E H. Rather than assigning prior probabilities to each ofthe 5083 hypotheses individually, I adopted a hierarchical approach based on the intuitive division of 11. into mathematical properties and magnitude intervals. A fraction A of the total probability was allocated to the mathematical hypotheses as a group, leaving (1 - A) for J. B. Tenenbaum 62 the magnitude hypotheses. The ,\ probability was distributed uniformly across the mathematical hypotheses. The (1 - ,\) probability was distributed across the magnitude intervals as a function of interval size according to an Erlang distribution, p( h) ex (Ihl/ li 2 )e- 1hl /0', to capture the intuition that intervals of some intermediate size are more likely than those of very large or small size. ,\ and Ii are treated as free parameters of the model. The likelihood is determined by the assumption of randomly sampled positive examples. In the simplest case, each example in X is assumed to be independently sampled from a uniform density over the concept G. For n examples we then have: p(Xlh) l/lhl n if Vj, xU) E h (1) o otherwise, where Ih I denotes the size of the subset h. For example, if h denotes the even numbers, then Ihl = 50, because there are 50 even numbers between I and 100. Equation I embodies the size principle for scoring hypotheses: smaller hypotheses assign greater likelihood than do larger hypotheses to the same data, and they assign exponentially greater likelihood as the number of consistent examples increases. The size principle plays a key role in learning concepts from only positive examples [12], and, as we will see below, in determining the appearance of rule-like or similarity-like modes of generalization. Given these priors and likelihoods, the posterior p( hlX) follows directly from Bayes' rule. Finally, we compute the probability of generalization to a new object y by averaging the predictions of all hypotheses weighted by their posterior probabilities p( h IX): p(y E GIX) =L (2) p(y E Glh)p(hIX). hE1i Equation 2 follows from the conditional independence of X and the membership of y E G, given h. To evaluate Equation 2, note that p(y E Glh) is simply 1 ify E h, and 0 otherwise. = Model results. Figure Ib shows the predictions of this Bayesian model (with'\ 1/2, Ii = 10). The model captures the main features of the data, including convergence to the most specific rule on Class II trials and to appropriately shaped similarity gradients on Class III trials. We can understand the transitions between graded, similarity-like and all-or-none, rule-like regimes of generalization as arising from the interaction of the size principle (Equation 1) with hypothesis averaging (Equation 2). Because each hypothesis h contributes to the average in Equation 2 in proportion to its posterior probability p(hIX), the degree of uncertainty in p(hIX) determines whether generalization will be sharp or graded. When p( h IX) is very spread out, many distinct hypotheses contribute significantly, resulting in a broad gradient of generalization. When p(hIX) is concentrated on a single hypothesis h*, only h* contributes significantly and generalization appears all-or-none. The degree of uncertainty in p( h IX) is in tum a consequence ofthe size principle. Given a few examples consistent with one hypothesis that is significantly smaller than the next-best competitor - such as X = {16, 8, 2, 64}, where "powers of two" is significantly smaller than "even numbers" - then the smallest hypothesis becomes exponentially more likely than any other and generalization appears to follow this most specific rule. However, given only one example (such as X {16}), or given several examples consistent with many similarly sized hypothesessuch as X = {16, 23,19, 20}, where the top candidates are all very similar intervals: "numbers between 16 and 23", "numbers between 15 and 24", etc. - the size-based likelihood favors the smaller hypotheses only slightly, p(hIX) is spread out over many overlapping hypotheses and generalization appears to follow a gradient of similarity. That the Bayesian = Rules and Similarity in Concept Learning 63 model predicts the right shape for the magnitude-based similarity gradients on Class III trials is no accident. The characteristic distance ? of the Bayesian generalization gradient varies with the uncertainty in p( h IX), which (for interval hypotheses) can be shown to covary with the intuitively relevant factor of average separation between neighboring examples. Bayes vs. rules or similarity alone. It is instructive to consider two special cases of the Bayesian model that are equivalent to conventional similarity-based and rule-based algorithms from the concept learning literature. What I call the SIM algorithm was pioneered by [5] and also described in [2,3] as a Bayesian approach to learning concepts from both positive and negative evidence. SIM replaces the size-based likelihood with a binary likelihood that measures only whether a hypothesis is consistent with the examples: p( X Ih) :::: 1 ifVj, xli) E h, and 0 otherwise. Generalization under SIM is just a count of the features shared by y and all the examples in X, independent of the frequency of those features or the number of examples seen. As Figure Ic shows, SIM successfully models generalization from a single example (Class I) but fails to capture how generalization sharpens up after multiple examples, to either the most specific rule (Class II) or a magnitude-based similarity gradient with appropriate characteristic distance ? (Class III). What I call the MIN algorithm preserves the size principle but replaces the step of hypothesis averaging with maximization: p(y E GIX) :::: 1 ify E arg maXh p(Xlh), and 0 otherwise. MIN is perhaps the oldest algorithm for concept learning [3] and, as a maximum likelihood algorithm, is asymptotically equivalent to Bayes. Its success for finite amounts of data depends on how peaked p(hIX) is (Figure Id). MIN always selects the most specific consistent rule, which is reasonable when that hypothesis is much more probable than any other (Class II), but too conservative in other cases (Classes I and III). In quantitative terms, the predictions of Bayes correlate much more highly with the observed data (R 2 :::: 0.91) than do the predictions of either SIM (R 2 :::: 0.74) or MIN (R 2 :::: 0.47). In sum, only the full Bayesian framework can explain the full range of rule-like and similarity-like generalization patterns observed on this task. 4 Discussion Experiments in two other domains provide further support for Bayes as a unifying framework for concept learning. In the context of multidimensional continuous feature spaces, similarity gradients are the default mode of generalization [5]. Bayes successfully models how the shape of those gradients depends on the distribution and number of examples; SIM and MIN do not [12]. Bayes also successfully predicts how fast these similarity gradients converge to the most specific consistent rule. Convergence is quite slow in this domain (n "" 50) because the hypothesis space consists of densely overlapping subsets - axisparallel rectangles - much like the interval hypotheses in the Class III number tasks. Another experiment engaged a word-learning task, using photographs of real objects as stimuli and a cover story oflearning a new language [11]. On each trial, subjects saw either one example of a novel word (e.g., a toy animal labeled with "Here is a blicket."), or three examples at one of three different levels of specificity: subordinate (e.g., 3 dalmatians labeled with "Here are three blickets."), basic (e.g., 3 dogs), or superordinate (e.g., 3 animals). They then were asked to pick the other instances of that concept from a set of 24 test objects, containing matches to the example(s) at all levels (e.g., other dalmatians, dogs, animals) as well as many non-matching objects. Figure 2 shows data and predictions for all three models. Similarity-like generalization given one example rapidly converged to the most specific rule after only three examples were observed, just as in the number task (Classes I and II) but in contrast to the axis-parallel rectangle task or the Class III num- , 64 J. B. Tenenbaum ber tasks, where similarity-like responding was still the norm after three or four examples. For modeling purposes, a hypothesis space was constructed from a hierarchical clustering of subjects' similarity judgments (augmented by an a priori preference for basic-level concepts) [11] . The Bayesian model successfully predicts rapid convergence from a similarity gradient to the minimal rule, because the smallest hypothesis consistent with each example set is significantly smaller than the next-best competitor (e.g., "dogs" is significantly smaller than "dogs and cats", just as with "multiples often" vs. "multiples of five"). Bayes fits the full data extremely well (R 2 = 0.98); by comparison, SIM (R 2 = 0.83) successfully accounts for only the n = 1 trials and MIN (R 2 = 0.76), the n = 3 trials. In conclusion, a Bayesian framework is able to account for both rule- and similarity-like modes of generalization, as well as the dynamics of transitions between these modes, across several quite different domains of concept learning. The key features of the Bayesian model are hypothesis averaging and the size principle. The former allows either rule-like or similarity-like behavior depending on the uncertainty in the posterior probability. The latter determines this uncertainty as a function of the number and distribution of examples and the structure ofthe learner's hypothesis space. With sparsely overlapping hypotheses - i.e., the most specific hypothesis consistent with the examples is much smaller than its nearest competitors - convergence to a single rule occurs rapidly, after just a few examples. With densely overlapping hypotheses - i.e., many consistent hypotheses of comparable size - convergence to a single rule occurs much more slowly, and a gradient of similarity is the norm after just a few examples. Importantly, the Bayesian framework does not so much obviate the distinction between rules and similarity as explain why it might be useful in understanding the brain. As Figures 1 and 2 show, special cases of Bayes corresponding to the SIM and MIN algorithms consistently account for distinct and complementary regimes of generalization. SIM, without the size principle, works best given only one example or densely overlappipg hypotheses, when Equation I does not generate large differences in likelihood. MIN, without hypothesis averaging, works best given many examples or sparsely overlapping hypotheses, when the most specific hypothesis dominates the sum over 1i in Equation 2. In light of recent brain-imaging studies dissociating rule- and exemplarbased processing [8], the Bayesian theory may best be thought of as a computational-level account of concept learning, with multiple subprocesses - perhaps subserving SIM and MIN - implemented in distinct neural circuits. I hope to explore this possibility in future work. References [1] M. Erickson & J. Kruschke (1998). Rules and exemplars in category learning. JEP: General 127, 107-140. [2] D. Haussler, M. Kearns, & R. Schapire (1994). Bounds on the sample complexity of Bayesian learning using information theory and the VC-dimension. Machine Learning 14,83-113. [3] T. Mitchell (1997). Machine Learning. McGraw-Hill. [4] R. Nosofsky & T. Palmeri (1998). A rule-plus-exception model for classifying objects in continuous-dimension spaces. Psychonomic Bull. & Rev. 5,345-369. [5] R. Shepard (1987). Towards a universal law of generalization for psychological science. Science 237, 1317-1323. [6] R. Shepard & P. Arabie (1979). Additive clustering: Representation of similarities as combinations of discrete overlapping properties. Psych. Rev. 86, 87-123. [7] S. Sloman & L. Rips (1998). Similarity and Symbols in Human Thinking. MIT Press. [8] E. Smith, A. Patalano & 1. Jonides (1998). Alternative strategies of categorization. In [6]. [9] E. Smith & S. Sloman (1994). Similarity- vs. rule-based categorization. Mem. & Cog. 22,377. [10] J. Tenenbaum (1996). Learning the structure of similarity. NIPS 8. [11] J. Tenenbaum (1999). A Bayesian Framework/or Concept Learning. Ph. D. Thesis, MIT. [12] J. Tenenbaum (1999). Bayesian modeling of human concept learning. NIPS I I. Rules and Similarity in Concept Learning 65 (a) Average generalization judgments: o.g~II~"~IIII Class I o.g [111~1I1 I. Class II Class III l o.~o ..I Jill _II I X=60 I II X=168264 I ? 1I. ! X=60 80 1030 o1~ ?? III 0.5 11 I II I IIlmll,' III I I X = 1: ? !. 1._*_111 1 II ... X=1623 1920 ? ? X=60 52 57 55 of ... 0.5 1 II I. LI~~~~ ' _ _~~~_ _~~~~ 10 20 30 40 50 60 70 80 90100 11111*11111 ? I ?? **'* I 10 20 30 40 50 60 70 80 90100 (b) Bayesian model: X= 16 o.gtllllil~'1 II ? 1 ?. ?. ?. .I X = 16 8 2 64 o1t 0.5 I IIliI a _ ~ ?1 I? I X= 1623 1920 ? o.g X=60 II L II I I ?? " ??1.1111 I I 1 o.g ..I. 1r o.g II. 10 20 30 40 50 60 70 80 90 100 X=6080103O I. I II ...L1J.. I. I .. . II ... X = 60 52 57 55 uljl. ... .. 10 20 30 40 50 60 70 80 90 100 (c) Pure similarity model (SIM): 1 t 0'8llhk~~lll 1 X= 16 I 1 II . J o.gf!II~u!wUIII . I II X=168264 o?~f"I??11 J I X=1623 1920 * ? ?J I. X=60 o.g IlL 1t I.! I ..1I1.1~IJIl !d X=60 80 10 30 o.g[ "1L 1r o.g II. 10 20 30 40 50 60 70 80 90 100 III hlllll??? 111 II X = 60 52 57 55 II I 11111.1111 I. 10 20 30 40 50 60 70 80 90 100 (d) Pure rule model (MIN): X= 16 a1f.....L * ?.... ?. 0.5 .I * 1f .I..... a.... X=168264 X=1623 1920 X = 60 1 0.5 o ??? o.g 1.1. 1 ~ o.g .1. 0.5 l!Ett* 10 20 30 40 50 60 70 80 90100 ........1..... .. I. III 1I.1..l.1~1~ro 10 30 X=60525755 .1 ...Jl. "' .. *** 10 20 30 40 50 60 70 80 90 100 Figure 1: Data and model predictions for the number concept task. (a) Average generalization judgments: (b) Bayesian model: o?~~?~~?~~?~UL (c) Pure similarity model (SIM): Training examples: I 1 I. 3 subordmate r 1 I 3 basic r 1 I. 3 superordinate r 1 0.5~.5~.5~.5 o?~IIL?~IIL?~IIL?~UL (d) Pure rule model (MIN): o?~~?~~?~~?~UL Figure 2: Data and model predictions for the word learning task.
1666 |@word trial:16 briefly:1 sharpens:1 proportion:1 norm:2 instruction:1 pick:2 mention:1 selecting:1 glh:2 yet:1 assigning:1 written:1 must:1 numerical:1 additive:2 shape:2 blickets:1 designed:1 v:3 alone:1 oldest:1 smith:2 short:2 num:1 provides:1 contribute:1 preference:1 five:1 height:1 mathematical:8 along:1 constructed:1 ect:1 consists:3 rapid:1 behavior:1 roughly:2 brain:3 lll:1 becomes:1 provided:1 circuit:1 what:2 evolved:1 kind:2 psych:2 emerging:1 impractical:1 quantitative:1 multidimensional:2 ro:1 appear:2 positive:9 consequence:1 id:1 meet:1 might:3 plus:1 subprocesses:1 suggests:1 range:1 jep:1 digit:1 universal:1 thought:3 significantly:6 matching:1 word:4 ett:1 specificity:1 judged:1 context:2 applying:2 conventional:2 equivalent:2 missing:2 independently:1 kruschke:1 of3:1 simplicity:2 pure:4 rule:53 haussler:1 importantly:1 obviate:1 analogous:1 imagine:1 play:1 pioneered:1 rip:1 hypothesis:47 element:1 sparsely:2 predicts:3 labeled:2 observed:9 role:1 module:1 capture:5 principled:1 intuition:1 complexity:1 asked:2 dynamic:1 arabie:1 o1t:1 purely:1 dissociating:1 division:1 learner:2 various:1 cat:1 articulated:1 distinct:6 fast:1 offive:1 axisparallel:1 artificial:1 quite:3 modular:1 stanford:3 plausible:1 larger:1 say:1 otherwise:4 favor:1 interaction:1 neighboring:3 relevant:2 rapidly:2 intuitive:1 convergence:5 assessing:1 categorization:3 object:9 depending:1 measured:1 exemplar:3 nearest:1 odd:2 received:1 sim:12 implemented:1 come:1 correct:1 vc:1 human:3 subordinate:1 explains:1 require:1 assign:2 generalization:30 probable:1 extension:2 ic:1 iil:3 algorithmic:1 cognition:1 continuum:1 consecutive:2 smallest:4 purpose:1 currently:1 saw:2 individually:1 largest:1 successfully:5 weighted:1 hope:1 mit:2 clearly:1 always:1 rather:1 overwhelmingly:1 focus:2 consistently:1 likelihood:11 logically:3 contrast:3 inference:1 membership:1 typically:1 accept:1 selects:1 i1:2 arg:1 classification:1 ill:1 priori:1 animal:3 special:4 fairly:1 cube:1 equal:1 construct:1 shaped:1 broad:2 thinking:1 peaked:1 future:1 others:2 stimulus:1 blicket:1 few:6 employ:1 randomly:2 preserve:1 densely:3 acceptance:1 possibility:3 highly:1 light:1 amenable:1 erlang:1 loosely:1 theoretical:2 minimal:1 psychological:3 instance:1 modeling:2 cover:1 measuring:1 maximization:1 bull:1 applicability:1 oflearning:3 subset:6 uniform:1 too:1 stored:1 varies:1 density:1 fundamental:1 gix:2 systematic:1 off:1 probabilistic:2 nosofsky:1 thesis:1 reflect:1 opposed:1 containing:1 slowly:1 xlh:2 exemplarbased:1 return:1 li:2 toy:1 account:7 potential:1 includes:2 depends:2 view:2 apparently:2 bayes:11 participant:1 parallel:3 minimize:1 square:1 l1j:1 characteristic:4 judgment:7 identify:1 correspond:3 ofthe:3 yes:2 weak:1 bayesian:23 xli:1 none:4 converged:1 explain:3 implausible:1 sharing:1 competitor:3 nonetheless:2 acquisition:1 frequency:1 involved:1 obvious:1 naturally:1 sampled:2 popular:1 ihl:2 mitchell:1 knowledge:1 occured:1 subserving:1 appears:3 tum:1 higher:1 follow:2 just:6 working:1 hlx:1 overlapping:6 mode:8 reveal:1 perhaps:2 concept:38 inductive:1 former:1 entering:1 jbt:1 covary:1 game:2 criterion:1 generalized:1 hill:1 argues:1 bring:1 reasoning:1 ranging:1 consideration:1 novel:1 common:1 psychonomic:1 preceded:1 endpoint:1 exponentially:2 shepard:2 jl:1 belong:1 interpretation:1 relating:1 cix:1 similarly:1 language:2 similarity:51 maxh:1 etc:2 posterior:4 recent:2 belongs:1 prime:1 binary:1 success:1 joshua:1 scoring:1 seen:3 greater:2 accident:1 converge:1 ii:26 multiple:8 full:4 match:1 characterized:1 prediction:8 involving:1 satisified:2 basic:3 essentially:1 psychologically:2 represent:1 want:1 participated:1 iiii:1 interval:8 leaving:1 allocated:1 appropriately:1 fell:1 subject:9 seem:1 call:2 near:1 leverage:1 intermediate:1 iii:12 independence:1 fit:1 psychology:1 gave:1 identified:1 shift:2 whether:4 ul:3 useful:2 amount:1 tenenbaum:7 ph:1 concentrated:1 category:1 simplest:1 documented:1 generate:1 schapire:1 happened:1 arising:1 discrete:1 group:3 key:2 four:2 salient:1 falling:1 rectangle:3 imaging:1 asymptotically:1 merely:1 fraction:2 sum:2 uncertainty:5 reasonable:1 separation:2 draw:1 acceptable:2 summarizes:1 comparable:1 bound:1 followed:3 distinguish:1 replaces:2 adapted:1 subj:1 min:11 extremely:1 pseudorandom:1 department:1 according:5 combination:2 describes:1 slightly:2 across:3 smaller:7 rev:2 hl:1 intuitively:3 computationally:1 equation:8 discus:1 count:1 end:1 adopted:1 eight:1 observe:1 hierarchical:2 appropriate:1 alternative:1 denotes:3 remaining:2 include:1 clustering:3 top:1 responding:1 unifying:2 embodies:1 graded:2 implied:1 objective:1 question:1 occurs:2 strategy:1 erickson:1 gradient:15 sloman:2 distance:5 argue:1 reason:1 o1:1 palmeri:1 ify:2 negative:1 observation:1 finite:1 displayed:1 situation:2 sharp:1 rating:2 introduced:2 dog:4 dalmatian:2 adclus:4 distinction:2 nip:2 address:1 beyond:1 bar:3 superordinate:2 below:2 pattern:6 able:1 regime:2 summarize:1 program:3 including:2 explanation:1 power:5 natural:2 hybrid:1 treated:1 representing:1 jill:1 rated:2 axis:2 hix:6 extract:1 gf:1 prior:4 literature:1 understanding:1 determining:1 law:1 fully:1 abstracting:1 degree:2 consistent:10 principle:8 story:1 classifying:1 free:1 bias:1 understand:1 ber:1 fall:2 emerge:1 distributed:2 dimension:4 default:1 ending:1 transition:2 rich:1 correlate:1 preferred:1 mcgraw:1 abstracted:2 mem:1 assumed:1 continuous:2 modularity:1 why:6 learn:1 ca:1 contributes:2 complex:2 domain:7 vj:1 did:1 main:1 spread:2 universe:2 complementary:1 xu:1 augmented:1 representative:2 slow:1 fails:1 position:1 candidate:3 ib:1 ix:4 cog:1 specific:11 symbol:1 evidence:1 dominates:1 ih:3 magnitude:11 generalizing:2 photograph:1 simply:2 explore:2 likely:2 appearance:1 satisfies:1 determines:2 conditional:1 marked:1 sized:1 lhl:1 towards:1 shared:1 typical:2 except:1 determined:2 uniformly:1 averaging:5 kearns:1 admittedly:1 total:1 conservative:1 engaged:1 experimental:2 la:1 meaningful:1 exception:1 formally:1 support:2 people:8 latter:1 meant:1 phenomenon:1 a1f:1 incorporate:1 evaluate:1 instructive:1 ex:1
729
1,667
Reinforcement Learning Using Approximate Belief States Andres Rodriguez * Artificial Intelligence Center SRI International 333 Ravenswood Avenue, Menlo Park, CA 94025 rodriguez@ai.sri.com Ronald Parr, Daphne Koller Computer Science Department Stanford University Stanford, CA 94305 {parr,koller}@cs.stanford.edu Abstract The problem of developing good policies for partially observable Markov decision problems (POMDPs) remains one of the most challenging areas of research in stochastic planning. One line of research in this area involves the use of reinforcement learning with belief states, probability distributions over the underlying model states. This is a promising method for small problems, but its application is limited by the intractability of computing or representing a full belief state for large problems. Recent work shows that, in many settings, we can maintain an approximate belief state, which is fairly close to the true belief state. In particular, great success has been shown with approximate belief states that marginalize out correlations between state variables. In this paper, we investigate two methods of full belief state reinforcement learning and one novel method for reinforcement learning using factored approximate belief states. We compare the performance of these algorithms on several well-known problem from the literature. Our results demonstrate the importance of approximate belief state representations for large problems. 1 Introduction The Markov Decision Processes (MDP) framework [2] is a good way of mathematically formalizing a large class of sequential decision problems involving an agent that is interacting with an environment. Generally, an MDP is defined in such a way that the agent has complete knowledge of the underlying state of the environment. While this formulation poses very challenging research problems, it is still a very optimistic modeling assumption that is rarely realized in the real world. Most of the time, an agent must face uncertainty or incompleteness in the information available to it. An extension of this formalism that generalizes MDPs to deal with this uncertainty is given by partially observable Markov Decision Processes (POMDPs) [1, 11] which are the focus of this paper. Solving a POMDP means finding an optimal behavior policy 7l'*, that maps from the agent's available knowledge of the environment, its belief state, to actions. This is usually done through a function, V, that assigns values to belief states. In the fully observable (MDP) "The work presented in this paper was done while the first author was at Stanford University. Reinforcement Learning Using Approximate Belief States 1037 case, a value function can be computed efficiently for reasonably sized domains. The situation is somewhat different for POMDPs, where finding the optimal policy is PSPACEhard in the number of underlying states [6]. To date, the best known exact algorithms to solve POMDPs are taxed by problems with a few dozen states [5]. There are several general approaches to approximating POMDP value functions using reinforcement learning methods and space does not permit a full review of them. The approach upon which we focus is the use of a belief state as a probability distribution over underlying model states. This is in contrast to methods that manipulate augmented state descriptions with finite memory [9, 12] and methods that work directly with observations [8] . The main advantage of a probability distribution is that it summarizes all of the information necessary to make optimal decisions [1]. The main disadvantages are that a model is required to compute a belief state, and that the task of representing and updating belief states in large problems is itself very difficult. In this paper, we do not address the problem of obtaining a model; our focus is on the the most effective way of using a model. Even with a known model, reinforcement learning techniques can be quite competitive with exact methods for solving POMDPs [lO]. Hence, we focus on extending the model-based reinforcement learning approach to larger problems through the use of approximate belief states. There are risks to such an approach: inaccuracies introduced by belief state approximation could give an agent a hopelessly inaccurate perception of its relationship to the environment. Recent work [4], however, presents an approximate tracking approach, and provides theoretical guarantees that the result of this process cannot stray too far from the exact belief state. In this approach, rather than maintaining an exact belief state, which is infeasible in most realistically large problems, we maintain an approximate belief state, usually from some restricted class of distributions. As the approximate belief state is updated (due to actions and observations), it is continuously projected back down into this restricted class. Specifically, we use decomposed belief states, where certain correlations between state variables are ignored. In this paper we present empirical results comparing three approaches to belief state reinforcement learning. The most direct approach is the use of a neural network with one input for each element of the full belief state. The second is the SPOVA method [lO], which uses a function approximator designed for POMDPs and the third is the use of a neural network with an approximate belief state as input. We present results for several well-known problems in the POMDP literature, demonstrating that while belief state approximation is ill-suited for some problems, it is an effective means of attacking large problems. 2 Basic Framework and Algorithms A POMDP is defined as a tuple < S, A, 0, T, R, 0 > of three sets and three functions. S is a set of states, A is a set of actions and is a set of observations. The transition function T : S x A ~ II( S) specifies how the actions affect the state of the world. It can be viewed as T( Si, a, S j) = P( S j la, sd, the probability that the agent reaches state S j if it currently is in state Si and takes action a. The reward function R : S x A ~ 1R determines the immediate reward received by the agent The observation model 0 : S x A ~ II( 0) determines what the agent perceives, depending on the environment state and the action taken. O(s, a, 0) = P( ola, s) is the probability that the agent observes 0 when it is in state s, having taken the action a. ? 1038 2.1 A. Rodriguez, R. Parr and D. Koller POMDP belief states A beliefstate, b, is defined as a probability distribution over all states S E S, where b(s), represents probability that the environment is in state s. After taking action a and observing 0, the belief state is updated using Bayes rule: 1 1 b (s ) O(S', a, 0) L.SES T(Si, a, s')b(sd = P( s I a, 0, b) = =---::::-:--:,-,:'=--~--:-:-:---:? L.sjES O(Sj, a, 0) L.siES T(Si' a, Sj)b(Si) 1 The size of an exact belief state is equal to the number of states in the model. For large problems, maintaining and manipulating an exact belief state can be problematic even if the the transition model has a compact representation [4]. For example, suppose the state space is described via a set of random variables X = {Xl, ... ,Xn }, where each Xi takes on values in some finite domain Val(Xi ), a particular S defines a value Xi E VaJ(Xi ) for each variable Xi. The full belief state representation will be exponential in n. We use the approximation method analyzed by Boyen and Koller [4], where the variables are partitioned into a set of disjoint clusters C I ... Ck and belief functions, bl ... bk are maintained over the variables in each cluster. At each time step, we compute the exact belief state, then compute the individual belief functions by marginalizing out inter-cluster correlations. For some assignment, Ci, to variables in C i , we obtain bi(Ci) = L.ygCl P(Ci' y). An approximation of the original, full belief state is then reconstructed as b( s) = n~=l bi (Ci). By representing the belief state as a product of marginal probabilities, we are projecting the belief state into a reduced space. While a full belief state representation for n state variables would be exponential in n, the size of decomposed belief state representation is exponential in the size of the largest cluster and additive in the number of clusters. For processes that mix rapidly enough, the errors introduced by approximation will stay bounded over time [4]. As discussed by Boyen and Koller [4], this type of decomposed belief state is particularly suitable for processes that can themselves be factored and represented as a dynamic Bayesian network [3]. In such cases we can avoid ever representing an exponentially sized belief state. However, the approach is fully general, and can be applied in any setting where the state is defined as an assignment of values to some set of state variables. 2.2 Value functions and policies for POMDPs If one thinks of a POMOP as an MOP defined over belief states, then the well-known fixed point equations for MOPs still hold. Specifically, V*(b) = m~x [L sES b(s)R(s, a) + 'Y L P(ola, b)V*(bl )] oED where'Y is the discount factor and b' (defined above) is the next belief state. The optimal policy is determined by the maximizing action for each belief state. In principle, we could use Q-Iearning or value iteration directly to solve POMOPs. The main difficulty lies in the fact that there are uncountably many belief states, making a tabular representation of the value function impossible. Exact methods for POMOPs use the fact that finite horizon value functions are piecewiselinear and convex [11], ensuring a finite representation. While finite, this representation can grow exponentially with the horizon, making exact approaches impractical in most settings. Function approximation is an attractive alternative to exact methods. We implement function approximation using a set of parameterized Q-functions, where Qa(b, W a ) is the reward-to-go for taking action a in belief state b. A value function is reconstructed from the Q-functions as V(b) maxa(Qa(b, W a )), and the update rule for Wa when a transition = Reinforcement Learning Using Approximate Belief States 1039 from state b to b' under action a with reward R is: 2.3 Function approximation architectures We consider two types of function approximators. The first is a two-layer feedforward neural network with sigmoidal internal units and a linear outermost layer. We used one network for each Q function. For full belief state reinforcement learning, we used networks hidden nodes. For with lSI inputs (one for each component of the belief state) and approximate belief state reinforcement learning, we used networks with one input for each assignment to the variables in each cluster. If we had two clusters, for example, each with 3 binary variables, then our Q networks would each have 2 3 + 2 3 = 16 inputs. We kept the number of hidden nodes for each network as the square root of the number of inputs. v'fSf Our second function approximator is SPOVA [10], which is a soft max function designed to exploit the piecewise-linear structure of POMDP value functions. A SPOVA Q function maintains a set of weight vectors Wal . . . W ai, and is evaluated as: In practice, a small value of k (usually 1.2) is adopted at the start of learning, making the function very smooth. This is increased during learning until SPOVA closely approximates a PWLC function of b (usually k = 8). We maintained one SPOVA Q function for each vectors to each function. This gave O(IAIISI parameters action and assigned to both SPOVA and the full belief state neural network. JiST 3 JiST) Empirical Results We present results on several problems from the POMDP literature and present an extension to a known machine repair problem that is designed to highlight the effects of approximate belief states. Our results are presented in the form of performance graphs, where the value of the current policy is obtained by taking a snapshot of the value function and measuring the discounted sum of reward obtained by the resulting policy in simulation. We use "NN" to refer to the neural network trained reinforcement learner trained with the full belief state and the term "Decomposed NN" to refer to the neural network trained with an approximate belief which is decomposed as a product of marginals. We used a simple exploration strategy, starting with a 0.1 probability of acting randomly, which decreased linearly to 0.01. Due to space limitations, we are not able to describe each model in detail. However, we used publicly available model description files from [5].1 Table 3.4 shows the running times of the different methods. These are generally much lower than what would be required to solve these problems using exact methods. 3.1 Grid Worlds We begin by considering two grid worlds, a 4 x 3 world from [10] and a 60-state world from [7]. The 4 x 3 world contains only 11 states and does not have a natural decomposition into state variables, so we compared SPOVA only with the full belief state neural network. I See hup:/Iwww.cs.brown.edu/research/ai/pomdp/index.html. Note that this file format specifies a starting distribution for each problem and our results are reported with respect to this starting distribution. A. Rodriguez, R. Parr and D. Koller 1040 S POVA NN ------Oecompo.-l NN ______/ " 01 ~ ! 05 o .0.5 / o. f? / J -. ?1 50'---"""OOOO""""""20000-'--"""".L---"-""""'"c:--:::-"--60000-'--= ' OOOO.L--",,, '"c:--90000 ,,-.--' ,OOOOO ISOOOO 1IOOOO Figure 1: a) 3 x 4 Grid World, b) 60-state maze The experimental results, which are averaged over 25 training runs and 100 simulations per policy snapshot, are presented in Figure 1a. They show that SPOVA learns faster than the neural network, but that the network does eventually catch up. The 60-state robot navigation problem [7] was amenable to a decomposed belief state approximation since its underlying state space comes from the product of 15 robot positions and 4 robot orientations. We decomposed the belief state with two clusters, one containing a position state variable and the other containing an orientation state variable. Figure 1b shows results in which SPOVA again dominates. The decomposed NN has trouble with this problem because the effects of position and orientation on the value function are not easily decoupled, i.e., the effect of orientation on value is highly state-dependent. This meant that the decomposed NN was forced to learn a much more complicated function of its inputs than the function learned by the network using the full belief state. 3.2 Aircraft Identification Aircraft identification is another problem studied in Cassandra's thesis. It includes sensing actions for identifying incoming aircraft and actions for attacking threatening aircraft. Attacks against friendly aircraft are penalized, as are failures to intercept hostile aircraft. This is a challenging problem because there is tension in deciding between the various sensors. " Better sensors tend to make the base more visible to hostile aircraft, while more stealthy sensors are less accurate. The sensors give information about both the aircraft's type and distance from the base. The state space of this problem is comprised of three main components. aircraft type - eitherthe aircraft is a friend orit is a foe; distance -how far the aircraft is currently from the base discretized into an adjustable number, d, of distinct distances; vis ibi 1 i ty - a measure of how visible the base is to the approaching aircraft, which is discretized into 5 levels. We chose d = 10, gaving this problem 104 states. The problem has a natural decomposition into state variables for aircraft type, distance and base visibility. The results for the three algorithms are shown in Figure 2(a). This is the first problem where we start to see an advantage from decomposing the belief state. For the decomposed NN, we used three separate clusters, one for each variable, which meant that the network had only 17 inputs. Not only did the simpler network learn faster, but it learned a better policy overall. We believe that this illustrates an important point: even though SPOVA and the full belief state neural network may be more expressive than the decomposed NN, the decomposed NN is able to search the space of functions it can represent much more efficiently due to the reduced number of parameters. 1041 Reinforcement Learning Using Approximate BeliefStates SPOVA, NN ----- DKompo!ll8d NN . so SPOVA NN O.:.ompo!ll8d NN ?20 ?20 O'--'-'~ OOOOO "-:--:-200000 '-'--"""""' OOO:---:-: """"" """""'SOOOOO ~-:eooooo -:":-:-:--:: 700000 """""""' 800000 ":::::-900000 """""'--' ''''' 1*IIIIonI o 10000 2IXlOO 30000 40000 50000 lteratbns 60000 70000 aoooo 90000 l00c00 Figure 2.: a) Aircraft Identification, b) Machine Maintenance 3.3 Machine Maintenance Our last problem was the machine maintenance problem from Cassandra's database. The problem assumes that there is a machine with a certain number of components. The quality of the parts produced by the machine is determined by the condition of the components. Each component can be in one of four conditions: good - the component is in good condition; fair - the component has some amount of wear, and would benefit from some maintenance; bad - the part is very worn and could use repairs; broken - the part is broken and must be replaced. The status of the components is observable only if the machine is completely disassembled. Figure 2(b) shows performance results for this problem for the 4 component version of this problem. At 256 states, it was at the maximum size for which a full belief state approach was manageable. However, the belief state for this problem decomposes naturally into clusters describing the status of each machine, creating a decomposed belief state with just four components. The graph shows the dominance of this this simple decomposition approach. We believe that this problem clearly demonstrates the advantage of belief state decomposition: The decomposed NN learns a function of 16 inputs in fraction of the time it takes for the full net or SPOVA to learn a lower-quality function of 256 inputs. 3.4 Running Times The table below shows the running times for the different problems presented above. These are generally much less than what would be required to solve these problems exactly. The full NN and SPOVA are roughly comparable, but the decomposed neural network is considerably faster. We did not exploit any problem structure in our approximate belief state computation, so the time spent computing belief states is actually larger for the decomposed NN. The savings comes from the the reduction in the number of parameters used, which reduced the number of partial derivatives computed. We expect the savings to be significantly more substantial for processes represented in a factored way [3], as the approximate belief state propagation algorithm can also take advantage of this additional structure. 4 Concluding Remarks We have a proposed a new approach to belief state reinforcement learning through the use of approximate belief states. Using well-known examples from the POMDP literature, we have compared approximate belief state reinforcement learning with two other methods A. Rodriguez, R. Parr and D. Koller 1042 Problem 3x4 Hallway Aircraft ID MachineM. SPOVA 19.1 s 32.8 min 38.3 min 2.5 h NN 13.0s 47 .1 min 49.9 min 2.6 h Decomposed NN 3.2 min 4.4 min 4.7 min Table 1: Run times (in seconds, minutes or hours) for the different algorithms that use exact belief states. Our results demonstrate that, while approximate belief states may not be ideal for tightly coupled problem features, such as the position and orientation of a robot, they are a natural and effective means of addressing some large problems. Even for the medium-sized problems we showed here, approximate belief state reinforcement learning can outperform full belief state reinforcement learning using fewer trials and much less CPU time. For many problems, exact belief state methods will simply be impractical and approximate belief states will provide a tractable alternative. Acknowledgements This work was supported by the ARO under the MURI program "Integrated Approach to Intelligent Systems," by ONR contract N66001-97-C-8554 under DARPA's HPKB program, and by the generosity of the Powell Foundation and the Sloan Foundation. References [1] K. J. Astrom. Optimal control of Markov decision processes with incomplete state estimation. l. Math. Anal. Applic., 10:174-205,1965. [2] R.E. Bellman. Dynamic Programming. Princeton University Press, 1957. [3] C. Boutilier, T. Dean, and S. Hanks . Decision theoretic planning: Structural assumptions and computational leverage. Journal of Artijiciallntelligence Research, 1999. [4] X. Boyen and D. Koller. Tractable inference for complex stochastic processes. In Proc. UAI, 1998. [5] A. Cassandra. Exact and approximate Algorithms for partially observable Markov Decision Problems. PhD thesis, Computer Science Dept., Brown Univ., 1998. [6] M. Littman. Algorithms for Sequential Decision Making. PhD thesis, Computer Science Dept., Brown Univ., 1996. [7] M. Littman, A. Cassandra, and L.P. Kaelbling. Learning policies for partially observable environments: Scaling up. In Proc. ICML, pages 362-370, 1996. [8] J. Loch and S. Singh. Using eligibility traces to find the best memory less policy in partially observable markov decision processes. In Proc. ICML. Morgan Kaufmann, 1998. [9] Andrew R. McCallum. Overcoming incomplete perception with utile distinction memory. In Proc.ICML, pages 190-196, 1993. [10] Ronald Parr and Stuart Russell. Approximating optimal policies for partially observable stochastic domains. In Proc. IlCAI, 1995. [11] R. D. Smallwood and E. J. Sondik. The optimal control of partially observable Markov processes over a finite horizon. Operations Research, 21: 1071-1088,1973. [12] M. Wiering and J. Schmidhuber. HQ-leaming: Discovering Markovian subgoals for non-Markovian reinforcement learning. Technical report, Istituo Daile Molle di Studi sull'Intelligenza Artificiale, 1996.
1667 |@word aircraft:15 trial:1 version:1 sri:2 manageable:1 simulation:2 decomposition:4 reduction:1 contains:1 current:1 com:1 comparing:1 si:5 artijiciallntelligence:1 must:2 ronald:2 visible:2 additive:1 visibility:1 designed:3 update:1 intelligence:1 fewer:1 discovering:1 hallway:1 mccallum:1 utile:1 provides:1 math:1 node:2 attack:1 sigmoidal:1 simpler:1 daphne:1 direct:1 inter:1 roughly:1 themselves:1 planning:2 behavior:1 discretized:2 bellman:1 discounted:1 decomposed:17 cpu:1 considering:1 perceives:1 begin:1 underlying:5 formalizing:1 bounded:1 medium:1 what:3 sull:1 maxa:1 finding:2 impractical:2 guarantee:1 friendly:1 iearning:1 exactly:1 demonstrates:1 control:2 unit:1 sd:2 pomop:1 id:1 ibi:1 chose:1 studied:1 challenging:3 limited:1 ola:2 bi:2 averaged:1 practice:1 implement:1 powell:1 area:2 empirical:2 significantly:1 cannot:1 close:1 marginalize:1 risk:1 impossible:1 intercept:1 map:1 dean:1 center:1 maximizing:1 go:1 worn:1 starting:3 convex:1 pomdp:9 identifying:1 assigns:1 factored:3 rule:2 smallwood:1 updated:2 suppose:1 exact:14 programming:1 us:1 element:1 particularly:1 updating:1 muri:1 database:1 wiering:1 russell:1 observes:1 substantial:1 environment:7 broken:2 reward:5 littman:2 dynamic:2 oed:1 trained:3 singh:1 solving:2 upon:1 learner:1 completely:1 easily:1 darpa:1 represented:2 various:1 univ:2 forced:1 distinct:1 effective:3 describe:1 artificial:1 quite:1 stanford:4 solve:4 larger:2 s:2 think:1 itself:1 advantage:4 net:1 aro:1 product:3 date:1 rapidly:1 realistically:1 description:2 cluster:10 extending:1 artificiale:1 spent:1 depending:1 friend:1 andrew:1 pose:1 received:1 c:2 involves:1 come:2 closely:1 stochastic:3 exploration:1 molle:1 mathematically:1 extension:2 hold:1 deciding:1 great:1 mop:2 parr:6 estimation:1 proc:5 currently:2 largest:1 ravenswood:1 sensor:4 clearly:1 rather:1 ck:1 avoid:1 focus:4 contrast:1 generosity:1 inference:1 dependent:1 nn:18 inaccurate:1 integrated:1 hidden:2 koller:8 manipulating:1 overall:1 ill:1 html:1 orientation:5 fairly:1 marginal:1 equal:1 saving:2 having:1 x4:1 represents:1 park:1 stuart:1 icml:3 tabular:1 report:1 piecewise:1 intelligent:1 few:1 randomly:1 tightly:1 individual:1 replaced:1 vaj:1 maintain:2 investigate:1 highly:1 threatening:1 analyzed:1 navigation:1 amenable:1 accurate:1 tuple:1 partial:1 necessary:1 decoupled:1 incomplete:2 theoretical:1 increased:1 formalism:1 modeling:1 soft:1 markovian:2 disadvantage:1 measuring:1 assignment:3 kaelbling:1 addressing:1 comprised:1 too:1 reported:1 fsf:1 considerably:1 international:1 stay:1 contract:1 continuously:1 again:1 thesis:3 containing:2 creating:1 derivative:1 includes:1 sloan:1 vi:1 root:1 sondik:1 optimistic:1 observing:1 competitive:1 bayes:1 maintains:1 start:2 complicated:1 square:1 publicly:1 kaufmann:1 efficiently:2 wal:1 bayesian:1 identification:3 andres:1 produced:1 pomdps:7 foe:1 reach:1 against:1 failure:1 ty:1 naturally:1 di:1 knowledge:2 oooo:2 actually:1 back:1 tension:1 taxed:1 ooo:1 formulation:1 done:2 evaluated:1 though:1 hank:1 just:1 correlation:3 until:1 expressive:1 disassembled:1 propagation:1 rodriguez:5 defines:1 quality:2 mdp:3 believe:2 effect:3 brown:3 true:1 hence:1 assigned:1 deal:1 attractive:1 piecewiselinear:1 during:1 eligibility:1 maintained:2 complete:1 demonstrate:2 theoretic:1 novel:1 exponentially:2 subgoals:1 discussed:1 approximates:1 marginals:1 refer:2 ai:3 grid:3 had:2 wear:1 robot:4 base:5 recent:2 showed:1 schmidhuber:1 certain:2 hostile:2 binary:1 success:1 onr:1 approximators:1 morgan:1 additional:1 somewhat:1 pwlc:1 attacking:2 ii:2 full:17 mix:1 smooth:1 technical:1 faster:3 manipulate:1 ensuring:1 involving:1 basic:1 maintenance:4 iteration:1 represent:1 decreased:1 grow:1 file:2 tend:1 structural:1 leverage:1 ideal:1 feedforward:1 enough:1 affect:1 gave:1 architecture:1 approaching:1 avenue:1 action:14 remark:1 ignored:1 generally:3 boutilier:1 ooooo:2 amount:1 discount:1 reduced:3 specifies:2 outperform:1 ilcai:1 lsi:1 problematic:1 disjoint:1 per:1 dominance:1 four:2 demonstrating:1 kept:1 n66001:1 graph:2 fraction:1 sum:1 run:2 parameterized:1 uncertainty:2 decision:10 summarizes:1 scaling:1 incompleteness:1 comparable:1 layer:2 min:7 concluding:1 format:1 department:1 developing:1 partitioned:1 making:4 projecting:1 restricted:2 repair:2 taken:2 equation:1 remains:1 describing:1 eventually:1 tractable:2 adopted:1 available:3 generalizes:1 decomposing:1 permit:1 operation:1 alternative:2 original:1 assumes:1 running:3 trouble:1 maintaining:2 exploit:2 approximating:2 bl:2 realized:1 strategy:1 hq:1 distance:4 separate:1 intelligenza:1 studi:1 loch:1 index:1 relationship:1 difficult:1 trace:1 anal:1 policy:12 adjustable:1 observation:4 snapshot:2 markov:7 finite:6 immediate:1 situation:1 ever:1 interacting:1 spova:15 overcoming:1 introduced:2 bk:1 required:3 learned:2 distinction:1 hour:1 inaccuracy:1 applic:1 qa:2 address:1 able:2 usually:4 perception:2 below:1 boyen:3 program:2 max:1 memory:3 belief:76 suitable:1 difficulty:1 natural:3 representing:4 mdps:1 catch:1 coupled:1 review:1 literature:4 acknowledgement:1 val:1 marginalizing:1 fully:2 expect:1 highlight:1 limitation:1 approximator:2 foundation:2 agent:9 principle:1 intractability:1 lo:2 uncountably:1 jist:2 penalized:1 supported:1 last:1 infeasible:1 face:1 taking:3 benefit:1 outermost:1 xn:1 world:8 transition:3 maze:1 author:1 reinforcement:19 projected:1 far:2 sj:2 approximate:24 observable:9 compact:1 reconstructed:2 status:2 incoming:1 uai:1 xi:5 search:1 decomposes:1 table:3 promising:1 learn:3 reasonably:1 ca:2 menlo:1 obtaining:1 pomops:2 complex:1 domain:3 did:2 main:4 linearly:1 fair:1 augmented:1 astrom:1 stray:1 position:4 exponential:3 xl:1 lie:1 third:1 learns:2 dozen:1 down:1 minute:1 bad:1 sensing:1 dominates:1 sequential:2 importance:1 ci:4 phd:2 illustrates:1 horizon:3 cassandra:4 suited:1 simply:1 hopelessly:1 tracking:1 partially:7 determines:2 sized:3 viewed:1 leaming:1 specifically:2 determined:2 acting:1 experimental:1 la:1 rarely:1 hup:1 internal:1 meant:2 dept:2 princeton:1
730
1,668
Learning sparse codes with a mixture-of-Gaussians prior Bruno A. Olshausen Department of Psychology and Center for Neuroscience, UC Davis 1544 Newton Ct. Davis, CA 95616 baolshausen@ucdavis.edu K. Jarrod Millman Center for Neuroscience, UC Davis 1544 Newton Ct. Davis, CA 95616 kjmillman@ucdavis. edu Abstract We describe a method for learning an overcomplete set of basis functions for the purpose of modeling sparse structure in images. The sparsity of the basis function coefficients is modeled with a mixture-of-Gaussians distribution. One Gaussian captures nonactive coefficients with a small-variance distribution centered at zero, while one or more other Gaussians capture active coefficients with a large-variance distribution. We show that when the prior is in such a form, there exist efficient methods for learning the basis functions as well as the parameters of the prior. The performance of the algorithm is demonstrated on a number of test cases and also on natural images. The basis functions learned on natural images are similar to those obtained with other methods, but the sparse form of the coefficient distribution is much better described. Also, since the parameters of the prior are adapted to the data, no assumption about sparse structure in the images need be made a priori, rather it is learned from the data. 1 Introduction The general problem we address here is that of learning a set of basis functions for representing natural images efficiently. Previous work using a variety of optimization schemes has established that the basis functions which best code natural images in terms of sparse, independent components resemble a Gabor-like wavelet basis in which the basis functions are spatially localized, oriented and bandpass in spatial-frequency [1, 2, 3, 4]. In order to tile the joint space of position, orientation, and spatial-frequency in a manner that yields useful image representations, it has also been advocated that the basis set be overcomplete [5], where the number of basis functions exceeds the dimensionality of the images being coded. A major challenge in learning overcomplete bases, though, comes from the fact that the posterior distribution over the coefficients must be sampled during learning. When the posterior is sharply peaked, as it is when a sparse prior is imposed, then conventional sampling methods become especially cumbersome. 842 B. A. Olshausen and K. J. Millman One approach to dealing with the problems associated with overcomplete codes and sparse priors is suggested by the form of the resulting posterior distribution over the coefficients averaged over many images. Shown below is the posterior distribution of one of the coefficients in a 4 x's overcomplete representation. The sparse prior that was imposed in learning was a Cauchy distribution and is overlaid (dashed line). It would seem that the coefficients do not fit this imposed prior very well, and instead want to occupy one of two states: an inactive state in which the coefficient is set nearly to zero, and an active state in which the coefficient takes on some significant non-zero value along a continuum. This suggests that the appropriate choice of prior is one that is capable of capturing these two discrete states. II I I ? I I . I I -2 \ 0 2 coo!IIdon1v.... Figure 1: Posterior distribution of coefficients with Cauchy prior overlaid. Our approach to modeling this form of sparse structure uses a mixture-of-Gaussians prior over the coefficients. A set of binary or ternary state variables determine whether the coefficient is in the active or inactive state, and then the coefficient distribution is Gaussian distributed with a variance and mean that depends on the state variable. An important advantage of this approach, with regard to the sampling problems mentioned above, is that the use of Gaussian distributions allows an analytical solution for integrating over the posterior distribution for a given setting of the state variables. The only sampling that needs to be done then is over the binary or ternary state variables. We show here that this problem is a tractable one. This approach differs from that taken previously by Attias [6] in that we do not use variational methods to approximate the posterior, but rather we rely on sampling to adequately characterize the posterior distribution over the coefficients. 2 Mixture-of-Gaussians model An image, I(x, y), is modeled as a linear superposition of basis functions, ?i(X, y), with coefficients ai, plus Gaussian noise II( x, y) : (1) In what follows this will be expressed in vector-matrix notation as 1= q. a + II. The prior probability distribution over the coefficients is factorial, with the distribution over each coefficient ai modeled as a mixture-of-Gaussians distribution with either two or three Gaussians (fig. 2). A set of binary or ternary state variables Si then determine which Gaussian is used to describe the coefficients. The total prior over both sets of variables, a and s, is of the form (2) 843 Learning Sparse Codes with a Mixture-of-Gaussians Prior Two Gaussians (binary state variables) Three Gaussians (ternary state variables) P(lIj) P(lIj) _,;=-1 .<;=1 ai ai Figure 2: Mixture-of-Gaussians prior. where P(Si) determines the probability of being in the active or inactive states, and P(ailsi) is a Gaussian distribution whose mean and variance is determined by the current state Si. The total image probability is then given by P(IIO) = L P(s/O) JP(I/a, O)P(als, O)da (3) s where P(Ila, 0) 1 e -~II-4>aI2 -2 (4) ZAN P(als,O) _l_e-t(a-Il(s))t Aa(s) (a-Il(s)) (5) ZAa(s) P(sIO) 1 _1 s t --e 2 ZA. A s (6) ? and the parameters 0 include AN, 4), Aa(s), f.L(s), and As . Aa(s) is a diagonal inverse covariance matrix with elements Aa(S)ii = Aa; (Si). (The notations Aa(s) and f.L(s) are used here to explicitly reflect the dependence of the means and variances of the ai on sd As is also diagonal (for now) with elements ASii = As;. The model is illustrated graphically in figure 3. Si (binary or ternary) Figure 3: Image model. B. A. Olshausen and K. J. Millman 844 3 Learning The objective function for learning the parameters of the model is the average log-likelihood: (7) ? = (log P(IIO)) Maximizing this objective will minimize the lower bound on coding length. Learning is accomplished via gradient ascent on the objective, ?. The learning rules for the parameters As, Aa (s), J-t( s) and ~ are given by: ex = {)? {)>"Si 1 2 [(Si)P(SiI 9) - (8) (si)P(sII,9)] {)? {)>"ai (u) ! [(8(Si - u))P(sII,9)_ 2 >"ai (u) (8(Si - u) (Kii(U) - 2ai(U)J-ti(U) + J-t~(u)))P(sII,9)] (9) {)? {)J-ti( u) >"ai (u) (8(Si - u) (ai(u) - J-ti(U))) (10) {)? {)~ >"N [I (a(s)) P(sII,9) - ~ (K(s)) P(SII,9)] (11) where u takes on values 0,1 (binary) or -1,0,1 (ternary) and K(s) H- 1 (s) + a(s) a(s)T. (a and H are defined in eqs. 15 and 16 in the next section.) Note that in these expressions we have dropped the outer brackets averaging over images simply to reduce clutter. Thus, for each image we must sample from the posterior P(sll, 0) in order to collect the appropriate statistics needed for learning. These statistics must be accumulated over many different images, and then the parameters are updated according to the rules above. Note that this approach differs from that of Attias [6] in that we do not attempt to sum over all states, s, or to use the variational approximation to approximate the posterior. Instead, we are effectively summing only over those states that are most probable according to the posterior. We conjecture that this scheme will work in practice because the posterior has significant probability only for a small fraction of states s, and so it can be well-characterized by a relatively small number of samples. Next we present an efficient method for Gibbs sampling from the posterior. 4 Sampling and inference In order to sample from the posterior P(sll,O), we first cast it in Boltzmann form: P(sll,O) ex e-E(s) where E(s) = -logP(s,IIO) = -logP(sIO) J P(lla,O)P(als,O)da (12) Learning Sparse Codes with a Mixture-of-Gaussians Prior 845 ~ST Ass + 10gZAa(S) + Eals(a,s) + ~IOgdetH(s) + const. (13) and (14) a = argminEals(a,s) H(s) = \7\7 aEals(a, s) (15) a = )\Nif!T if! + Aa(s) (16) Gibbs-sampling on P(sII, (}) can be performed by flipping state variables Si according to P(Si t- sa) = l+e~E(?i~?a) 1 P(Si t- sa) = (binary) (17) (ternary) (18) Where sa = Si in the binary case, and sa and sf3 are the two alternative states in the ternary case. AE(Si t- sa) denotes the change in E(s) due to changing Si to sa and is given by: (19) where ASi = sa - Si, AAai = Aai (sa) - Aai (Si), J = H- 1 , and Vi = Aai (Si) J.Li(Si). Note that all computations for considering a change of state are local and involve only terms with index i. Thus, deciding whether or not to change state can be computed quickly. However, if a change of state is accepted, then we must update J. Using the Sherman-Morrison formula, this can be kept to an O(N2) computation: J t- J - [ AAak 1 + AAak ] J k Jk (20) Jkk As long as accepted state changes are rare (which we have found to be the case for sparse distributions), then Gibbs sampling may be performed quickly and efficiently. In addition, Hand J are generally very sparse matrices, so as the system is scaled up the number of elements of a that are affected by a flip of Si will be relatively few. In order to code images under this model, a single state of the coefficients must be chosen for a given image. We use for this purpose the MAP estimator: argmaxP(aII,s, (}) (21) arg max P(sII, (}) (22) a s Maximizing the posterior distribution over s is accomplished by assigning a temperature, P(sII, (}) ex e-E(s)/T and gradually lowering it until there are no more state changes. (23) B. A. Olshausen and K. 1. Millman 846 5 5.1 Results Test cases We first trained the algorithm on a number of test cases containing known forms of both sparse and non-sparse (bi-modal) structure, using both critically sampled (complete) and 2x's overcomplete basis sets. The training sets consisted of 6x6 pixel image patches that were created by a sparse superposition of basis functions (36 or 72) with P(ISil = 1) = 0.2, Aa; (0) = 1000, and Aa; (1) = 10. The results of these test cases confirm that the algorithm is capable of correctly extracting both sparse and non-sparse structure from data, and they are not shown here for lack of space. 5.2 Natural images We trained the algorithm on 8x8 image patches extracted from pre-whitened natural images. In all cases, the basis functions were initialized to random functions (white noise) and the prior was initialized to be Gaussian (both Gaussians of roughly equal variance). Shown in figure 4a, b are the results for a set of 128 basis functions (2 x 's overcomplete) in the two-Gausian case. In the three-Gaussian case, the prior was initialized to be platykurtic (all three Gaussians of equal variance but offset at three different positions). Thus, in this case the sparse form of the prior emerged completely from the data. The resulting priors for two of the coefficients are shown in figure 4c, with the posterior distribution averaged over many images overlaid. For some of the coefficients the posterior distribution matches the mixture-of-Gaussians prior well, but for others the tails appear more Laplacian in form. Also, it appears that the extra complexity offered by having three Gaussians is not utilized: Both Gaussians move to the center position and have about the same mean. When a non-sparse, bimodal prior is imposed, the basis function solution does not become localized, oriented, and bandpass as it does with sparse priors. 5.3 Coding efficiency We evaluated the coding efficiency by quantizing the coefficients to different levels and calculating the total coefficient entropy as a function of the distortion introduced by quantization. This was done for basis sets containing 48, 64, 96, and 128 basis functions. At high SNR's the overcomplete basis sets yield better coding efficiency, despite the fact that there are more coefficients to code. However, the point at which this occurs appears to be well beyond the point where errors are no longer perceptually noticeable (around 14 dB). 6 Conclusions We have shown here that both the prior and basis functions of our image model can be adapted to natural images. Without sparseness being imposed, the model both seeks distributions that are sparse and learns the appropriate basis functions for this distribution. Our conjecture that a small number of samples allows the posterior to be sufficiently characterized appears to hold. In all cases here, averages were collected over 40 Gibbs sweeps, with 10 sweeps for initialization. The algorithm proved capable of extracting the structure in challenging datasets in high dimensional spaces. The overcomplete image codes have the lowest coding cost at high SNR levels, but at levels that appear higher than is practically useful. On the other hand, the 847 Learning Sparse Codes with a Mixture-of-Gaussians Prior a. c. : ~~::~~ . - '~ ". ~:. " 10'" . '0" -2 " " 10.... 10~ -. -1 0 I 2 -Z .\ J./ "'\. _I 0 1 2 Figure 4: An overcomplete set of 128 basis functions (a) and priors (b, vertical axis is log-probability) learned from natural images. c, Two of the priors learned from a three-Gaussian mixture using 64 basis functions, with the posterior distribution averaged over many coefficients overlaid. d, Rate distortion curve comparing the coding efficiency of different learned basis sets. sum of marginal entropies likely underestimates the true entropy of the coefficients considerably, as there are certainly statistical dependencies among the coefficients. So it may still be the case that the overcomplete bases will show a win at lower SNR's when these dependencies are included in the model (through the coupling term As). Acknowledgments This work was supported by NIH grant R29-MH057921. References [1] Olshausen BA, Field DJ (1997) "Sparse coding with an overcomplete basis set: A strategy employed by VI?" Vision Research, 37: 3311-3325. [2] Bell AJ, Sejnowski TJ (1997) "The independent components of natural images are edge filters," Vision Research, 37: 3327-3338. [3] van Hateren JH, van der Schaaff A (1997) "Independent component filters of natural images compared with simple cells in primary visual cortex," Proc. Royal Soc. Lond. B, 265: 359-366. [4] Lewicki MS , Olshausen BA (1999) "A probabilistic framework for the adaptation and comparison of image codes," JOSA A, 16(7): 1587-160l. [5] Simoncelli EP, Freeman WT, Adelson EH, Heeger DJ (1992) "Shiftable multiscale transforms," IEEE Transactions on Information Theory, 38(2): 587-607. [6] Attias H (1999) "Independent factor analysis," Neural Computation, 11: 803-852.
1668 |@word isil:1 seek:1 covariance:1 current:1 comparing:1 si:22 assigning:1 must:5 update:1 along:1 sii:9 become:2 manner:1 roughly:1 freeman:1 considering:1 notation:2 lowest:1 what:1 ti:3 scaled:1 grant:1 appear:2 dropped:1 local:1 sd:1 despite:1 plus:1 initialization:1 suggests:1 collect:1 challenging:1 bi:1 averaged:3 acknowledgment:1 ternary:8 practice:1 differs:2 lla:1 asi:1 gabor:1 bell:1 pre:1 integrating:1 ila:1 aai:3 conventional:1 imposed:5 demonstrated:1 center:3 maximizing:2 map:1 graphically:1 rule:2 estimator:1 updated:1 us:1 element:3 jk:1 utilized:1 ep:1 capture:2 mentioned:1 complexity:1 gausian:1 trained:2 efficiency:4 basis:25 completely:1 joint:1 aii:1 describe:2 sejnowski:1 whose:1 emerged:1 distortion:2 statistic:2 advantage:1 quantizing:1 analytical:1 adaptation:1 sll:3 argmaxp:1 coupling:1 noticeable:1 advocated:1 sa:8 eq:1 soc:1 resemble:1 come:1 iio:3 filter:2 centered:1 kii:1 probable:1 hold:1 practically:1 around:1 sufficiently:1 deciding:1 overlaid:4 major:1 continuum:1 purpose:2 proc:1 superposition:2 jkk:1 gaussian:9 rather:2 likelihood:1 inference:1 accumulated:1 pixel:1 arg:1 among:1 orientation:1 priori:1 spatial:2 uc:2 marginal:1 equal:2 field:1 having:1 sampling:8 adelson:1 nearly:1 peaked:1 others:1 few:1 oriented:2 attempt:1 baolshausen:1 certainly:1 mixture:11 bracket:1 tj:1 edge:1 capable:3 initialized:3 overcomplete:12 modeling:2 logp:2 cost:1 rare:1 snr:3 shiftable:1 characterize:1 dependency:2 considerably:1 st:1 probabilistic:1 quickly:2 reflect:1 aaai:1 containing:2 tile:1 li:1 coding:7 coefficient:28 explicitly:1 depends:1 vi:2 performed:2 minimize:1 il:2 variance:7 efficiently:2 yield:2 critically:1 zan:1 za:1 cumbersome:1 underestimate:1 frequency:2 associated:1 josa:1 sampled:2 proved:1 dimensionality:1 appears:3 higher:1 x6:1 modal:1 done:2 though:1 evaluated:1 nif:1 until:1 hand:2 multiscale:1 lack:1 aj:1 olshausen:6 consisted:1 true:1 adequately:1 spatially:1 ai2:1 zaa:1 illustrated:1 white:1 during:1 davis:4 m:1 complete:1 temperature:1 image:29 variational:2 nih:1 jp:1 tail:1 significant:2 gibbs:4 ai:10 bruno:1 sherman:1 dj:2 longer:1 cortex:1 base:2 posterior:19 binary:8 accomplished:2 der:1 employed:1 determine:2 dashed:1 ii:5 morrison:1 simoncelli:1 exceeds:1 match:1 characterized:2 long:1 coded:1 laplacian:1 ae:1 whitened:1 vision:2 bimodal:1 cell:1 addition:1 want:1 extra:1 ascent:1 db:1 seem:1 extracting:2 variety:1 fit:1 psychology:1 reduce:1 attias:3 inactive:3 whether:2 expression:1 useful:2 generally:1 involve:1 factorial:1 transforms:1 clutter:1 occupy:1 exist:1 neuroscience:2 correctly:1 discrete:1 affected:1 changing:1 kept:1 lowering:1 fraction:1 sum:2 inverse:1 patch:2 capturing:1 bound:1 ct:2 adapted:2 sharply:1 lond:1 relatively:2 conjecture:2 department:1 according:3 coo:1 sf3:1 gradually:1 taken:1 previously:1 needed:1 flip:1 tractable:1 gaussians:18 mh057921:1 appropriate:3 alternative:1 denotes:1 include:1 newton:2 const:1 calculating:1 especially:1 sweep:2 objective:3 move:1 flipping:1 occurs:1 strategy:1 primary:1 dependence:1 diagonal:2 gradient:1 win:1 outer:1 cauchy:2 collected:1 length:1 code:10 modeled:3 index:1 ba:2 boltzmann:1 vertical:1 datasets:1 introduced:1 cast:1 learned:5 ucdavis:2 established:1 address:1 beyond:1 suggested:1 below:1 sparsity:1 challenge:1 royal:1 max:1 natural:10 rely:1 eh:1 representing:1 scheme:2 axis:1 created:1 x8:1 lij:2 r29:1 prior:27 millman:4 localized:2 offered:1 supported:1 jh:1 sparse:24 distributed:1 regard:1 curve:1 van:2 made:1 transaction:1 approximate:2 dealing:1 confirm:1 active:4 summing:1 ca:2 as:1 da:2 noise:2 n2:1 fig:1 position:3 heeger:1 bandpass:2 wavelet:1 learns:1 formula:1 offset:1 quantization:1 effectively:1 perceptually:1 sparseness:1 entropy:3 simply:1 likely:1 visual:1 expressed:1 lewicki:1 aa:10 determines:1 extracted:1 change:6 included:1 determined:1 averaging:1 wt:1 total:3 accepted:2 sio:2 hateren:1 ex:3
731
1,669
An Oscillatory Correlation Framework for Computational Auditory Scene Analysis GuyJ.Brown Department of Computer Science University of Sheffield Regent Court, 211 Portobello Street, Sheffield S 1 4DP, UK Email: g.brown@dcs.shefac.uk DeLiang L. Wang Department of Computer and Information Science and Centre for Cognitive Science The Ohio State University Columbus, OH 43210-1277, USA Email: dwang@cis.ohio-state.edu Abstract A neural model is described which uses oscillatory correlation to segregate speech from interfering sound sources. The core of the model is a two-layer neural oscillator network. A sound stream is represented by a synchronized population of oscillators, and different streams are represented by desynchronized oscillator populations. The model has been evaluated using a corpus of speech mixed with interfering sounds, and produces an improvement in signal-to-noise ratio for every mixture. 1 Introduction Speech is seldom heard in isolation: usually, it is mixed with other environmental sounds. Hence, the auditory system must parse the acoustic mixture reaching the ears in order to retrieve a description of each sound source, a process termed auditory scene analysis (ASA) [2] . Conceptually, ASA may be regarded as a two-stage process . The first stage (which we term 'segmentation') decomposes the acoustic stimulus into a collection of sensory elements. In the second stage ('grouping'), elements that are likely to have arisen from the same environmental event are combined into a perceptual structure called a stream. Streams may be further interpreted by higher-level cognitive processes. Recently, there has been a growing interest in the development of computational systems that mimic ASA [4], [1], [5]. Such computational auditory scene analysis (CASA) systems are inspired by auditory function but do not model it closely; rather, they employ symbolic search or high-level inference engines. Although the performance of these systems is encouraging, they are no match for the abilities of a human listener; also, they tend to be complex and computationally intensive. In short, CASA currently remains an unsolved problem for real-time applications such as automatic speech recognition. Given that human listeners can segregate concurrent sounds with apparent ease, computational systems that are more closely modelled on the neurobiological mechanisms of hearing may offer a performance advantage over existing CAS A systems. This observation - together with a desire to understand the neurobiological basis of ASA - has led some investigators to propose neural network models of ASA. Most recently, Brown and Wang [3] have given an account of concurrent vowel separation based on oscillatory correlation. In this framework, oscillators that represent a perceptual stream are synchronized (phase locked with zero phase lag), and are desynchronized from oscillators that represent different streams [8]. Evidence for the oscillatory correlation theory comes from neurobiological studies which report synchronised oscillations in the auditory, visual and olfactory cortices (see [10] for a review). G. J. Brown and D. L. Wang 748 In this paper, we propose a neural network model that uses oscillatory correlation as the underlying neural mechanism for ASA; streams are formed by synchronizing oscillators in a two-dimensional time-frequency network. The model is evaluated on a task that involves the separation of two time-varying sounds. It therefore extends our previous study [3], which only considered the segregation of vowel sounds with static spectra. 2 Model description The input to the model consists of a mixture of speech and an interfering sound source, sampled at a rate of 16 kHz with 16 bit resolution. This input signal is processed in four stages described below (see [10] for a detailed account). 2.1 Peripheral auditory processing Peripheral auditory frequency selectivity is modelled using a bank of 128 gammatone filters with center frequencies equally distributed on the equivalent rectangular bandwidth (ERB) scale between 80 Hz and 5 kHz [1]. Subsequently, the output of each filter is processed by a model of inner hair cell function. The output of the hair cell model is a probabilistic representation of auditory nerve firing activity. 2.2 Mid-level auditory representations Mechanisms similar to those underlying pitch perception can contribute to the perceptual separation of sounds that have different fundamental frequencies (FOs) [3]. Accordingly, the second stage of the model extracts periodicity information from the simulated auditory nerve firing patterns. This is achieved by computing a running autocorrelation of the auditory nerve activity in each channel , forming a representation known as a correlogram [1], [5]. At time step j, the autocorrelation A(iJ,'t) for channel i with time lag 't is given by: K-I I. r(i,j-k)r(i,j-k-'t)w(k) A(i, j,'t) = (1) k=O Here, r is the output of the hair cell model and w is a rectangular window of width K time steps. We use K = 320, corresponding to a window width of 20 ms. The autocorrelation lag 't is computed in L steps of the sampling period between 0 and L-1 ; we use L = 201, corresponding to a maximum delay of 12.5 ms. Equation (1) is computed for M time frames, taken at 10 ms intervals (i .e., at intervals of 160 steps of the time indexj). For periodic sounds, a characteristic 'spine' appears in the correlogram which is centered on the lag corresponding to the stimulus period (Figure 1A). This pitch-related structure can be emphasized by forming a 'pooled' correlogram s(j,'t), which exhibits a prominent peak at the delay corresponding to perceived pitch: N s(j, 't) I. A (i, j, 't) = i (2) =I It is also possible to extract harmonics and formants from the correlogram, since frequency channels that are excited by the same acoustic component share a similar pattern of periodicity. Bands of coherent periodicity can be identified by cross-correlating adjacent correlogram channels; regions of high correlation indicate a harmonic or formant [1] . The cross-correlation C(iJ) between channels i and i+ 1 at time frame j is defined as: L-I C(i,j) = IL.A(i,j, 't)A(i+l,j, 't) t=O (l~i~N-l) (3) Here, A(i, j , 't) is the autocorrelation function of (1) which has been normalized to have zero mean and unity variance. A typical cross-correlation function is shown in Figure 1A. 749 Oscillatory Correlation for CASA 2.3 Neural oscillator network: overview Segmentation and grouping take place within a two-layer oscillator network (Figure IB). The basic unit of the network is a single oscillator, which is defined as a reciprocally connected excitatory variable x and inhibitory variable y [7]. Since each layer of the network takes the form of a time-frequency grid, we index each oscillator according to its frequency channel (i) and time frame (j): Xij = 3xij-xt+2-Yij+lij+Sij+P (4a) Yij = ?(y(1 + tanh(xi/~? (4b) - Yij) Here, Ii} represents external input to the oscillator, Si} denotes the coupling from other oscillators in the network, c, 'Y and ~ are parameters, and p is the amplitude of a Gaussian noise term. If coupling and noise are ignored and Ii} is held constant, (4) defines a relaxation oscillator with two time scales. The x-nullcline, i.e. Xii' = 0, is a cubic function and the y-nullcline is a sigmoid function. If Ii" > 0, the two nul clines intersect only at a point along the middle branch of the cubic with ~ chosen small. In this case, the oscillator exhibits a stable limit cycle for small values of c, and is referred to as enabled. The limit cycle alternates between silent and active phases of near steady-state behaviour. Compared to motion within each phase, the alternation between phases takes place rapidly, and is referred to as jumping. If Ii" < 0, the two nullclines intersect at a stable fixed point. In this case, no oscillation occurs. Hence, oscillations in (4) are stimulus-dependent. 2.4 Neural oscillator network: segment layer In the first layer of the network, segments are formed - blocks of synchronised oscillators that trace the evolution of an acoustic component through time and frequency. The first layer is a two-dimensional time-frequency grid of oscillators with a global inhibitor (see Figure IB). The coupling term Sij in (4a) is defined as ~ Sij = kl E (5) Wij ,k/H(xk/-e x )- WzH(z-e z ) N(i, j) where H is the Heaviside function (i.e., H(x) = I for x ~ 0, and zero otherwise), Wij,kl is the connection weight from an oscillator (iJ) to an oscillator (k,/) and N(iJ) is the four nearest neighbors of (iJ). The threshold ex is chosen so that an oscillator has no influence on its A B 5000 'N ::z:: ';:: 2741 u <= I!.l '"i3" d::... ~I!.l 1457 U Q) <= <= ~ ..c: U i n.D i , 2.5 5.0 j , I 7.5 10.0 12.5 Autocorrelation Lag (ms) Figure I: A. Correlogram of a mixture of speech and trill telephone, taken 450 ms after the start of the stimulus. The pooled correlogram is shown in the bottom panel, and the crosscorrelation function is shown on the right. B. Structure of the two-layer oscillator network. G. J. Brown and D. L. Wang 750 neighbors unless it is in the active phase. The weight of neighboring connections along the time axis is uniformly set to 1. The connection weight between an oscillator (iJ) and its vertical neighbor (i+lJ) is set to 1 if C(iJ) exceeds a threshold Se; otherwise it is set to O. Wz is the weight of inhibition from the global inhibitor z, defined as (6) where <roo = 1 if xi} 2:: Sz for at least one oscillator (iJ), and <roo = 0 otherwise. Hence Sz is a threshold. If <roo = 1, z ~ 1. Small segments may form which do not correspond to perceptually significant acoustic components. In order to remove these noisy fragments, we introduce a lateral potential Pi} for oscillator (iJ), defined as [11]: Pij = (1 - Pij)H[ L.. H(x kl - ex) - epJ - ?Pij (7) kleNp(i,j) Here, Sp is a threshold. Nf(i J ) is called the potential neighborhood of (iJ), which is chosen to be (iJ-l) and (iJ+l). I both neighbors of (iJ) are active, Pi} approaches 1 on a fast time scale; otherwise, Pij relaxes to 0 on a slow time scale determined by c. The lateral potential plays its role by gating the input to an oscillator. More specifically, we replace (4a) with (4a') iij = 3xij-x:j +2-Yij+ lijH(pij-e) +Sij+P With Pij initialized to 1, it follows that Pij will drop below the threshold S unless the oscillator (iJ) receives excitation from its entire potential neighborhood. Given our choice of neighborhood in (5), this implies that a segment must extend for at least three consecutive time frames. Oscillators that are stimulated but cannot maintain a high potential are relegated to a discontiguous 'background' of noisy activity. An oscillator (iJ) is stimulated if its corresponding input lij > O. Oscillators are stimulated only if the energy in their corresponding correlogram channel exceeds a threshold Sa. It is evident from (1) that the energy in a correlogram channel i at time j corresponds to A(iJ,O); thus we set Ii} =0.2 if A(iJ,O) > Sa' and Iij =-5 otherwise. Figure 2A shows the segmentation of a mixture of speech and trill telephone. The network was simulated by the LEGION algorithm [8], producing 94 segments (each represented by a distinct gray level) plus the background (shown in black). For convenience we show all segments together in Figure 2A, but each actually arises during a unique time interval. B g 5000 2741 >. <.) c: <) ~ 1457 ~ !:S 1: <) 729 U 03 c: c: ~ ..c: 315 U 80 0.0 Time (seconds) 1.5 Time (seconds) Figure 2: A. Segments formed by the first layer of the network for a mixture of speech and trill telephone. B. Categorization of segments according to FO. Gray pixels represent the set P, and white pixels represent regions that do not agree with the FO. 751 Oscillatory Correlation for CASA 2.5 Neural oscillator network: grouping layer The second layer is a two-dimensional network of laterally coupled oscillators without global inhibition. Oscillators in this layer are stimulated if the corresponding oscillator in the first layer is stimulated and does not form part of the background. Initially, all oscillators have the same phase, implying that all segments from the first layer are allocated to the same stream. This initialization is consistent with psychophysical evidence suggesting that perceptual fusion is the default state of auditory organisation [2]. In the second layer, an oscillator has the same form as in (4), except that Xu is changed to: iii = 3x ij - x~ + 2 - (4a") Yij + Ii) 1 + !1H(Pij - a)] + Sij + P Here, Jl is a small positive parameter; this implies that an oscillator with a high lateral potential gets a slightly higher external input. We choose NpCiJ) and aR so that oscillators which correspond to the longest segment from the first layer are the first to jump to the active phase. The longest segment is identified by using the mechanism described in [9]. The coupling term in (4a") consists of two types of coupling: e v Sij = Sij + Sij (8) Here, S;j represents mutual excitation between oscillators within each segment. We set S~ = 4 if the active oscillators from the same segment occupy more than half of the length of the segment; otherwise S~j = 0.1 if there is at least one active oscillator from the same segment. S; denotes vertical connections between oscillators corresponding to The coupling term different frequency channels and different segments, but within the same time frame. At each time frame, an FO is estimated from the pooled correlogram (2) and this is used to classify frequency channels into two categories: a set of channels, P, that are consistent with the FO, and a set of channels that are not (Figure 2B). Given the delay 'tm at which the largest peak occurs in the pooled correlogram, for each channel i at time frame j, i E P if AU, j, 't m ) / A(i, j, 0) > ad (9) Since AUJ,O) is the energy in correlogram channel i at time j, (9) amounts to classification on the basis of an energy threshold. We use ad = 0.95. The delay 'tm can be found by using a winner-take-all network, although for simplicity we currently apply a maximum selector. A 5IMM) N N :r:'-' >. :r:'-' 2741 u >. u Q) Q) ;:::l ;:::l c: .. g' c: go 1457 .... Q) c 729 Q) U U "0 c: Q) c: ; ..c: 1457 ~ .... Q) u.. C Q) 2741 ? 315 ..c: U U 80 Time (seconds) Time (seconds) Figure 3: A. Snapshot showing the activity of the second layer shortly after the start of simulation. Active oscillators (white pixels) correspond to the speech stream. B. Another snapshot, taken shortly after A. Active oscillators correspond to the telephone stream. G. J. Brown and D. L. Wang 752 The FO classification process operates on channels, rather than segments. As a result, channels within the same segment at a particular time frame may be allocated to different FO categories. Since segments cannot be decomposed, we enforce a rule that all channels of the same frame within each segment must belong to the same FO category as that of the majority of channels. After this conformational step, vertical connections are fonned such that, at each time frame, two oscillators of different segments have mutual excitatory links if the two corresponding channels belong to the same FO category; otherwise they have mutual inhibitory links. S~ is set to -O.S if (iJ) receives an input from its inhibitory links; similarly, s~ is set to O.S if (iJ) receives an input from its vertical excitatory links. At present, our model has no mechanism for grouping segments that do not overlap in time. Accordingly, we limit operation of the second layer to the time span of the longest segment. After fonning lateral connections and trimming by the longest segment, the network is numerically solved using the singular limit method [6]. Figure 3 shows the response of the second layer to the mixture of speech and trill telephone. The figure shows two snapshots of the second layer, where a white pixel indicates an active oscillator and a black pixel indicates a silent oscillator. The network quickly forms two synchronous blocks, which desynchronize from each other. Figure 3A shows a snapshot taken when the oscillator block (stream) corresponding to the segregated speech is in the active phase; Figure 3B shows a subsequent snapshot when the oscillator block corresponding to the trill telephone is in the active phase. Hence, the activity in this layer of the network embodies the result of ASA; the components of an acoustic mixture have been separated using FO infonnation and represented by oscillatory correlation. 2.6 Resynthesis The last stage of the model is a resynthesis path. Phase-corrected output from the gammatone filterbank is divided into 20 ms sections, overlapping by 10 ms and windowed with a raised cosine. A weighting is then applied to each section, which is unity if the corresponding oscillator is in its active phase, and zero otherwise. The weighted filter outputs are summed across all channels to yield a resynthesized wavefonn. A B 70 90 'c:Q "'""' 60 '~ "'""' "0 '-' '-' .9 50 <;:; .... 0 til 0 0 .... 70 > 0 u ~ ?0 I B e;; 20 s= ft- tlO . r;; s= ?I :; 0 -10 ? F ?n,~If .i~ ::1 10 ? ? I ?? I I ..c: u 0 0 .J1l NO Nl N2 N3 N4 N5 N6 N7 N8 N9 Intrusion type 0.. CIl -,,- ",.- n I" ?' Ir r-- F 60 ? 50 f:.Il o 40 s= 0 - I, I 0 80 "0 40 r-- s= 30 100 :;-'\ H< ~ 'i itllE 30 20 Bi IO 0 ' J "f J U:, ~ ", mfi t "1 ',' il t' f l f' j: ,d NO Nl N2 N3 N4 N5 N6 N7 N8 N9 Intrusion type Figure 4: A. SNR before (black bar) and after (grey bar) separation by the model. Results are shown for voiced speech mixed with ten intrusions (NO = 1 kHz tone; Nl = random noise; N2 = noise bursts; N3 = 'cocktail party' noise; N4 = rock music; NS = siren; N6 = trill telephone; N7 = female speech; N8 = male speech; N9 = female speech). B. Percentage of speech energy recovered from each mixture after separation by the model.
1669 |@word middle:1 grey:1 simulation:1 excited:1 n8:3 tlo:1 fragment:1 existing:1 recovered:1 si:1 must:3 subsequent:1 remove:1 drop:1 implying:1 half:1 tone:1 accordingly:2 xk:1 core:1 short:1 contribute:1 windowed:1 along:2 burst:1 consists:2 regent:1 autocorrelation:5 olfactory:1 introduce:1 discontiguous:1 spine:1 growing:1 formants:1 inspired:1 decomposed:1 nullcline:2 encouraging:1 window:2 underlying:2 panel:1 interpreted:1 every:1 nf:1 laterally:1 filterbank:1 uk:2 unit:1 producing:1 positive:1 before:1 limit:4 io:1 mfi:1 firing:2 path:1 black:3 plus:1 initialization:1 au:1 ease:1 locked:1 bi:1 unique:1 block:4 intersect:2 symbolic:1 get:1 cannot:2 convenience:1 influence:1 equivalent:1 center:1 go:1 rectangular:2 resolution:1 simplicity:1 rule:1 regarded:1 oh:1 retrieve:1 enabled:1 population:2 play:1 us:2 element:2 recognition:1 bottom:1 role:1 ft:1 wang:5 solved:1 region:2 connected:1 cycle:2 segment:24 asa:7 basis:2 represented:4 listener:2 separated:1 distinct:1 fast:1 neighborhood:3 apparent:1 lag:5 roo:3 otherwise:8 ability:1 formant:1 noisy:2 advantage:1 rock:1 propose:2 neighboring:1 rapidly:1 gammatone:2 description:2 produce:1 categorization:1 coupling:6 nearest:1 ij:20 ex:2 sa:2 involves:1 come:1 indicate:1 synchronized:2 implies:2 closely:2 filter:3 subsequently:1 centered:1 human:2 cline:1 behaviour:1 yij:5 auj:1 considered:1 consecutive:1 perceived:1 currently:2 tanh:1 infonnation:1 concurrent:2 largest:1 weighted:1 inhibitor:2 gaussian:1 i3:1 reaching:1 rather:2 nullclines:1 varying:1 improvement:1 longest:4 indicates:2 intrusion:3 inference:1 dependent:1 lj:1 entire:1 initially:1 relegated:1 wij:2 pixel:5 classification:2 development:1 raised:1 summed:1 mutual:3 legion:1 sampling:1 represents:2 synchronizing:1 mimic:1 report:1 stimulus:4 employ:1 phase:12 vowel:2 maintain:1 interest:1 trimming:1 male:1 mixture:9 nl:3 held:1 jumping:1 unless:2 initialized:1 classify:1 ar:1 hearing:1 snr:1 delay:4 periodic:1 combined:1 fundamental:1 peak:2 probabilistic:1 together:2 quickly:1 ear:1 choose:1 cognitive:2 external:2 til:1 crosscorrelation:1 account:2 potential:6 suggesting:1 pooled:4 ad:2 stream:11 start:2 voiced:1 formed:3 il:3 ir:1 variance:1 characteristic:1 correspond:4 yield:1 conceptually:1 modelled:2 oscillatory:8 fo:10 email:2 energy:5 frequency:11 unsolved:1 static:1 sampled:1 auditory:13 segmentation:3 amplitude:1 actually:1 nerve:3 appears:1 higher:2 response:1 evaluated:2 stage:6 correlation:11 receives:3 parse:1 overlapping:1 defines:1 gray:2 columbus:1 usa:1 brown:6 normalized:1 evolution:1 hence:4 white:3 adjacent:1 during:1 width:2 portobello:1 steady:1 excitation:2 cosine:1 m:7 prominent:1 evident:1 motion:1 harmonic:2 ohio:2 recently:2 sigmoid:1 overview:1 khz:3 winner:1 jl:1 extend:1 belong:2 numerically:1 significant:1 automatic:1 seldom:1 grid:2 similarly:1 centre:1 stable:2 cortex:1 inhibition:2 deliang:1 female:2 termed:1 selectivity:1 alternation:1 period:2 signal:2 ii:6 branch:1 sound:11 exceeds:2 match:1 offer:1 cross:3 divided:1 equally:1 pitch:3 basic:1 sheffield:2 n5:2 hair:3 represent:4 arisen:1 achieved:1 cell:3 background:3 interval:3 singular:1 source:3 desynchronized:2 allocated:2 hz:1 tend:1 resynthesized:1 n7:3 near:1 iii:1 relaxes:1 isolation:1 bandwidth:1 identified:2 silent:2 inner:1 tm:2 court:1 intensive:1 synchronous:1 speech:16 cocktail:1 ignored:1 heard:1 detailed:1 se:1 conformational:1 j1l:1 amount:1 mid:1 band:1 ten:1 processed:2 category:4 occupy:1 xij:3 percentage:1 inhibitory:3 estimated:1 xii:1 erb:1 four:2 threshold:7 relaxation:1 extends:1 place:2 separation:5 oscillation:3 bit:1 layer:20 activity:5 scene:3 n3:3 span:1 department:2 according:2 alternate:1 peripheral:2 across:1 slightly:1 unity:2 dwang:1 wavefonn:1 n4:3 sij:8 taken:4 computationally:1 segregation:1 equation:1 remains:1 agree:1 desynchronize:1 mechanism:5 operation:1 apply:1 enforce:1 shortly:2 n9:3 denotes:2 running:1 music:1 embodies:1 psychophysical:1 occurs:2 exhibit:2 dp:1 link:4 simulated:2 lateral:4 street:1 majority:1 length:1 index:1 ratio:1 trace:1 vertical:4 observation:1 snapshot:5 segregate:2 dc:1 frame:10 fonning:1 kl:3 connection:6 acoustic:6 engine:1 coherent:1 bar:2 usually:1 below:2 perception:1 pattern:2 trill:6 wz:1 reciprocally:1 event:1 overlap:1 siren:1 axis:1 extract:2 coupled:1 lij:2 n6:3 review:1 segregated:1 mixed:3 pij:8 consistent:2 bank:1 share:1 interfering:3 pi:2 periodicity:3 excitatory:3 changed:1 last:1 understand:1 neighbor:4 distributed:1 default:1 sensory:1 collection:1 jump:1 fonned:1 party:1 selector:1 neurobiological:3 sz:2 global:3 correlating:1 active:12 imm:1 corpus:1 xi:2 spectrum:1 search:1 decomposes:1 stimulated:5 channel:20 ca:1 complex:1 sp:1 noise:6 n2:3 xu:1 referred:2 cubic:2 slow:1 cil:1 iij:2 n:1 perceptual:4 ib:2 weighting:1 casa:4 emphasized:1 xt:1 gating:1 showing:1 evidence:2 grouping:4 fusion:1 organisation:1 ci:1 perceptually:1 led:1 likely:1 forming:2 visual:1 desire:1 correlogram:12 corresponds:1 environmental:2 oscillator:49 replace:1 typical:1 telephone:7 uniformly:1 determined:1 specifically:1 except:1 operates:1 corrected:1 called:2 synchronised:2 arises:1 investigator:1 heaviside:1 resynthesis:2
732
167
794 NEURAL ARCHITECTURE Valentino Braitenberg Max Planck Institute Federal Republic of Germany While we are waiting for the ultimate biophysics of cell membranes and synapses to be completed, we may speculate on the shapes of neurons and on the patterns of their connections. Much of this will be significant whatever the outcome of future physiology. Take as an example the isotropy, anisotropy and periodicity of different kinds of neural networks. The very existence of these different types in different parts of the brain (or in different brains) defeats explanation in terms of embryology; the mechanisms of development are able to make one kind of network or another. The reasons for the difference must be in the functions they perform. The tasks which they solve in one case apparently refer to some space which is intrinsically isotropic, in another to a situation in which different coordinates mean different things. In the periodic case, the tasks obviously refer to some kind of modules and to their relations. The examples I have in mind are first the cerebral cortex, quite isotropic in the plane of the cortex, second the cerebellar cortex with very different sets of fibers at right angles to each other (one excitatory, as we know today, and the other inhibitory) and third some of the nerve nets behind the eye of the fly. Besides general patterns of symmetry, some simple statements of a statistical nature can be read off the histological picture. If a neuron is a device picking up excitation (and/or inhibition) on its ten to ten thousand afferent synapses and producing excitation (or inhibition) on ten to ten thousand synapses on other neurons, the density, geometrical distribution and reciprocal overlap of the clouds of afferent and of efferent synapses of individual neurons provide unquestionable constraints to neural computation. In the simple terms of the histological practitioner, this translates into the description of the shapes of dendritic and axonal trees, into counts of neurons and synapses, differential counts of synapses of the excitatory and inhibitory kind and measurements of the axonal and dendritic lengths.
167 |@word read:1 excitation:2 dendritic:2 reason:1 besides:1 geometrical:1 length:1 must:1 statement:1 shape:2 defeat:1 cerebral:1 perform:1 device:1 significant:1 refer:2 neuron:5 isotropic:2 plane:1 measurement:1 reciprocal:1 federal:1 situation:1 cortex:3 inhibition:2 differential:1 connection:1 brain:2 able:1 pattern:2 anisotropy:1 relation:1 max:1 germany:1 explanation:1 isotropy:1 overlap:1 kind:4 development:1 eye:1 biophysics:1 picture:1 future:1 braitenberg:1 cerebellar:1 whatever:1 cell:1 planck:1 producing:1 individual:1 thing:1 practitioner:1 axonal:2 behind:1 periodicity:1 excitatory:2 architecture:1 histological:2 institute:1 translates:1 tree:1 physiology:1 ultimate:1 republic:1 ten:4 periodic:1 inhibitory:2 density:1 valentino:1 nature:1 off:1 coordinate:1 picking:1 symmetry:1 waiting:1 today:1 cloud:1 module:1 angle:1 speculate:1 fly:1 thousand:2 afferent:2 apparently:1 third:1 constraint:1 fiber:1 synapsis:6 outcome:1 membrane:1 quite:1 solve:1 unquestionable:1 efferent:1 intrinsically:1 obviously:1 net:1 count:2 mechanism:1 mind:1 know:1 nerve:1 description:1 existence:1 completed:1
733
1,670
Emergence of Topography and Complex Cell Properties from Natural Images using Extensions of ICA Aapo Hyviirinen and Patrik Hoyer Neural Networks Research Center Helsinki University of Technology P.O. Box 5400, FIN-02015 HUT, Finland aapo.hyvarinen~hut.fi, patrik.hoyer~hut.fi http://www.cis.hut.fi/projects/ica/ Abstract Independent component analysis of natural images leads to emergence of simple cell properties, Le. linear filters that resemble wavelets or Gabor functions. In this paper, we extend ICA to explain further properties of VI cells. First, we decompose natural images into independent subspaces instead of scalar components. This model leads to emergence of phase and shift invariant features, similar to those in VI complex cells. Second, we define a topography between the linear components obtained by ICA. The topographic distance between two components is defined by their higher-order correlations, so that two components are close to each other in the topography if they are strongly dependent on each other. This leads to simultaneous emergence of both topography and invariances similar to complex cell properties. 1 Introduction A fundamental approach in signal processing is to design a statistical generative model of the observed signals. Such an approach is also useful for modeling the properties of neurons in primary sensory areas. The basic models that we consider here express a static monochrome image J (x, y) as a linear superposition of some features or basis functions bi (x, y): n J(x, y) = 2: bi(x, Y)Si (1) i=l where the Si are stochastic coefficients, different for each image J(x, y). Estimation of the model in Eq. (1) consists of determining the values of Si and bi(x, y) for all i and (x, y), given a sufficient number of observations of images, or in practice, image patches J(x,y). We restrict ourselves here to the basic case where the bi(x,y) form an invertible linear system. Then we can invert Si =< Wi, J > where the Wi denote the inverse filters, and < Wi, J >= L.x,y Wi(X, y)J(x, y) denotes the dot-product. 828 A. Hyviirinen and P Hoyer The Wi (x, y) can then be identified as the receptive fields of the model simple cells, and the Si are their activities when presented with a given image patch I(x, y). In the basic case, we assume that the Si are nongaussian, and mutually independent. This type of decomposition is called independent component analysis (ICA) [3, 9, 1, 8], or sparse coding [13]. Olshausen and Field [13] showed that when this model is estimated with input data consisting of patches of natural scenes, the obtained filters Wi(X,y) have the three principal properties of simple cells in VI: they are localized, oriented, and bandpass (selective to scale/frequency). Van Hateren and van der Schaaf [15] compared quantitatively the obtained filters Wi(X, y) with those measured by single-cell recordings of the macaque cortex, and found a good match for most of the parameters. We show in this paper that simple extensions of the basic ICA model explain emergence of further properties of VI cells: topography and the invariances of complex cells. Due to space limitations, we can only give the basic ideas in this paper. More details can be found in [6, 5, 7]. First, using the method of feature subspaces [11], we model the response of a complex cell as the norm of the projection of the input vector (image patch) onto a linear subspace, which is equivalent to the classical energy models. Then we maximize the independence between the norms of such projections, or energies. Thus we obtain features that are localized in space, oriented, and bandpass, like those given by simple cells, or Gabor analysis. In contrast to simple linear filters, however, the obtained feature subspaces also show emergence of phase invariance and (limited) shift or translation invariance. Maximizing the independence, or equivalently, the sparseness of the norms of the projections to feature subspaces thus allows for the emergence of exactly those invariances that are encountered in complex cells. Second, we extend this model of independent subspaces so that we have overlapping subspaces, and every subspace corresponds to a neighborhood on a topographic grid. This is called topographic ICA, since it defines a topographic organization between components. Components that are far from each other on the grid are independent, like in ICA. In contrast, components that are near to each other are not independent: they have strong higher-order correlations. This model shows emergence of both complex cell properties and topography from image data. 2 Independent subspaces as complex cells In addition to the simple cells that can be modelled by basic ICA, another important class of cells in VI is complex cells. The two principal properties that distinguish complex cells from simple cells are phase invariance and (limited) shift invariance. The purpose of the first model in this paper is to explain the emergence of such phase and shift invariant features using a modification of the ICA model. The modification is based on combining the principle of invariant-feature subspaces [11] and the model of multidimensional independent component analysis [2]. Invariant feature subspaces. The principle of invariant-feature subspaces states that one may consider an invariant feature as a linear subspace in a feature space. The value of the invariant, higher-order feature is given by (the square of) the norm of the projection of the given data point on that subspace, which is typically spanned by lower-order features. A feature subspace, as any linear subspace, can always be represented by a set of orthogonal basis vectors, say Wi(X, y), i = 1, ... , m, where m is the dimension of the subspace. Then the value F(I) of the feature F with input vector I(x, y) is given by F(I) = L::l < Wi, I >2, where a square root 829 Emergence of VI properties using Extensions ofleA might be taken. In fact, this is equivalent to computing the distance between the input vector I (X, y) and a general linear combination of the basis vectors (filters) Wi(X, y) of the feature subspace [11]. In [11], it was shown that this principle, when combined with competitive learning techniques, can lead to emergence of invariant image features. Multidimensional independent component analysis. In multidimensional independent component analysis [2] (see also [12]), a linear generative model as in Eq. (1) is assumed. In contrast to ordinary leA, however, the components (responses) Si are not assumed to be all mutually independent. Instead, it is assumed that the Si can be divided into couples, triplets or in general m-tuples, such that the Si inside a given m-tuple may be dependent on each other, but dependencies between different m-tuples are not allowed. Every m-tuple of Si corresponds to m basis vectors bi(x, y). The m-dimensional probability densities inside the m-tuples of Si is not specified in advance in the general definition of multidimensional leA [2]. In the following, let us denote by J the number of independent feature subspaces, and by Sj,j = 1, ... , J the set of the indices of the Si belonging to the subspace of index j . Independent feature subspaces. Invariant-feature subspaces can be embedded in multidimensional independent component analysis by considering probability distributions for the m-tuples of Si that are spherically symmetric, i.e. depend only on the norm. In other words, the probability density Pj (.) of the m-tuple with index j E {1, ... , J}, can be expressed as a function of the sum of the squares of the si,i E Sj only. For simplicity, we assume further that the Pj(') are equal for all j, i.e. for all subspaces. Assume that the data consists of K observed image patches I k (x, y), k = 1, ... , K. Then the logarithm of the likelihood L of the data given the model can be expressed as K 10gL(wi(x, y), i = L.n) = L J L 10gp(L < Wi, h >2) + Klog Idet WI k=1 j=1 (2) iESj where P(LiESj sT) = pj(si,i E Sj) gives the probability density inside the j-th m-tuple of Si, and W is a matrix containing the filters Wi(X, y) as its columns. As in basic leA, prewhitening of the data allows us to consider the Wi(X, y) to be orthonormal, and this implies that log I det WI is zero [6]. Thus we see that the likelihood in Eq. (2) is a function of the norms of the projections of Ik(x,y) on the subspaces indexed by j, which are spanned by the orthonormal basis sets given by Wi(X, y), i E Sj. Since the norm of the projection of visual data on practically any subspace has a supergaussian distribution, we need to choose the probability density P in the model to be sparse [13], i.e. supergaussian [8]. For example, we could use the following probability distribution logp( L st) = -O:[L s~11/2 + {3, iESj (3) iESj which could be considered a multi-dimensional version of the exponential distribution. Now we see that the estimation of the model consists of finding subspaces such that the norms of the projections of the (whitened) data on those subspaces have maximally sparse distributions. The introduced "independent (feature) subspace analysis" is a natural generalization of ordinary leA. In fact, if the projections on the subspaces are reduced to dotproducts, i.e. projections on 1-D subs paces , the model reduces to ordinary leA A. Hyviirinen and P. Hoyer 830 (provided that, in addition, the independent components are assumed to have nonskewed distributions). It is to be expected that the norms of the projections on the subspaces represent some higher-order, invariant features. The exact nature of the invariances has not been specified in the model but will emerge from the input data, using only the prior information on their independence. When independent subspace analysis is applied to natural image data, we can identify the norms of the projections (2:iESj st)1/2 as the responses of the complex cells. If the individual filter vectors Wi(X, y) are identified with the receptive fields of simple cells, this can be interpreted as a hierarchical model where the complex cell response is computed from simple cell responses Si, in a manner similar to the classical energy models for complex cells. Experiments (see below and [6]) show that the model does lead to emergence of those invariances that are encountered in complex cells. 3 Topographic leA The independent subspace analysis model introduces a certain dependence structure for the components Si. Let us assume that the distribution in the subspace is sparse, which means that the norm of the projection is most of the time very near to zero. This is the case, for example, if the densities inside the subspaces are specified as in (3). Then the model implies that two components Si and Sj that belong to the same subspace tend to be nonzero simultaneously. In other words, and S] are positively correlated. This seems to be a preponderant structure of dependency in most natural data. For image data, this has also been noted by Simoncelli [14). s; Now we generalize the model defined by (2) so that it models this kind of dependence not only inside the m-tuples, but among all ''neighboring'' components. A neighborhood relation defines a topographic order [10). (A different generalization based on an explicit generative model is given in [5].) We define the model by the following likelihood: K n n 10gL(wi(x,y),i = 1, ... ,n) = LLG(Lh(i,j) k=I j=l < Wi,h >2) +KlogldetWI (4) i=l Here, h(i, j) is a neighborhood function, which expresses the strength of the connection between the i-th and j-th units. The neighborhood function can be defined in the same way as with the self-organizing map [10). Neighborhoods can thus be defined as one-dimensional or two-dimensional; 2-D neighborhoods can be square or hexagonal. A simple example is to define a 1-D neighborhood relation by h(i,j) = {I, 0, if Ii - ~I ~ m otherwIse. (5) The constant m defines here the width of the neighborhood. The function G has a similar role as the log-density of the independent components in classic ICA. For image data, or other data with a sparse structure, G should be chosen as in independent subspace analysis, see Eq. (3). Properties of the topographic leA model. Here, we consider for simplicity only the case of sparse data. The first basic property is that all the components Si are uncorrelated, as can be easily proven by symmetry arguments [5]. Moreover, their variances can be defined to be equal to unity, as in classic ICA. Second, components Si and S j that are near to each other, Le. such that h( i, j) is significantly non-zero, Emergence oj VI properties using Extensions ojleA 831 tend to be active (non-zero) at the same time. In other words, their energies sf and s; are positively correlated. Third, latent variables that are far from each other are practically independent. Higher-order correlation decreases as a function of distance, assuming that the neighborhood is defined in a way similar to that in (5). For details, see [5]. Let us note that our definition of topography by higher-order correlations is very different from the one used in practically all existing topographic mapping methods. Usually, the distance is defined by basic geometrical relations like Euclidean distance or correlation. Interestingly, our principle makes it possible to define a topography even among a set of orthogonal vectors whose Euclidean distances are all equal. Such orthogonal vectors are actually encountered in leA, where the basis vectors and filters can be constrained to be orthogonal in the whitened space. 4 Experiments with natural image data We applied our methods on natural image data. The data was obtained by taking 16 x 16 pixel image patches at random locations from monochrome photographs depicting wild-life scenes (animals, meadows, forests, etc.). Preprocessing consisted of removing the De component and reducing the dimension of the data to 160 by peA. For details on the experiments, see [6, 5]. Fig. 1 shows the basis vectors of the 40 feature subspaces (complex cells), when subspace dimension was chosen to be 4. It can be seen that the basis vectors associated with a single complex cell all have approximately the same orientation and frequency. Their locations are not identical, but close to each other. The phases differ considerably. Every feature subspace can thus be considered a generalization of a quadrature-phase filter pair as found in the classical energy models, enabling the cell to be selective to some given orientation and frequency, but invariant to phase and somewhat invariant to shifts. Using 4 dimensions instead of 2 greatly enhances the shift invariance of the feature subspace. In topographic leA, the neighborhood function was defined so that every neighborhood consisted of a 3 x 3 square of 9 units on a 2-D torus lattice [10]. The obtained basis vectors, are shown in Fig. 2. The basis vectors are similar to those obtained by ordinary leA of image data [13, 1]. In addition, they have a clear topographic organization. In addition, the connection to independent subspace analysis is clear from Fig. 2. Two neighboring basis vectors in Fig. 2 tend to be of the same orientation and frequency. Their locations are near to each other as well. In contrast, their phases are very different. This means that a neighborhood of such basis vectors, i.e. simple cells, is similar to an independent subspace. Thus it functions as a complex cell. This was demonstrated in detail in [5]. 5 Discussion We introduced here two extensions of leA that are especially useful for image modelling. The first model uses a subspace representation to model invariant features. It turns out that the independent subspaces of natural images are similar to complex cells. The second model is a further extension of the independent subspace model. This topographic leA model is a generative model that combines topographic mapping with leA. As in all topographic mappings, the distance in the representation space (on the topographic "grid") is related to some measure of distance between represented components. In topographic leA, the distance between represented components is defined by higher-order correlations, which gives 832 A. Hyviirinen and P Hoyer the natural distance measure in the context of leA. An approach closely related to ours is given by Kohonen's Adaptive Subspace SelfOrganizing Map [11). However, the emergence of shift invariance in [11) was conditional to restricting consecutive patches to come from nearby locations in the image, giving the input data a temporal structure like in a smoothly changing image sequence. Similar developments were given by F6ldiak [4). In contrast to these two theories, we formulated an explicit image model. This independent subspace analysis model shows that emergence of complex cell properties is possible using patches at random, independently selected locations, which proves that there is enough information in static images to explain the properties of complex cells. Moreover, by extending this subspace model to model topography, we showed that the emergence of both topography and complex cell properties can be explained by a single principle: neighboring cells should have strong higher-order correlations. References [1] A.J. Bell and T.J. Sejnowski. The 'independent components' of natural scenes are edge filters. Vision Research, 37:3327-3338, 1997. [2] J.-F. Cardoso. Multidimensional independent component analysis. In Proc. IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP'98), Seattle, WA, 1998. [3] P. Comon. Independent component analysis - a new concept? Signal Processing, 36:287-314, 1994. [4] P. Foldiak. Learning invariance from transformation sequences. Neural Computation, 3:194-200, 1991. [5] A. Hyvarinen and P. O. Hoyer. Topographic independent component analysis. 1999. Submitted, available at http://www.cis.hut.firaapo/. [6] A. Hyvarinen and P. O. Hoyer. Emergence of phase and shift invariant features by decomposition of natur:al images into independent feature subspaces. Neural Computation, 2000. (in press). [7] A. Hyvarinen, P. O. Hoyer, and M. Inki. The independence assumption: Analyzing the independence of the components by topography. In M. Girolami, editor, Advances in Independent Component Analysis. Springer-Verlag, 2000. in press. [8] A. Hyvarinen and E. Oja. A fast fixed-point algorithm for independent component analysis. Neural Computation, 9(7):1483-1492, 1997. [9] C. Jutten and J. Herault. Blind separation of sources, part I: An adaptive algorithm based on neuromimetic architecture. Signal Processing, 24:1-10, 1991. [10] T. Kohonen. Self-Organizing Maps. Springer-Verlag, Berlin, Heidelberg, New York, 1995. [11] T. Kohonen. Emergence of invariant-feature detectors in the adaptive-subspace selforganizing map. Biological Cybernetics, 75:281-291, 1996. [12] J. K. Lin. Factorizing multivariate function classes. In Advances in Neural Information Processing Systems, volume 10, pages 563-569. The MIT Press, 1998. [13] B. A. Olshausen and D. J. Field. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature, 381:607-609, 1996. [14] E. P. Simoncelli and O. Schwartz. Modeling surround suppression in VI neurons with a statistically-derived normalization model. In Advances in Neural Information Processing Systems 11, pages 153-159. MIT Press, 1999. [15] J. H. van Hateren and A. van der Schaaf. Independent component filters of natural images compared with simple cells in primary visual cortex. Proc. Royal Society ser. B, 265:359-366, 1998. Emergence of Vi properties using Extensions of leA - ", 833 ;II :II ? ? ? .. ? ? .. -, ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? '11'1 .; ? Figure 1: Independent subspaces of natural image data. The model gives Gaborlike basis vectors for image windows. Every group of four basis vectors corresponds to one independent feature subspace, or complex cell. Basis vectors in a subspace are similar in orientation, location and frequency. In contrast, their phases are very different. .- ? ? ~ I ? ? I; IiioiII iii ? ? " I . iii I , 'i ~ ." . I ? ? Figure 2: Topographic leA of natural image data. This gives Gabor-like basis vectors as well. Basis vectors that are similar in orientation, location and/or frequency are close to each other. The phases of near by basis vectors are very different, giving each neighborhood properties similar to a complex cell.
1670 |@word version:1 norm:11 seems:1 decomposition:2 ours:1 interestingly:1 existing:1 si:21 generative:4 selected:1 location:7 ik:1 consists:3 wild:1 combine:1 inside:5 manner:1 ica:12 expected:1 multi:1 window:1 considering:1 project:1 provided:1 moreover:2 kind:1 interpreted:1 finding:1 transformation:1 temporal:1 every:5 multidimensional:6 exactly:1 schwartz:1 ser:1 unit:2 analyzing:1 approximately:1 might:1 limited:2 klog:1 bi:5 statistically:1 practice:1 area:1 bell:1 gabor:3 significantly:1 projection:12 word:3 onto:1 close:3 context:1 www:2 equivalent:2 map:4 demonstrated:1 center:1 maximizing:1 independently:1 simplicity:2 spanned:2 orthonormal:2 classic:2 exact:1 us:1 observed:2 role:1 decrease:1 depend:1 basis:18 easily:1 icassp:1 represented:3 fast:1 sejnowski:1 neighborhood:13 whose:1 say:1 otherwise:1 topographic:17 gp:1 emergence:20 sequence:2 product:1 neighboring:3 kohonen:3 combining:1 organizing:2 seattle:1 extending:1 measured:1 eq:4 strong:2 resemble:1 implies:2 come:1 girolami:1 differ:1 closely:1 filter:12 stochastic:1 pea:1 generalization:3 decompose:1 biological:1 extension:7 practically:3 hut:5 considered:2 mapping:3 finland:1 consecutive:1 purpose:1 estimation:2 proc:2 superposition:1 supergaussian:2 mit:2 always:1 derived:1 monochrome:2 modelling:1 likelihood:3 greatly:1 contrast:6 suppression:1 dependent:2 typically:1 relation:3 selective:2 pixel:1 among:2 orientation:5 herault:1 development:1 animal:1 constrained:1 schaaf:2 field:5 equal:3 identical:1 quantitatively:1 oriented:2 oja:1 simultaneously:1 individual:1 phase:11 ourselves:1 consisting:1 organization:2 llg:1 introduces:1 prewhitening:1 tuple:4 edge:1 lh:1 orthogonal:4 indexed:1 euclidean:2 logarithm:1 column:1 modeling:2 logp:1 lattice:1 ordinary:4 dependency:2 considerably:1 combined:1 st:3 density:6 fundamental:1 invertible:1 nongaussian:1 containing:1 choose:1 conf:1 de:1 coding:1 coefficient:1 int:1 vi:9 blind:1 root:1 competitive:1 square:5 variance:1 identify:1 generalize:1 modelled:1 cybernetics:1 submitted:1 explain:4 simultaneous:1 detector:1 definition:2 energy:5 frequency:6 associated:1 static:2 couple:1 actually:1 higher:8 response:5 maximally:1 box:1 strongly:1 correlation:7 overlapping:1 defines:3 jutten:1 olshausen:2 consisted:2 concept:1 spherically:1 symmetric:1 nonzero:1 self:2 width:1 noted:1 geometrical:1 image:31 fi:3 inki:1 volume:1 extend:2 belong:1 surround:1 grid:3 dot:1 cortex:2 meadow:1 etc:1 multivariate:1 showed:2 foldiak:1 dotproducts:1 certain:1 verlag:2 life:1 der:2 seen:1 somewhat:1 maximize:1 signal:5 ii:3 simoncelli:2 reduces:1 match:1 lin:1 divided:1 aapo:2 basic:9 whitened:2 vision:1 represent:1 normalization:1 invert:1 cell:40 lea:17 addition:4 source:1 recording:1 tend:3 near:5 iii:2 enough:1 independence:5 architecture:1 restrict:1 identified:2 idea:1 det:1 shift:8 speech:1 york:1 useful:2 clear:2 cardoso:1 selforganizing:2 reduced:1 http:2 estimated:1 pace:1 express:2 group:1 four:1 changing:1 pj:3 sum:1 inverse:1 patch:8 separation:1 distinguish:1 encountered:3 activity:1 strength:1 helsinki:1 scene:3 nearby:1 argument:1 hexagonal:1 combination:1 belonging:1 unity:1 wi:20 modification:2 comon:1 explained:1 invariant:15 taken:1 mutually:2 turn:1 idet:1 neuromimetic:1 available:1 hierarchical:1 denotes:1 giving:2 especially:1 prof:1 classical:3 society:1 receptive:3 primary:2 dependence:2 hoyer:8 enhances:1 subspace:53 distance:10 berlin:1 assuming:1 code:1 index:3 equivalently:1 f6ldiak:1 design:1 neuron:2 observation:1 fin:1 enabling:1 introduced:2 pair:1 specified:3 connection:2 acoustic:1 macaque:1 below:1 usually:1 oj:1 royal:1 natural:16 technology:1 patrik:2 prior:1 determining:1 embedded:1 topography:11 limitation:1 proven:1 localized:2 sufficient:1 principle:5 editor:1 uncorrelated:1 translation:1 gl:2 taking:1 emerge:1 sparse:7 van:4 dimension:4 sensory:1 adaptive:3 preprocessing:1 far:2 hyvarinen:5 sj:5 active:1 assumed:4 tuples:5 factorizing:1 latent:1 triplet:1 hyviirinen:4 nature:2 symmetry:1 depicting:1 forest:1 heidelberg:1 complex:23 allowed:1 quadrature:1 positively:2 fig:4 sub:1 explicit:2 bandpass:2 exponential:1 sf:1 torus:1 third:1 wavelet:1 removing:1 restricting:1 ci:2 sparseness:1 smoothly:1 photograph:1 visual:2 expressed:2 scalar:1 springer:2 corresponds:3 conditional:1 formulated:1 reducing:1 principal:2 called:2 invariance:12 hateren:2 correlated:2
734
1,671
Training Data Selection for Optimal Generalization in Trigonometric Polynomial Networks Masashi Sugiyama*and Hidemitsu Ogawa Department of Computer Science, Tokyo Institute of Technology, 2-12-1, O-okayama, Meguro-ku, Tokyo, 152-8552, Japan. sugi@cs. titeck. ac.jp Abstract In this paper, we consider the problem of active learning in trigonometric polynomial networks and give a necessary and sufficient condition of sample points to provide the optimal generalization capability. By analyzing the condition from the functional analytic point of view, we clarify the mechanism of achieving the optimal generalization capability. We also show that a set of training examples satisfying the condition does not only provide the optimal generalization but also reduces the computational complexity and memory required for the calculation of learning results. Finally, examples of sample points satisfying the condition are given and computer simulations are performed to demonstrate the effectiveness of the proposed active learning method. 1 Introduction Supervised learning is obtaining an underlying rule from training examples , and can be formulated as a function approximation problem. If sample points are actively designed, then learning can be performed more efficiently. In this paper, we discuss the problem of designing sample points, referred to as active learning, for optimal generalization. Active learning is classified into two categories depending on the optimality. One is global optimal, where a set of all training examples is optimal (e.g. Fedorov [3]) . The other is greedy optimal, where the next training example to sample is optimal in each step (e.g. MacKay [5], Cohn [2], Fukumizu [4], and Sugiyama and Ogawa [10]). In this paper, we focus on the global optimal case and give a new active learning method in trigonometric polynomial networks. The proposed method does not employ any approximations in its derivation, so that it provides exactly the optimal generalization capability. Moreover, the proposed method reduces the computational complexity and memory required for the calculation of learning results. Finally, the effectiveness of the proposed method is demonstrated through computer simulations. '' http://ogawa-www.cs.titech.ac?jprsugi. 625 Training Data Selection for Optimal Generalization 2 Formulation of supervised learning In this section, the supervised learning problem is formulated from the functional analytic point of view (see Ogawa [7]). Then, our learning criterion and model are described. 2.1 Supervised learning as an inverse problem Let us consider the problem of obtaining the optimal approximation to a target function f(x) of L variables from a set of M training examples. The training examples are made up of sample points Xm E V, where V is a subset of the Ldimensional Euclidean space R L, and corresponding sample values Ym E C: {(xm, Ym) 1Ym = f(xm) + nm}~=l' (1) where Ym is degraded by zero-mean additive noise n m . Let nand Y be Mdimensional vectors whose m-th elements are nm and Ym, respectively. Y is called a sample value vector. In this paper, the target function f(x) is assumed to belong to a reproducing kernel Hilbert space H (Aronszajn [1]). If H is unknown, then it can be estimated by model selection methods (e.g. Sugiyama and Ogawa [9]). Let K(?,?) be the reproducing kernel of H. If a function 'l/Jm(x) is defined as 'l/Jm (x) = K (x, x m ), then the value of f at a sample point Xm is expressed as f(x m ) = (I, 'l/Jm), where (-,.) stands for the inner product. For this reason, 'l/Jm is called a sampling function. Let A be an operator defined as M A= ~ (em0~, (2) m=l where em is the m-th vector of the so-called standard basis in C M and (. 0 7) stands for the Neumann-Schatten productl. A is called a sampling operator. Then, the relationship between f and Y can be expressed as Y = Af +n. (3) Let us denote a mapping from Y to a learning result fo by X: fo = Xy, (4) where X is called a learning operator. Then, the supervised learning problem is reformulated as an inverse problem of obtaining X providing the best approximation fa to f under a certain learning criterion. 2.2 Learning criterion and model As mentioned above, function approximation is performed on the basis of a learning criterion. Our purpose of learning is to minimize the generalization error of the learning result fa measured by Je = Enllfo - f11 2 , (5) where En denotes the ensemble average over noise. In this paper, we adopt projection learning as our learning criterion. Let A*, R(A*), and PR(AO) be the adjoint operator of A, the range of A*, and the orthogonal projection operator onto R(A*), respectively. Then, projection learning is defined as follows. IFor any fixed 9 in a Hilbert space HI and any fixed f in a Hilbert space H2, the Neumann-Schatten product (f ? g) is an operator from HI to H2 defined by using any hE HI as (f?g)h = (h,g)f? 626 M Sugiyama and H. Ogawa Definition 1 (Projection learning) (Ogawa !6j) An operator X is called the projection learning operator if X minimizes the functional J p [X] = En II X nll 2 under the constraint XA = Pn(A*). It is well-known that Eq.(5) can be decomposed into the bias and variance: JG = IIPn(A*)f - fl12+ En11Xn112. (6) Eq.(6) implies that the projection learning criterion reduces the bias to a certain level and minimizes the variance. Let us consider the following function space. Definition 2 (Trigonometric polynomial space) Let x = (e(I),e(2), .. ?,e(L))T. For 1 :S l :S L, let Nl be a positive integer and Vl = [-7r,7r]. Then, a function space H is called a trigonometric polynomial space of order (N1 , N 2 , ... , N L) if H is spanned by (7) (8) The dimension J.l of a trigonometric polynomial space of order (N1 , N 2 , ... , N L) is J.l = (2Nl + 1), and the reproducing kernel of this space is expressed as nf=1 L K(x, x') = II Kl(e(l), e(l)I), (9) l=1 where K, 3 (~{l) ~('}') = { , if e(l) if -=1= e(l)' , e(l) = e(l)I. (10) Active learning in trigonometric polynomial space The problem of active learning is to find a set {Xm}~=1 of sample points providing the optimal generalization capability. In this section, we give the optimal solution to the active learning problem in the trigonometric polynomial space. Let At be the Moore-Penrose generalized inverse 2 of A. Then, the following proposition holds. Proposition 1 If the noise covariance matrix Q is given as Q = a 2 I with a 2 > 0, then the projection learning operator X is expressed as X = At. Note that the sampling operator A is uniquely determined by {Xm}~=1 (see Eq.(2)). From Eq.(6), the bias of a learning result fo becomes zero for all f in H if and only if N(A) = {O}, where NO stands for the null space of an operator. For this reason, 2 An operator X is called the Moore-Penrose generalized inverse of an operator A if X satisfies AXA = A, XAX = X, (AX)'" = AX, and (XA)" = XA . 627 Training Data Selection for Optimal Generalization H Figure 1: Mechanism of noise suppression by Theorem 1. If a set {xm}~= l of sample points satisfies A* A = MI, then XAf = f, IIXntll = JMlln111, and Xn2 = o. we consider the case where a set {Xm}~=l of sample points satisfies N(A) = {o}. In this case, Eq.(6) is reduced to (11) which is equivalent to the noise variance in H. Consequently, the problem of active learning becomes the problem of finding a set {Xm }~= 1 of sample points minimizing Eq.(l1) under the constraint N(A) = {a} . First, we derive a condition for optimal generalization in terms of the sampling operator A. Theorem 1 Assume that the noise covariance matrix Q is given as Q = (721 with > o. Then, Je in Eq.{11) is minimized under the constraint N(A) = {O} if and only if (12) A*A =MI, where I denotes the identity operator on H. In this case, the minimum value of Je is (72J.L/M, where J.L is the dimension of H . (72 Eq.(12) implies that {:;k1jJm}~=l forms a pseudo orthonormal basis (Ogawa [8]) in H, which is an extension of orthonormal bases. The following lemma gives interpretation of Theorem 1. Lemma 1 When a set {Xm}~=l of sample points satisfies Eq.(12}, it holds that for all f E H, (13) XAf f IIAfl1 IIXul1 rullfll for all f E H, { *llull for u E 'R.(A), o (14) (15) for u E'R.(A).l. Eqs.(14) and (15) imply that k A becomes an isometry and VMX becomes a partial isometry with the initial space 'R.(A) , respectively. Let us decompose the noise n as n = nl + n2, where nl E 'R.(A) and n2 E 'R.(A).l. Then, the sample value vector y is rewritten as y = Af + nl + n2. It follows from Eq.(13) that the signal component Af is transformed into the original function f by X. From Eq.(15) , X suppresses the magnitude of noise nl in 'R.(A) by and completely removes the k 628 M Sugiyama and H. Ogawa -71" ? ? C .----. 211" M . -71" 71" ? C (a) Theorem 2 ? ? (b) Theorem 3 Figure 2: Two examples of sample points such that Condition (12) holds (1-? and M = 6). =3 noise n2 in R(A).l. This analysis is summarized in Fig.1. Note that Theorem 1 and its interpretation are valid for all Hilbert spaces such that K(x, x) is a constant for any x. In Theorem 1, we have given a necessary and sufficient condition to minimize Ja in terms of the sampling operator A. Now we give two examples of sample points {X m };;:[=l such that Condition (12) holds. From here on, we focus on the case when the dimension L of the input x is 1 for simplicity. However, the following results can be easily scaled to the case when L > 1. Theorem 2 Let M 2: 1-?, where 1-? is the dimension of H. Let c be an arbitrary constant such that -71" < c :::; -71" + ~. If a set {X m };;:[=l of sample points is determined as (16) then Eq.(12) holds. Theorem 3 Let M = kl-? where k is a positive integer. Let c be an arbitrary constant such that -71" :::; c :::; -71" + ~. If a set {X m };;:[=l of sample points is determined as 271" (17) Xm =c+-r, where r = m - 1 (mod 1-?), 1-? then Eq. (12) holds. Theorem 2 means that M sample points are fixed to 271"1M intervals in the domain [-71",71"] and sample values are gathered once at each point (see Fig.2 (a? . In contrast, Theorem 3 means that 1-? sample points are fixed to 271"11-? intervals in the domain and sample values are gathered k times at each point (see Fig.2 (b?. Now, we discuss calculation methods ofthe projection learning result fo(x). Let h m be the m-th column vector of the M-dimensional matrix (AA*)t. Then, for general sample points, the projection learning result fo(x) can be calculated as M (y, hm )1/Jm(x). fo(x) = L (18) m=l When we use the optimal sample points satisfying Condition (12), the following theorems hold. Theorem 4 When Eq.(12) holds, the projection learning result fo(x) can be calculated as 1 M (19) fo(x) = M LYm1/Jm(X). m=l Training Data Selection for Optimal Generalization 629 Theorem 5 When sample points are determined following Theorem 3, the projection learning result fo (x) can be calculated as 1 I-' 1 J.I. fo(x) = - LYp'I/Jp(x), p=l where YP = k k LYp+J.I.(q-l). (20) q=l In Eq.(18), the coefficient of 'l/Jm(x) is obtained by the inner product (y, h m). In contrast, it is replaced with Ym/M in Eq.(19) , which implies that the Moore-Penrose generalized inverse of AA* is not required for calculating fo(x). This property is quite useful when the number M of training examples is very large since the calculation of the Moore-Penrose generalized inverse of high dimensional matrices is sometimes unstable. In Eq.(20), the number of basis functions is reduced to I-' and the coefficient of 'l/Jp(x) is obtained by Yp/I-', where YP is the mean sample values at xp. For general sample points, the computational complexity and memory required for calculating fo(x) by Eq.(18) are both O(M2). In contrast, Theorem 4 states that if a set of sample points satisfies Eq.(12) , then both the computational complexity and memory are reduced to O(M). Hence, Theorem 1 and Theorem 4 do not only provide the optimal generalization but also reduce the computational complexity and memory. Moreover, if we determine sample points following Theorem 3 and calculate the learning result fo(x) by Theorem 5, then the computational complexity and memory are reduced to 0 (1-'). This is extremely efficient since I-' does not depend on the number M of training examples. The above results are shown in Tab.1. 4 Simulations In this section, the effectiveness of the proposed active learning method is demonstrated through computer simulations. Let H be a trigonometric polynomial space of order 100, and the noise covariance matrix Q be Q = I . Let us consider the following three sampling schemes. (A) Optimal sampling: Training examples are gathered following Theorem 3. (B) Experimental design: Eq.(2) in Cohn [2] is adopted as the active learning criterion. The value of this criterion is evaluated by 30 reference points. The sampling location is determined by multi-point-search with 3 candidates. (C) Passive learning: Training examples are given unilaterally. Fig.3 shows the relation between the number of training examples and the generalization error. The horizontal and vertical axes display the number of training examples and the generalization error Je measured by Eq.(5), respectively. The solid line shows the sampling scheme (A). The dashed and dotted lines denote the averages of 10 trials of the sampling schemes (B) and (C), respectively. When the number of training examples is 201, the generalization error of the sampling scheme (A) is 1 while the generalization errors of the sampling schemes (B) and (C) are 3.18 x 104 and 8.75 x 104 , respectively. This graph illustrates that the proposed sampling scheme gives much better generalization capability than the sampling schemes (B) and (C) especially when the number of training examples is not so large. 5 Conclusion We proposed a new active learning method in the trigonometric polynomial space. The proposed method provides exactly the optimal generalization capability and M Sugiyama and H Ogawa 630 Table 1: Computational complexity and memory required for projection learning. Calculation methods Eq.(18) Computational Complexity and Memory O(M2) 1 Or----,---:---,.-~---r---____." (jj c:: ,gtil ?M O(M) Theorem S? O(J.L) = kJ.L where J.L is the dimension of Hand k is a positive integer. I I I 6 I Cii (jj c:: CI) I .' I ..... I I 4 " Ol CI) I- , I .~ .?: Theorem 4 Optimal sampling Experimental design Passive learning ... 8 E 2 ' .. , . ..... '.;. .... ' "- ........ ........... - .... - OL-----~------~-----~-----~ 300 400 500 600 The number of training examples Figure 3: Relation between the number of training examples and the generalization error. at the same time, it reduces the computational complexity and memory required for the calculation of learning results. The mechanism of achieving the optimal generalization was clarified from the functional analytic point of view. References [1] N. Aronszajn. Theory of reproducing kernels. Transactions on American Mathematical Society, 68:337-404, 1950. [2] D. Cohn, Neural network exploration using optimal experiment design. In J. Cowan et al. (Eds.), Advances in Neural Information Processing Systems 6, pp. 679-686. Morgan-Kaufmann Publishers Inc., San Mateo, CA, 1994. [3J V. V . Fedorov. Theory of Optimal Experiments. Academic Press, New York, 1972. [4] K. Fukumizu. Active learning in multilayer perceptrons. In D. Touretzky et al. (Eds.), Advances in Neural Information Processing Systems 8, pp. 295-301. The MIT Press, Cambridge, 1996. [5] D. MacKay. Information-based objective functions for active data selection. Neural Computation, 4(4):590-604, 1992. [6] H. Ogawa, Projection filter regularization of ill-conditioned problem. In Proceedings of SPIE, 808, Inverse Problems in Optics, pp. 189-196, 1987. [7] H, Ogawa. Neural network learning, generalization and over-learning. In Proceedings of the ICIIPS'92, International Conference on Intelligent Information Processing fj System, vol. 2, pp. 1-6, Beijing, China, 1992. [8] H. Ogawa. Theory of pseudo biorthogonal bases and its application. In Research Institute for Mathematical Science, RIMS Kokyuroku, 1067, Reproducing Kernels and their Applications, pp. 24-38, 1998. [9] M. Sugiyama and H. Ogawa. Functional analytic approach to model selectionSubspace information criterion. In Proceedings of 1999 Workshop on InformationBased Induction Sciences (IBIS'99), pp. 93-98, Syuzenji, Shizuoka, Japan, 1999 (Its complete version is available at ftp://ftp.cs.titech.ac.jp/pub/TR/99/TR990009.ps.gz). [10] M. Sugiyama and H. Ogawa. Incremental active learning in consideration of bias, Technical Report of IEICE, NC99-56, pp. 15-22, 1999 (Its complete version is available at ftp://ftp.cs.titech.ac.jp/pub/TR/99/TR99-001O.ps.gz).
1671 |@word especially:1 trial:1 c:4 version:2 implies:3 polynomial:10 society:1 hence:1 regularization:1 lyp:2 objective:1 tokyo:2 moore:4 filter:1 simulation:4 fa:2 exploration:1 covariance:3 uniquely:1 tr:2 solid:1 ja:1 schatten:2 initial:1 ao:1 generalization:22 criterion:9 pub:2 decompose:1 proposition:2 generalized:4 demonstrate:1 okayama:1 complete:2 unstable:1 extension:1 l1:1 clarify:1 hold:8 passive:2 fj:1 reason:2 induction:1 consideration:1 relationship:1 providing:2 minimizing:1 mapping:1 additive:1 functional:5 analytic:4 adopt:1 remove:1 designed:1 purpose:1 jp:5 belong:1 he:1 interpretation:2 greedy:1 design:3 unknown:1 vertical:1 fedorov:2 cambridge:1 fukumizu:2 provides:2 mit:1 sugiyama:8 clarified:1 location:1 jg:1 reproducing:5 pn:1 arbitrary:2 mathematical:2 base:2 fl12:1 isometry:2 ax:3 focus:2 required:6 llull:1 kl:2 certain:2 axa:1 contrast:3 suppression:1 f11:1 multi:1 ol:2 biorthogonal:1 morgan:1 vl:1 decomposed:1 minimum:1 nand:1 cii:1 xm:11 jm:7 relation:2 determine:1 transformed:1 becomes:4 signal:1 ii:2 underlying:1 moreover:2 xn2:1 dashed:1 ill:1 null:1 reduces:4 memory:9 technical:1 minimizes:2 academic:1 suppresses:1 mackay:2 calculation:6 af:3 finding:1 once:1 scheme:7 technology:1 sampling:15 imply:1 pseudo:2 masashi:1 nf:1 kokyuroku:1 multilayer:1 hm:1 titech:3 exactly:2 minimized:1 scaled:1 report:1 intelligent:1 kernel:5 employ:1 sometimes:1 kj:1 positive:3 interval:2 replaced:1 publisher:1 analyzing:1 n1:2 h2:2 cowan:1 mateo:1 china:1 mod:1 effectiveness:3 sufficient:2 integer:3 xp:1 nl:6 range:1 partial:1 necessary:2 xy:1 inner:2 reduce:1 bias:4 orthogonal:1 institute:2 euclidean:1 dimension:5 projection:13 calculated:3 mdimensional:1 stand:3 column:1 valid:1 made:1 reformulated:1 onto:1 york:1 selection:6 operator:16 jj:2 san:1 transaction:1 useful:1 subset:1 www:1 equivalent:1 demonstrated:2 tr99:1 ifor:1 global:2 active:15 category:1 reduced:4 simplicity:1 http:1 assumed:1 m2:2 rule:1 dotted:1 search:1 international:1 spanned:1 orthonormal:2 estimated:1 unilaterally:1 table:1 ku:1 ca:1 ym:6 vol:1 obtaining:3 target:2 nm:2 achieving:2 designing:1 domain:2 element:1 satisfying:3 american:1 noise:10 graph:1 yp:3 actively:1 japan:2 n2:4 beijing:1 inverse:7 fig:4 summarized:1 informationbased:1 referred:1 calculate:1 coefficient:2 inc:1 je:4 en:2 performed:3 view:3 mentioned:1 tab:1 candidate:1 complexity:9 hi:3 capability:6 display:1 theorem:23 depend:1 minimize:2 degraded:1 optic:1 variance:3 kaufmann:1 efficiently:1 ensemble:1 gathered:3 basis:4 completely:1 ofthe:1 constraint:3 easily:1 workshop:1 xax:1 optimality:1 extremely:1 ci:2 derivation:1 magnitude:1 illustrates:1 conditioned:1 department:1 classified:1 fo:13 touretzky:1 ed:2 whose:1 quite:1 definition:2 em:1 penrose:4 xaf:2 pp:7 sugi:1 expressed:4 gz:2 ibis:1 mi:2 spie:1 pr:1 aa:2 satisfies:5 nll:1 discus:2 identity:1 hilbert:4 mechanism:3 formulated:2 product:3 rim:1 consequently:1 adopted:1 available:2 supervised:5 rewritten:1 trigonometric:10 determined:5 meguro:1 adjoint:1 formulation:1 evaluated:1 lemma:2 called:8 xa:3 experimental:2 ogawa:15 hand:1 p:2 neumann:2 horizontal:1 original:1 cohn:3 incremental:1 aronszajn:2 denotes:2 perceptrons:1 ftp:4 depending:1 derive:1 ac:4 measured:2 calculating:2 ieice:1 eq:23
735
1,672
Variational Inference for Bayesian Mixtures of Factor Analysers Zoubin Ghahramani and Matthew J. Beal Gatsby Computational Neuroscience Unit University College London 17 Queen Square, London WC1N 3AR, England {zoubin,m.beal}Ggatsby.ucl.ac.uk Abstract We present an algorithm that infers the model structure of a mixture of factor analysers using an efficient and deterministic variational approximation to full Bayesian integration over model parameters. This procedure can automatically determine the optimal number of components and the local dimensionality of each component (Le. the number of factors in each factor analyser) . Alternatively it can be used to infer posterior distributions over number of components and dimensionalities. Since all parameters are integrated out the method is not prone to overfitting. Using a stochastic procedure for adding components it is possible to perform the variational optimisation incrementally and to avoid local maxima. Results show that the method works very well in practice and correctly infers the number and dimensionality of nontrivial synthetic examples. By importance sampling from the variational approximation we show how to obtain unbiased estimates of the true evidence, the exact predictive density, and the KL divergence between the variational posterior and the true posterior, not only in this model but for variational approximations in general. 1 Introduction Factor analysis (FA) is a method for modelling correlations in multidimensional data. The model assumes that each p-dimensional data vector y was generated by first linearly transforming a k < p dimensional vector of unobserved independent zero-mean unit-variance Gaussian sources, x, and then adding a p-dimensional zeromean Gaussian noise vector, n, with diagonal covariance matrix \}!: i.e. y = Ax+n. Integrating out x and n, the marginal density of y is Gaussian with zero mean and covariance AA T + \}!. The matrix A is known as the factor loading matrix. Given data with a sample covariance matrix I:, factor analysis finds the A and \}! that optimally fit I: in the maximum likelihood sense. Since k < p, a single factor analyser can be seen as a reduced parametrisation of a full-covariance Gaussian. 1 IFactor analysis and its relationship to principal components analysis (peA) and mixture models is reviewed in (10). Z. Ghahramani and M. J. Heal 450 A mixture of factor analysers (MFA) models the density for y as a weighted average of factor analyser densities s P(yjA, q,,7r) = LP(sj7r)P(yjs,AS, '11), (1) s=1 where 7r is the vector of mixing proportions, s is a discrete indicator variable, and AS is the factor loading matrix for factor analyser s which includes a mean vector for y. By exploiting the factor analysis parameterisation of covariance matrices, a mixture of factor analysers can be used to fit a mixture of Gaussians to correlated high dimensional data without requiring O(P2) parameters or undesirable compromises such as axis-aligned covariance matrices. In an MFA each Gaussian cluster has intrinsic dimensionality k (or ks if the dimensions are allowed to vary across clusters). Consequently, the mixture of factor analysers simultaneously addresses the problems of clustering and local dimensionality reduction. When '11 is a multiple of the identity the model becomes a mixture of probabilistic PCAs. Tractable maximum likelihood procedure for fitting MFA and MPCA models can be derived from the Expectation Maximisation algorithm [4, 11]. The maximum likelihood (ML) approach to MFA can easily get caught in local maxima. 2 Ueda et al. [12] provide an effective deterministic procedure for avoiding local maxima by considering splitting a factor analyser in one part of space and merging two in a another part. But splits and merges have to be considered simultaneously because the number of factor analysers has to stay the same since adding a factor analyser is always expected to increase the training likelihood. A fundamental problem with maximum likelihood approaches is that they fail to take into account model complexity (Le. the cost of coding the model parameters) . So more complex models are not penalised, which leads to overfitting and the inability to determine the best model size and structure (or distributions thereof) without resorting to costly cross-validation procedures. Bayesian approaches overcome these problems by treating the parameters 0 as unknown random variables and averaging over the ensemble of models they define: P(Y) = / dO P(YjO)P(O). (2) P(Y) is the evidence for a data set Y = {yl, .. . ,yN}. Integrating out parameters penalises models with more degrees of freedom since these models can a priori model a larger range of data sets. All information inferred from the data about the parameters is captured by the posterior distribution P(OjY) rather than the ML point estimate 0. 3 While Bayesian theory deals with the problems of overfitting and model selection/averaging, in practice it is often computationally and analytically intractable to perform the required integrals. For Gaussian mixture models Markov chain Monte Carlo (MCMC) methods have been developed to approximate these integrals by sampling [8, 7]. The main criticism of MCMC methods is that they are slow and 2 Technically, the log likelihood is not bounded above if no constraints are put on the determinant of the component covariances. So the real ML objective for MFA is to find the highest finite local maximum of the likelihood. 3We sometimes use () to refer to the parameters and sometimes to all the unknown quantities (parameters and hidden variables). Formally the only difference between the two is that the number of hidden variables grows with N, whereas the number of parameters usually does not. 451 Variational Inference for Bayesian Mixtures of Factor Analysers it is usually difficult to assess convergence. Furthermore, the posterior density over parameters is stored as a set of samples, which can be inefficient. Another approach to Bayesian integration for Gaussian mixtures [9] is the Laplace approximation which makes a local Gaussian approximation around a maximum a posteriori parameter estimate. These approximations are based on large data limits and can be poor, particularly for small data sets (for which, in principle, the advantages of Bayesian integration over ML are largest). Local Gaussian approximations are also poorly suited to bounded or positive parameters such as the mixing proportions of the mixture model. Finally, it is difficult to see how this approach can be applied to online incremental changes to model structure. In this paper we employ a third approach to Bayesian inference: variational approximation. We form a lower bound on the log evidence using Jensen's inequality: 1: == In P(Y) = In / dO P(Y, 0) ~/ dO Q(O) In P6~~~) == F, (3) which we seek to maximise. Maximising F is equivalent to minimising the KLdivergence between Q(O) and P(OIY), so a tractable Q can be used as an approximation to the intractable posterior. This approach draws its roots from one way of deriving mean field approximations in physics, and has been used recently for Bayesian inference [13, 5, 1]. The variational method has several advantages over MCMC and Laplace approximations. Unlike MCMC, convergence can be assessed easily by monitoring F. The approximate posterior is encoded efficiently in Q(O) . Unlike Laplace approximations, the form of Q can be tailored to each parameter (in fact the optimal form of Q for each parameter falls out of the optimisation), the approximation is global, and Q optimises an objective function. Variational methods are generally fast, F is guaranteed to increase monotonically and transparently incorporates model complexity. To our knowledge, no one has done a full Bayesian analysis of mixtures of factor analysers. Of course, vis-a-vis MCMC, the main disadvantage of variational approximations is that they are not guaranteed to find the exact posterior in the limit. However, with a straightforward application of sampling, it is possible to take the result of the variational optimisation and use it to sample from the exact posterior and exact predictive density. This is described in section 5. In the remainder of this paper we first describe the mixture of factor analysers in more detail (section 2). We then derive the variational approximation (section 3). We show empirically that the model can infer both the number of components and their intrinsic dimensionalities, and is not prone to overfitting (section 6). Finally, we conclude in section 7. 2 The Model Starting from (1), the evidence for the Bayesian MFA is obtained by averaging the likelihood under priors for the parameters (which have their own hyperparameters): P(Y) / d7rP(7rIa:) / dvP(vla,b) / dA P(Alv), g[.t, P(s?I1r) J dx?P(xn)p(ynlx?,sn,A', q;)]. (4) Z. Ghahramani and M. J. Beal 452 Here {a, a, b, "Ill} are hyperparameters 4 , v are precision parameters (Le. inverse variances) for the columns of A. The conditional independence relations between the variables in this model are shown graphically in the usual belief network representation in Figure 1. While arbitrary choices could be made for the priors on the first line of (4), choosing priors that are conjugate to the likelihood terms on the second line of (4) greatly simplifies inference and interpretability.5 So we choose P(7rJa) to be symmetric Dirichlet, which is conjugate to the multinomial P(sJ7r). The prior for the factor loading matrix plays a key role in this model. Each component of the mixture has a Gaussian prior P(ABJV B), where each element of the vector VB is the precision of a column of A. IT one of these precisions vi -t 00, :''!:~",~.................. 1 then the outgoing weights for factor Xl will go to zero, which allows the model to reduce the inFigure 1: Generative model for trinsic dimensionality of X if the data does not variational Bayesian mixture of factor analysers. Circles denote warrant this added dimension. This method of random variables, solid rectangles intrinsic dimensionality reduction has been used denote hyperparameters, and the by Bishop [2] for Bayesian peA, and is closely dashed rectangle shows the plate related to MacKay and Neal's method for automatic relevance determination (ARD) for inputs (i.e. repetitions) over the data. to a neural network [6]. To avoid overfitting it is important to integrate out all parameters whose cardinality scales with model complexity (Le. number of components and their dimensionalities). We therefore also integrate out the precisions using Gamma priors, P(vJa, b). 3 The Variational Approximation Applying Jensen's inequality repeatedly to the log evidence (4) we lower bound it using the following factorisation of the distribution of parameters and hidden variables: Q(A)Q(7r, v)Q(s, x). Given this factorisation several additional factorisations fallout of the conditional independencies in the model resulting in the variational objective function: F= jd-n;Q(-n;) In + t, .t, PJ7;~) + Q(s") t, j dv'Q(v') lIn P6;~~) b) + jdA'Q(A') In P6~~~') 1 [j d-n; Q(-n;) Pci~:~~) + j In + jdABQ(A B) j dx"Q(x"Js") In Q~~:~") dxnQ(xnJsn)lnp(ynJxn,sn,AB, "Ill)] The variational posteriors Q('), as given in the Appendix, are derived by performing a free-form extremisation of F w.r.t. Q. It is not difficult to show that these extrema are indeed maxima of F. The optimal posteriors Q are of the same conjugate forms as the priors. The model hyperparameters which govern the priors can be estimated in the same fashion (see the Appendix). 4We currently do not integrate out 1lJ', although this can also be done. 5Conjugate priors have the same effect as pseudo-observations. (5) Variational lriference for Bayesian Mixtures of Factor Analysers 4 453 Birth and Death When optimising F , occasionally one finds that for some s: Ln Q(sn) = O. These zero responsibility components are the result of there being insufficient support from the local data to overcome the dimensional complexity prior on the factor loading matrices. So components of the mixture die of natural causes when they are no longer needed. Removing these redundant components increases F . Component birth does not happen spontaneously, so we introduce a heuristic. Whenever F has stabilised we pick a parent-component stochastically with probability proportional to e- f3F? and attempt to split it into two; Fa is the s-specific contribution to F with the last bracketed term in (5) normalised by Ln Q(sn). This works better than both cycling through components and picking them at random as it concentrates attempted births on components that are faring poorly. The parameter distributions of the two Gaussians created from the split are initialised by partitioning the responsibilities for the data, Q(sn), along a direction sampled from the parent's distribution. This usually causes F to decrease, so by monitoring the future progress of F we can reject this attempted birth if F does not recover. Although it is perfectly possible to start the model with many components and let them die, it is computationally more efficient to start with one component and allow it to spawn more when necessary. 5 Exact Predictive Density, True Evidence, and KL By importance sampling from the variational approximation we can obtain unbiased estimates of three important quantities: the exact predictive density, the true log evidence [" and the KL divergence between the variational posterior and the true posterior. Letting 0 = {A, 7r}, we sample Oi '" Q (0). Each such sample is an instance of a mixture of factor analysers with predictive density given by (1). We weight these predictive densities by the importance weights Wi = P(Oi, Y)/Q(Oi), which are easy to evaluate. This results in a mixture of mixtures of factor analysers, and will converge to the exact predictive density, P(ylY), as long as Q(O) > 0 wherever P(OIY) > O. The true log evidence can be similarly estimated by [, = In(w), where (.) denotes averaging over the importance samples. Finally, the KL divergence is given by: KL(Q(O)IIP(OIY)) = In(w) - (In w). This procedure has three significant properties. First, the same importance weights can be used to estimate all three quantities. Second, while importance sampling can work very poorly in high dimensions for ad hoc proposal distributions, here the variational optimisation is used in a principled manner to pick Q to be a good approximation to P and therefore hopefully a good proposal distribution. Third, this procedure can be applied to any variational approximation. A detailed exposition can be found in [3]. 6 Results Experiment 1: Discovering the number of components. We tested the model on synthetic data generated from a mixture of 18 Gaussians with 50 points per cluster (Figure 2, top left). The variational algorithm has little difficulty finding the correct number of components and the birth heuristics are successful at avoiding local maxima. After finding the 18 Gaussians repeated splits are attempted and rejected. Finding a distribution over number of components using F is also simple. Experiment 2: The shrinking spiral. We used the dataset of 800 data points from a shrinking spiral from [12] as another test of how well the algorithm could 454 Z. Ghahramani and M. J. Beal Figure 2: (top) Exp 1: The frames from left to right are the data, and the 2 S.D. Gaussian ellipses after 7, 14, 16 and 22 accepted births. (bottom) Exp 2: Shrinking spiral data and 1 S.D. Gaussian ellipses after 6, 9, 12, and 17 accepted births. Note that the number of Gaussians increases from left to right. number intrinsic dlmensionalnies of points 7 per cluster 8 8 16 32 64 128 -7600 - 76OOQ 500 1000 1500 I I 1 1 1 1 1 3 4 2 2 2 2 2 2 2 2 2 2 2 4 I 6 7 7 3 4 4 3 3 3 2000 Figure 3: (left) Exp 2: :F as function of iteration for the spiral problem on a typical run. Drops in :F constitute component births. Thick lines are accepted attempts, thin lines are rejected attempts. (middle) Exp 3: Means of the factor loading matrices. These results are analogous to those given by Bishop [2] for Bayesian peA. (right) Exp 3: Table with learned number of Gaussians and dimension ali ties as training set size increases. Boxes represent model components that capture several of the clusters. escape local maxima and how robust it was to initial conditions (Figure 2, bottom). Again local maxima did not pose a problem and the algorithm always found between 12-14 Gaussians regardless of whether it was initialised with 0 or 200. These runs took about 3-4 minutes on a 500MHz Alpha EV6 processor. A plot of:F shows that most of the compute time is spent on accepted moves (Figure 3, left). Experiment 3: Discovering the local dimensionalities. We generated a synthetic data set of 300 data points in each of 6 Gaussians with intrinsic dimensionalities (7432 2 1) embedded in 10 dimensions. The variational Bayesian approach correctly inferred both the number of Gaussians and their intrinsic dimensionalities (Figure 3, middle). We varied the number of data points and found that as expected with fewer points the data could not provide evidence for as many components and intrinsic dimensions (Figure 3, right). 7 Discussion Search over model structures for MFAs is computationally intractable if each factor analyser is allowed to have different intrinsic dimensionalities . In this paper we have shown that the variational Bayesian approach can be used to efficiently infer this model structure while avoiding overfitting and other deficiencies of ML approaches. One attraction of our variational method, which can be exploited in other models, is that once a factorisation of Q is assumed all inference is automatic and exact. We can also use :F to get a distribution over structures if desired. Finally we derive 455 Variational Inference for Bayesian Mixtures ofFactor Analysers a generally applicable importance sampler that gives us unbiased estimates of the true evidence, the exact predictive density, and the KL divergence between the variational posterior and the true posterior. Encouraged by the results on synthetic data, we have applied the Bayesian mixture of factor analysers to a real-world unsupervised digit classification problem. We will report the results of these experiments in a separate article. Appendix: Optimal Q Distributions and Hyperparameters Q(xnls n ) '" N(x n ,", E S ) Q(A~) ""' N(X:, Eq ,S) In Q(sn) = [1jJ(wu s ) -1jJ(w)] + xn,s=EsxsT'lT-lyn, 1 2 In IE s I + Q(vl) ""' Q(ai,bl) (In p(ynlxn, sn, AS , 'IT)) + c X:= ['IT-l"tQ(sn)ynxn 'STEq,S] , ai=a+~, bi=b+~:t(A~12) q n=l E s - 1= (A ST 'IT -1 AS) + I, Eq ,S -~ 'IT;ql Q(1r) '" D(wu) N L Q(sn)(xnxnT) +diag(vS), n=l q=l wU s = ~ + N L Q(sn) n=l where {N, Q, D} denote Normal, Gamma and Dirichlet distributions respectively, (-) denotes expectation under the variational posterior, and 1jJ (x) is the digamma function 1jJ(x) == lnr(x). Note that the optimal distributions Q(A S) have block diagonal covariance structure; even though each AS is a p x q matrix, its covariance only has O(pq2) parameters. Differentiating:F with respect to the parameters, a and b, of the precision prior we get fixed point equations 1jJ(a) = (In v)+lnb and b = a/(v). Similarly the fixed point for the parameters of the Dirichlet prior is 1jJ(a) -1jJ(a/S) + 2: [1jJ(wu s ) -1jJ(w)]/S = o. tx References [1] H. Attias. Inferring parameters and structure of latent variable models by variational Bayes. In Proc. 15th Conf. on Uncertainty in Artificial Intelligence, 1999. [2] C.M. Bishop. Variational PCA. In Proc. Ninth Int. Conf. on Artificial Neural Networks. ICANN, 1999. [3] Z. Ghahramani, H. Attias, and M.J. Beal. Learning model structure. Technical Report GCNU-TR-1999-006, (in prep.) Gatsby Unit, Univ. College London, 1999. [4] Z. Ghahramani and G.E. Hinton. The EM algorithm for mixtures of factor analyzers. Technical Report CRG-TR-96-1 [http://~.gatsby . ucl. ac. uk/ ~zoubin/papers/tr-96-1.ps.gz], Dept. of Compo Sci. , Univ. of Toronto, 1996. [5] D.J.C. MacKay. Ensemble learning for hidden Markov models. Technical report, Cavendish Laboratory, University of Cambridge, 1997. [6] R.M. Neal. Assessing relevance determination methods using DELVE. In C.M. Bishop, editor, Neural Networks and Machine Learning, 97-129. Springer-Verlag, 1998. [7] C.E. Rasmussen. The infinite gaussian mixture model. In Adv. Neur. Inf. Pmc. Sys. 12. MIT Press, 2000. [8] S. Richardson and P.J. Green. On Bayesian analysis of mixtures with an unknown number of components. J. Roy. Stat. Soc.-Ser. B, 59(4) :731-758, 1997. [9] S.J. Roberts , D. Husmeier, 1. Rezek, and W. Penny. Bayesian approaches to Gaussian mixture modeling. IEEE PAMI, 20(11):1133- 1142, 1998. [10] S. T . Roweis and Z. Ghahramani. A unifying review of linear Gaussian models. Neural Computation, 11(2):305- 345, 1999. [11] M.E. Tipping and C.M. Bishop. Mixtures of probabilistic principal component analyzers. Neural Computation, 11(2):443- 482, 1999. [12] N. Ueda, R. Nakano, Z. Ghahramani, and G.E. Hinton. SMEM algorithm for mixture models. In Adv. Neur. Inf. Proc. Sys. 11 . MIT Press, 1999. [13] S. Waterhouse, D.J.C. Mackay, and T. Robinson. Bayesian methods for mixtures of experts. In Adv. Neur. Inf. Proc. Sys. 1. MIT Press, 1995.
1672 |@word determinant:1 middle:2 loading:5 proportion:2 seek:1 yja:1 covariance:9 pick:2 tr:3 solid:1 reduction:2 initial:1 dx:2 happen:1 treating:1 drop:1 plot:1 v:1 generative:1 discovering:2 fewer:1 intelligence:1 ria:1 sys:3 compo:1 toronto:1 penalises:1 along:1 fitting:1 manner:1 introduce:1 indeed:1 expected:2 automatically:1 little:1 considering:1 cardinality:1 becomes:1 bounded:2 developed:1 unobserved:1 extremum:1 finding:3 kldivergence:1 pseudo:1 multidimensional:1 f3f:1 tie:1 uk:2 partitioning:1 unit:3 ser:1 yn:1 positive:1 maximise:1 local:13 limit:2 pami:1 k:1 delve:1 range:1 bi:1 spontaneously:1 practice:2 maximisation:1 block:1 digit:1 procedure:7 reject:1 integrating:2 lnr:1 zoubin:3 get:3 undesirable:1 selection:1 put:1 applying:1 equivalent:1 deterministic:2 straightforward:1 graphically:1 starting:1 caught:1 go:1 regardless:1 splitting:1 factorisation:4 attraction:1 deriving:1 cavendish:1 laplace:3 analogous:1 play:1 exact:9 element:1 roy:1 particularly:1 bottom:2 role:1 capture:1 adv:3 decrease:1 highest:1 principled:1 transforming:1 govern:1 complexity:4 compromise:1 predictive:8 technically:1 ali:1 easily:2 ynxn:1 tx:1 univ:2 fast:1 effective:1 london:3 monte:1 describe:1 artificial:2 analyser:22 pci:1 choosing:1 birth:8 whose:1 encoded:1 larger:1 heuristic:2 richardson:1 online:1 beal:5 hoc:1 advantage:2 ucl:2 took:1 remainder:1 aligned:1 mixing:2 poorly:3 roweis:1 exploiting:1 convergence:2 cluster:5 parent:2 p:1 assessing:1 incremental:1 vja:1 derive:2 spent:1 ac:2 stat:1 pose:1 ard:1 progress:1 eq:2 p2:1 soc:1 concentrate:1 direction:1 thick:1 closely:1 correct:1 stochastic:1 pea:3 crg:1 around:1 considered:1 normal:1 exp:5 matthew:1 vary:1 proc:4 dvp:1 applicable:1 currently:1 largest:1 repetition:1 weighted:1 mit:3 gaussian:15 always:2 rather:1 avoid:2 ax:1 derived:2 modelling:1 likelihood:9 greatly:1 digamma:1 criticism:1 sense:1 posteriori:1 inference:7 vl:1 integrated:1 lj:1 hidden:4 relation:1 classification:1 ill:2 priori:1 integration:3 mackay:3 marginal:1 field:1 optimises:1 once:1 sampling:5 encouraged:1 optimising:1 unsupervised:1 warrant:1 thin:1 future:1 report:4 escape:1 employ:1 simultaneously:2 divergence:4 gamma:2 tq:1 ab:1 freedom:1 attempt:3 mixture:31 wc1n:1 chain:1 integral:2 pq2:1 necessary:1 circle:1 desired:1 instance:1 column:2 modeling:1 ar:1 disadvantage:1 mhz:1 queen:1 cost:1 successful:1 optimally:1 stored:1 synthetic:4 st:1 density:12 fundamental:1 ie:1 stay:1 probabilistic:2 yl:1 physic:1 picking:1 parametrisation:1 again:1 iip:1 choose:1 stochastically:1 conf:2 expert:1 inefficient:1 account:1 coding:1 includes:1 int:1 mcmc:5 bracketed:1 vi:3 ad:1 root:1 responsibility:2 start:2 recover:1 bayes:1 alv:1 contribution:1 ass:1 square:1 oi:3 variance:2 efficiently:2 ensemble:2 bayesian:22 carlo:1 monitoring:2 pcas:1 processor:1 penalised:1 whenever:1 initialised:2 thereof:1 gcnu:1 sampled:1 dataset:1 knowledge:1 infers:2 dimensionality:13 stabilised:1 tipping:1 done:2 box:1 though:1 zeromean:1 furthermore:1 rejected:2 p6:3 correlation:1 hopefully:1 incrementally:1 grows:1 effect:1 requiring:1 unbiased:3 true:8 analytically:1 symmetric:1 death:1 laboratory:1 neal:2 deal:1 die:2 plate:1 variational:31 recently:1 multinomial:1 empirically:1 refer:1 significant:1 cambridge:1 ai:2 automatic:2 resorting:1 similarly:2 analyzer:2 longer:1 j:1 posterior:16 own:1 inf:3 occasionally:1 verlag:1 inequality:2 lnp:1 exploited:1 seen:1 captured:1 additional:1 husmeier:1 determine:2 converge:1 redundant:1 monotonically:1 dashed:1 full:3 multiple:1 infer:3 technical:3 england:1 determination:2 cross:1 minimising:1 lin:1 long:1 ellipsis:2 optimisation:4 expectation:2 fallout:1 iteration:1 sometimes:2 tailored:1 represent:1 proposal:2 whereas:1 source:1 unlike:2 incorporates:1 split:4 easy:1 spiral:4 independence:1 fit:2 perfectly:1 reduce:1 simplifies:1 attias:2 whether:1 pca:1 cause:2 constitute:1 repeatedly:1 jj:9 generally:2 detailed:1 reduced:1 http:1 transparently:1 neuroscience:1 estimated:2 correctly:2 per:2 discrete:1 rezek:1 key:1 independency:1 rectangle:2 offactor:1 run:2 inverse:1 uncertainty:1 ueda:2 wu:4 draw:1 appendix:3 vb:1 bound:2 jda:1 guaranteed:2 nontrivial:1 constraint:1 deficiency:1 performing:1 neur:3 poor:1 conjugate:4 across:1 em:1 wi:1 lp:1 parameterisation:1 wherever:1 heal:1 dv:1 spawn:1 computationally:3 ln:2 equation:1 fail:1 needed:1 letting:1 tractable:2 infigure:1 gaussians:9 i1r:1 jd:1 assumes:1 clustering:1 dirichlet:3 denotes:2 top:2 yjs:1 unifying:1 nakano:1 ghahramani:8 bl:1 objective:3 move:1 added:1 quantity:3 fa:2 costly:1 usual:1 diagonal:2 cycling:1 separate:1 sci:1 evaluate:1 maximising:1 relationship:1 insufficient:1 pmc:1 difficult:3 ql:1 robert:1 unknown:3 perform:2 observation:1 markov:2 finite:1 lyn:1 hinton:2 frame:1 varied:1 ninth:1 arbitrary:1 inferred:2 required:1 kl:6 merges:1 learned:1 robinson:1 address:1 usually:3 interpretability:1 green:1 belief:1 mfa:7 natural:1 difficulty:1 indicator:1 smem:1 axis:1 created:1 gz:1 sn:10 prior:12 review:1 waterhouse:1 embedded:1 proportional:1 validation:1 integrate:3 degree:1 mpca:1 article:1 principle:1 editor:1 prone:2 course:1 last:1 free:1 rasmussen:1 normalised:1 allow:1 fall:1 differentiating:1 penny:1 overcome:2 dimension:6 xn:2 world:1 made:1 approximate:2 alpha:1 ml:5 global:1 overfitting:6 conclude:1 assumed:1 alternatively:1 search:1 latent:1 reviewed:1 table:1 robust:1 complex:1 da:1 diag:1 did:1 icann:1 main:2 linearly:1 noise:1 hyperparameters:5 prep:1 allowed:2 repeated:1 fashion:1 gatsby:3 slow:1 precision:5 shrinking:3 inferring:1 xl:1 third:2 removing:1 minute:1 bishop:5 specific:1 jensen:2 evidence:10 intrinsic:8 intractable:3 adding:3 merging:1 importance:7 suited:1 lt:1 springer:1 aa:1 conditional:2 identity:1 consequently:1 exposition:1 change:1 typical:1 infinite:1 averaging:4 sampler:1 principal:2 accepted:4 attempted:3 formally:1 college:2 support:1 inability:1 assessed:1 relevance:2 dept:1 outgoing:1 tested:1 avoiding:3 correlated:1
736
1,673
Mixture Density Estimation Jonathan Q. Li Department of Statistics Yale University P.O. Box 208290 New Haven, CT 06520 Andrew R. Barron Department of Statistics Yale University P.O. Box 208290 New Haven, CT 06520 Qiang.Li@aya.yale. edu Andrew. Barron@yale. edu Abstract Gaussian mixtures (or so-called radial basis function networks) for density estimation provide a natural counterpart to sigmoidal neural networks for function fitting and approximation. In both cases, it is possible to give simple expressions for the iterative improvement of performance as components of the network are introduced one at a time. In particular, for mixture density estimation we show that a k-component mixture estimated by maximum likelihood (or by an iterative likelihood improvement that we introduce) achieves log-likelihood within order 1/k of the log-likelihood achievable by any convex combination. Consequences for approximation and estimation using Kullback-Leibler risk are also given. A Minimum Description Length principle selects the optimal number of components k that minimizes the risk bound. 1 Introduction In density estimation, Gaussian mixtures provide flexible-basis representations for densities that can be used to model heterogeneous data in high dimensions. We introduce an index of regularity Cf of density functions f with respect to mixtures of densities from a given family. Mixture models with k components are shown to achieve Kullback-Leibler approximation error bounded by c}/k for every k. Thus in a manner analogous to the treatment of sinusoidal and sigmoidal networks in Barron [1],[2], we find classes of density functions f such that reasonable size networks (not exponentially large as function of the input dimension) achieve suitable approximation and estimation error. Consider a parametric family G = {<pe(x) , x E X C Rd' : fJ E e c Rd} of probability density functions parameterized by fJ E e. Then consider the class C = CONV(G) of density functions for which there is a mixture representation of the form fp(x) = Ie <pe(x)P(dfJ) where <pe(x) are density functions from G and P is a probability measure on (1) e. The main theme of the paper is to give approximation and estimation bounds of arbitrary densities by finite mixture densities. We focus our attention on densities J. Q. Li and A. R. Barron 280 inside C first and give an approximation error bound by finite mixtures for arbitrary f E C. The approximation error is measured by Kullback-Leibler divergence between two densities, defined as DUllg) = J f(x) log[f(x)jg(x)]dx. (2) In density estimation, D is more natural to use than the L2 distance often seen in the function fitting literature. Indeed, D is invariant under scale transformations (and other 1-1 transformation of the variables) and it has an intrinsic connection with Maximum Likelihood, one of the most useful methods in the mixture density estimation. The following result quantifies the approximation error. THEOREM 1 Let G = {4>8(X) : 0 E 8} and C= CONV(G). Let f(x) J 4>8 (x)P(dO) E C. There exists fk' a k-component mixture of 4>8, such that (3) In the bound, we have 2 Cf and 'Y = 4[log(3Ve) = JJJ4>8 4>~(x)P(dO) (x)P(dO) dx, (4) + a], where a = sup log 81,82,X 4>81 (x) . 4> 82 (x) (5) Here, a characterizes an upper bound of the log ratio of the densities in G, when the parameters are restricted to 8 and the variable to X . Note that the rate of convergence, Ijk, is not related to the dimensions of 8 or X. The behavior of the constants, though, depends on the choices of G and the target f? For example we may take G to be the Gaussian location family, which we restrict to a set X which is a cube of side-length A. Likewise we restrict the parameters to be in the same cube. Then, dA 2 a<-2? (7 (6) The value of c} depends on the target density M components, then f. Suppose f is a finite mixture with In this case, a is linear in dimension. (7) C} ~ M, with equality if and only if those M components are disjoint. Indeed, suppose f(x) = E!l Pi 4>8; (x), then Pi 4>8; (x)j E!l Pi 4>8; (x) ~ 1and hence c} = J "",M L..-i=l CI;;4>8;.(X)) 4>8; (x) dx Ei=l Pt4>8; (x) ~ JI)I)4>8; M (x)dx = M. (8) i=l Genovese and Wasserman [3] deal with a similar setting. A Kullback-Leibler approximation bound of order IjVk for one-dimensional mixtures of Gaussians is given by them. In the more general case that f is not necessarily in C, we have a competitive optimality result. Our density approximation is nearly at least as good as any gp in C. 281 Mixture Density Estimation THEOREM 2 For every gp(x) = f ?o(x)P(d8), DUIlIk) ~ DUllgp) Here, 2 C/,P = J(Jf c2 + ~p 'Y. ?~(x)P(d8) ?o(x)P(d8))2 f(x)dx. (9) (10) In particular, we can take infimum over all gp E C, and still obtain a bound. Let DUIIC) = infgEc DUlIg). A theory of information projection shows that if there exists a sequence of fk such that DUllfk) -t DUIIC), then fk converges to a function 1*, which achieves DUIIC). Note that 1* is not necessarily an element in C. This is developed in Li[4] building on the work of Bell and Cover[5]. As a consequence of Theorem 2 we have (11) where c},* is the smallest limit of cJ,p for sequences of P achieving DUlIgp) that approaches the infimum DUIIC). We prove Theorem 1 by induction in the following section. An appealing feature of such an approach is that it provides an iterative estimation procedure which allows us to estimate one component at a time. This greedy procedure is shown to perform almost as well as the full-mixture procedures, while the computational task of estimating one component is considerably easier than estimating the full mixtures. Section 2 gives the iterative construction of a suitable approximation, while Section 3 shows how such mixtures may be estimated from data. Risk bounds are stated in Section 4. 2 An iterative construction of the approximation We provide an iterative construction of Ik's in the following fashion. Suppose during our discussion of approximation that f is given. We seek a k-component mixture fk close to f. Initialize h by choosing a single component from G to minimize DUllh) = DUII?o). Now suppose we have fk-l(X). Then let fk(X) = (1 - a)fk-l(X) + a?o(x) where a and 8 are chosen to minimize DUIIIk). More generally let Ik be any sequence of k-component mixtures, for k = 1,2, ... such that DUIIIk) ~ mina,o DUII(l - a)fk-l + a?o). We prove that such sequences Ik achieve the error bounds in Theorem 1 and Theorem 2. Those familiar with the iterative Hilbert space approximation results of Jones[6], Barron[l]' and Lee, Bartlett and Williamson[7], will see that we follow a similar strategy. The use of L2 distance measures for density approximation involves L2 norms of component densities that are exponentially large with dimension. Naive Taylor expansion of the Kullback-Leibler divergence leads to an L2 norm approximation (weighted by the reciprocal of the density) for which the difficulty remains (Zeevi & Meir[8], Li[9]). The challenge for us was to adapt iterative approximation to the use of Kullback-Leibler divergence in a manner that permits the constant a in the bound to involve the logarithm of the density ratio (rather than the ratio itself) to allow more manageable constants. J. Q. Li and A. R. Barron 282 The proof establishes the inductive relationship Dk ::; (1 - a)D k- 1 + 0. 2 B, where B is bounded and Dk = DUllfk). By choosing 0.1 = 1,0.2 thereafter ak = 2/k, it's easy to see by induction that Dk ::; 4B/k. (12) = 1/2 and tr To get (12), we establish a quadratic upper bound for -log -log ?1-0:)"',_1+0:?e). Three key analytic inequalities regarding to the logarithm will be handy for us, for r ~ -log(r) ::; -(r - 1) + [-log(ro) + ro - l](r _ 1)2 (ro - 1)2 (13) 2[ -log(r) + r - 1] ::; Iog r, (14) -log(r)+r-l<I/2 I -() (r _ 1)2 + og r (15) ro > 0, r-l and where log- (-) is the negative part of the logarithm. The proof of of inequality (13) is done by verifying that -lo(~(.:it!-1 is monotone decreasing in r. Inequalities (14) and (15) are shown by separately considering the cases that r < 1 and r > 1 (as well as the limit as r -+ 1). To get the inequalities one multiplies through by (r -1) or (r - 1)2, respectively, and then takes derivatives to obtain suitable monotonicity in r as one moves away from r = 1. Now apply the inequality (13) with r = (1-0:)"'_1 +o:?e 9 and ro = (1-0:)'k-1, 9 where 9 is an arbitrary density in C with 9 = J ?9P(d9). Note that r ~ ro in this case because o:t e ~ O. Plug in r = ro + a~ at the right side of (13) and expand the square. Then we get -log(r) < -(ro + a: _ 1) + a? I ( ) - - - og ro [-IO~~o~~fo -1][(ro - 1) + (ag?W +ro -1] 2 ?[-log(ro) +ro + a 2?2[-log(ro) + 0.2 -1] . g2 (ro - 1) 9 ro - 1 Now apply (14) and (15) respectively. We get a? 2?2 _ ? -log(r) ::; -log(ro) - - + a 2(1/2 + log (ro)) + a-Iog(ro). (16) 9 9 9 9 Note that in our application, ro is a ratio of densities in C. Thus we obtain an upper bound for log-(ro) involving a. Indeed we find that (1/2 + log-(ro)) ::; "1/4 where "I is as defined in the theorem. In the case that f is in C, we take 9 = f. Then taking the expectation with respect to f of both sides of (16), we acquire a quadratic upper bound for Dk, noting that tr. r = Also note that D k is a function of 9. The greedy algorithm chooses 9 to minimize Dk(9). Therefore Dk ::; mjnD k (9) ::; / D k (9)P(d9). (17) Plugging the upper bound (16) for Dk(9) into (17), we have Dk::; ( ([-log(ro)- a? 19 Ix 9 +a2?:("f/4)+a~log(ro)]J(x)dxP(d9). 9 9 (18) 283 Mixture Density Estimation where Thus TO = (1 - a)fk-1 (x)jg(x) and P is chosen to satisfy Ie ?>e(x)P(dO) = g(x). Dk ~ (1- a)Dk- 1 + a 2! ?>~(x)P(dO) (g(x))2 f(x)dx{rj4) + a log(l- a) - a -log(l- a). It can be shown that alog(l- a) - a -log(l - a) inductive relationship, ~ (19) O. Thus we have the desired (20) Therefore, Dk ~ 'Yc 2 f. In the case that f does not have a mixture representation of the form I ?>eP(dO), i.e. f is outside the convex hull C, we take Dk to be I f(x) log dx for any given gp(x) = I ?>e(x)P(dO). The above analysis then yields Dk = DUllfk) -DUllgp) ::; j:f:? 'Yc 2 f 3 as desired. That completes the proof of Theorems 1 and 2. A greedy estimation procedure The connection between the K-L divergence and the MLE helps to motivate the following estimation procedure for /k if we have data Xl, ... , Xn sampled from f. The iterative construction of fk can be turned into a sequential maximum likelihood estimation by changing min DUllfk) to max 2:~1 log fk (Xi) at each step. A surprising result is that the resulting estimator A has a log likelihood almost at least as high as log likelihood achieved by any density gp in C with a difference of order 1jk. We formally state it as n n 1=1 1=1 1 '~logfk(Xi) " ~ 1 '~IOg9p(Xi) " ~ ~ ~ - 2 cF P 'k (21) for all gp E C. Here Fn is the empirical distribution, for which c2Fn,P (ljn) 2:~=1 c~;,p where 2 I ?>Hx)P(dO) (22) Cx,P = (f ?>e(x)P(dO))2 . The proof of this result (21) follows as in the proof in the last section, except that now we take Dk = EFn loggp(X)j fk(X) to be the expectation with respect to Fn instead of with respect to the density f. Let's look at the computation at each step to see the benefits this new greedy procedure can bring for us. We have ik(X) = (1- a)ik-1(X) + a?>e(x) with 0 and a chosen to maximize n L log[(l - a)f~-l (Xi) + a?>e(Xi )] (23) i=l which is a simple two component mixture problem, with one of the two components, f~-l(X), fixed. To achieve the bound in (21), a can either be chosen by this iterative maximum likelihood or it can be held fixed at each step to equal ak (which as before is ak = 2jk for k > 2). Thus one may replace the MLE-computation of a kcomponent mixture by successive MLE-computations of two-component mixtures. The resulting estimate is guaranteed to have almost at least as high a likelihood as is achieved by any mixture density. J. Q. Li and A. R. Barron 284 A disadvantage of the greedy procedure is that it may take a number of steps to adequately downweight poor initial choices. Thus it is advisable at each step to retune the weights of convex combinations of previous components (and even perhaps to adjust the locations of these components) , in which case, the result from the previous iterations (with k - 1 components) provide natural initialization for the search at step k. The good news is that as long as for each k, given ik-l, the A is chosen among k component mixtures to achieve likelihood at least as large as the choice achieving maxol:~=llog[(l - ak)fk-l (Xi) + ak<Po(Xi )), that is, we require that n n L log f~(Xd ~ mt" L log[(l- ak)f~-l (Xi) + ak<Po(X i )), (24) i=l i=l then the conclusion (21) will follow. In particular, our likelihood results and risk bound results apply both to the case that A is taken to be global maximizer of the likelihood over k-component mixtures as well as to the case that ik is the result of the greedy procedure. 4 Risk bounds for the MLE and the iterative MLE The metric entropy of the family G is controlled to obtain the risk bound and to determine the precisions with which the coordinates of the parameter space are allowed to be represented. Specifically, the following Lipschitz condition is assumed: for (} E e c Rd and x E X C R d, d sup Ilog <PO (x) -log <Po' (x)1 ~ B L IOj - 0jl xEX j=l (25) where OJ is the j-th coordinate of the parameter vector. Note that such a condition is satisfied by a Gaussian family with x restricted to a cube with sidelength A and has a location parameter 0 that is also prescribed to be in the same cube. In particular, if we let the variance be 0'2, we may set B = 2AI 0'2. Now we can state the bound on the K-L risk of A. THEOREM 3 Assume the condition {25}. Also assume e to be a cube with sidelength A. Let ik(X) be either the maximizer of the likelihood over k -component mixtures or more generally any sequence of density estimates f~ satisfying {24}. We have A E(DUllfk)) - DUIIC) ~ 'Y 2 2 cf 2kd k .. + 'Y-:;;: log(nABe). (26) From the bound on risk, a best choice of k would be of order roughly Vn leading to a bound on ED(fllf~) - DUIIC) of order 1/Vn to within logarithmic factors. However the best such bound occurs with k = 'Ycf, .. VnI.j2dlog(nABe) which is not available when the value of cf, .. is unknown. More importantly, k should not be chosen merely to optimize an upper bound on risk, but rather to balance whatever approximation and estimation sources of error actually occur. Toward this end we optimize a penalized likelihood criterion related to the minimum description length principle, following Barron and Cover [10] . Let l(k) be a function of k that satisfies l:~l e-l(k) ::; 1, such as l(k) = 2Iog(k+ 1). Mixture Densiry Estimation 285 A penalized MLE (or MDL) procedure picks k by minimizing ! n t log i=l 1 h(Xi ) A + 2kd log (nABe) + 21(k)jn. (27) n Then we have E(DUllh)) - DUlle) A ~ 2 2 Cf 2kd m1nb k * + r-;-log(nABe) + 21(k)jn}. (28) A proof of these risk bounds is given in Li[4]. It builds on general results for maximum likelihood and penalized maximum likelihood procedures. Recently, Dasgupta [11] has established a randomized algorithm for estimating mixtures of Gaussians, in the case that data are drawn from a finite mixture of sufficiently separated Gaussian components with common covariance, that runs in time linear in the dimension and quadratic in the sample size. However, present forms of his algorithm require impractically large sample sizes to get reasonably accurate estimates of the density. It is not yet known how his techniques will work for more general mixtures. Here we see that iterative likelihood maximization provides a better relationship between accuracy, sample size and number of components. References [1] Barron, Andrew (1993) Universal Approximation Bounds for Superpositions of a Sigmoidal Function. IEEE Transactions on Information Theory 39, No.3: 930-945 [2] Barron, Andrew (1994) Approximation and Estimation Bounds for Artificial Neural Networks. Machine Learning 14: 115-133. [3] Genovese, Chris and Wasserman, Larry (1998) Rates of Convergence for the Gaussian Mixture Seive. Manuscript. [4] Li, Jonathan Q. (1999) Estimation of Mixture Models. Ph.D Dissertation. The Department of Statistics. Yale University. [5] Bell, Robert and Cover, Thomas (1988) Game-theoretic optimal portfolios. Management Science 34: 724-733. [6] Jones, Lee (1992) A simple lemma on greedy approximation in Hilbert space and convergence rates for projection pursuit regression and neural network training. Annals of Statistics 20: 608-613. [7] Lee, W.S., Bartlett, P.L. and Williamson R.C. (1996) Efficient Agnostic Learning of Neural Networks with Bounded Fan-in. IEEE Transactions on Information Theory 42, No.6: 2118-2132. [8] Zeevi, Assaf and Meir Ronny (1997) Density Estimation Through Convex Combinations of Densities: Approximation and Estimation Bounds. Neural Networks 10, No.1: 99-109. [9] Li, Jonathan Q. (1997) Iterative Estimation of Mixture Models. Ph.D. Prospectus. The Department of Statistics. Yale University. [10] Barron, Andrew and Cover, Thomas (1991) Minimum Complexity Density Estimation. IEEE Transactions on Information Theory 37: 1034-1054. [11] Dasgupta, Sanjoy (1999) Learning Mixtures of Gaussians. Pmc. IEEE Conf. on Foundations of Computer Science, 634-644.
1673 |@word manageable:1 achievable:1 norm:2 seek:1 covariance:1 pick:1 tr:2 initial:1 surprising:1 yet:1 dx:7 fn:2 analytic:1 xex:1 greedy:7 reciprocal:1 dissertation:1 provides:2 location:3 successive:1 sigmoidal:3 c2:1 ik:8 prove:2 fitting:2 assaf:1 inside:1 manner:2 introduce:2 indeed:3 roughly:1 behavior:1 decreasing:1 considering:1 conv:2 estimating:3 bounded:3 agnostic:1 minimizes:1 developed:1 ag:1 transformation:2 dfj:1 every:2 xd:1 ro:24 whatever:1 before:1 vni:1 limit:2 consequence:2 io:1 ak:7 initialization:1 handy:1 procedure:10 universal:1 empirical:1 bell:2 projection:2 radial:1 sidelength:2 get:5 close:1 risk:10 ronny:1 optimize:2 attention:1 convex:4 wasserman:2 estimator:1 importantly:1 his:2 coordinate:2 analogous:1 annals:1 construction:4 target:2 suppose:4 element:1 satisfying:1 jk:2 ep:1 verifying:1 news:1 complexity:1 motivate:1 basis:2 po:4 represented:1 separated:1 artificial:1 choosing:2 outside:1 statistic:5 gp:6 itself:1 dxp:1 sequence:5 turned:1 achieve:5 description:2 convergence:3 regularity:1 converges:1 help:1 andrew:5 advisable:1 measured:1 involves:1 rj4:1 hull:1 larry:1 require:2 hx:1 sufficiently:1 zeevi:2 achieves:2 smallest:1 a2:1 estimation:23 superposition:1 establishes:1 weighted:1 gaussian:6 rather:2 og:2 focus:1 improvement:2 likelihood:18 expand:1 selects:1 among:1 flexible:1 multiplies:1 initialize:1 cube:5 equal:1 qiang:1 jones:2 look:1 genovese:2 nearly:1 haven:2 divergence:4 ve:1 familiar:1 adjust:1 mdl:1 mixture:37 held:1 accurate:1 taylor:1 logarithm:3 desired:2 cover:4 disadvantage:1 maximization:1 considerably:1 chooses:1 density:36 randomized:1 ie:2 lee:3 d9:3 satisfied:1 management:1 d8:3 conf:1 derivative:1 leading:1 li:10 sinusoidal:1 ioj:1 satisfy:1 depends:2 characterizes:1 sup:2 competitive:1 minimize:3 square:1 accuracy:1 variance:1 likewise:1 yield:1 fo:1 ed:1 proof:6 sampled:1 treatment:1 cj:1 hilbert:2 actually:1 manuscript:1 follow:2 done:1 box:2 though:1 ei:1 maximizer:2 infimum:2 perhaps:1 building:1 counterpart:1 inductive:2 equality:1 hence:1 adequately:1 leibler:6 deal:1 during:1 game:1 criterion:1 mina:1 theoretic:1 bring:1 fj:2 recently:1 common:1 mt:1 ji:1 exponentially:2 retune:1 jl:1 ai:1 rd:3 fk:13 jg:2 portfolio:1 inequality:5 seen:1 minimum:3 determine:1 maximize:1 full:2 adapt:1 plug:1 long:1 mle:6 iog:3 plugging:1 controlled:1 involving:1 regression:1 heterogeneous:1 expectation:2 metric:1 iteration:1 achieved:2 separately:1 completes:1 source:1 noting:1 easy:1 restrict:2 nabe:4 regarding:1 expression:1 bartlett:2 useful:1 generally:2 involve:1 ph:2 meir:2 estimated:2 disjoint:1 dasgupta:2 thereafter:1 key:1 achieving:2 drawn:1 changing:1 monotone:1 merely:1 run:1 parameterized:1 family:5 reasonable:1 aya:1 almost:3 vn:2 bound:27 ct:2 guaranteed:1 yale:6 fan:1 quadratic:3 occur:1 optimality:1 min:1 prescribed:1 downweight:1 department:4 combination:3 poor:1 kd:3 appealing:1 invariant:1 restricted:2 taken:1 remains:1 end:1 available:1 gaussians:3 pursuit:1 permit:1 apply:3 barron:11 away:1 jn:2 thomas:2 cf:6 build:1 establish:1 move:1 occurs:1 parametric:1 strategy:1 distance:2 chris:1 toward:1 induction:2 length:3 index:1 relationship:3 ratio:4 balance:1 acquire:1 minimizing:1 pmc:1 robert:1 stated:1 negative:1 unknown:1 perform:1 upper:6 ilog:1 alog:1 finite:4 arbitrary:3 introduced:1 connection:2 established:1 efn:1 yc:2 fp:1 challenge:1 max:1 oj:1 suitable:3 natural:3 difficulty:1 naive:1 literature:1 l2:4 foundation:1 principle:2 pi:3 lo:1 penalized:3 last:1 side:3 allow:1 taking:1 benefit:1 dimension:6 xn:1 transaction:3 kullback:6 monotonicity:1 global:1 assumed:1 xi:9 search:1 iterative:13 quantifies:1 reasonably:1 expansion:1 williamson:2 necessarily:2 da:1 main:1 allowed:1 fashion:1 precision:1 theme:1 xl:1 pe:3 ix:1 theorem:9 dk:14 intrinsic:1 exists:2 sequential:1 ci:1 ljn:1 easier:1 entropy:1 cx:1 logarithmic:1 g2:1 satisfies:1 jf:1 replace:1 lipschitz:1 specifically:1 except:1 impractically:1 llog:1 lemma:1 called:1 sanjoy:1 ijk:1 formally:1 jonathan:3
737
1,674
Model selection in clustering by uniform convergence bounds* Joachim M. Buhmann and Marcus Held Institut flir Informatik III, RomerstraBe 164, D-53117 Bonn, Germany {jb,held}@cs.uni-bonn.de Abstract Unsupervised learning algorithms are designed to extract structure from data samples. Reliable and robust inference requires a guarantee that extracted structures are typical for the data source, Le., similar structures have to be inferred from a second sample set of the same data source. The overfitting phenomenon in maximum entropy based annealing algorithms is exemplarily studied for a class of histogram clustering models. Bernstein's inequality for large deviations is used to determine the maximally achievable approximation quality parameterized by a minimal temperature. Monte Carlo simulations support the proposed model selection criterion by finite temperature annealing. 1 Introduction Learning algorithms are designed to extract structure from data. Two classes of algorithms have been widely discussed in the literature - supervised and unsupervised learning. The distinction between the two classes depends on supervision or teacher information which is either available to the learning algorithm or missing. This paper applies statistical learning theory to the problem of unsupervised learning. In particular, error bounds as a protection against overfitting are derived for the recently developed Asymmetric Clustering Model (ACM) for co-occurrence data [6]. These theoretical results show that the continuation method "deterministic annealing" yields robustness of the learning results in the sense of statistical learning theory. The computational temperature of annealing algorithms plays the role of a control parameter which regulates the complexity of the learning machine. Let us assume that a hypothesis class 1? of loss functions h(x; a) is given. These loss functions measure the quality of structures in data. The complexity of 1? is controlled by coarsening, i.e., we define a 'Y-cover of 1?. Informally, the inference principle advocated by us performs learning by two inference steps: (i) determine the optimal approximation level l' for consistent learning (in terms of large risk deviations); (ii) given the optimal approximation level 1', average over all hypotheses in an appropriate neighborhood of the empirical minimizer. The result of the inference *This work has been supported by the German Israel Foundation for Science and Research Development (GIF) under grant #1-0403-001.06/95. Model Selection in Clustering by Uniform Convergence Bounds 217 procedure is not a single hypothesis but a set of hypotheses. This set is represented either by an average of loss functions or, alternatively, by a typical member of this set. This induction approach is named Empirical Risk Approximation (ERA) [2]. The reader should note that the learning algorithm has to return an average structure which is typical in a 'Y-cover sense but it is not supposed to return the hypothesis with minimal empirical risk as in Vapnik's "Empirical Risk Minimization" (ERM) induction principle for classification and regression [9]. The loss function with minimal empirical risk is usually a structure with maximal complexity, e.g., in clustering the ERM principle will necessarily yield a solution with the maximal number of clusters. The ERM principle, therefore, is not suitable as a model selection principle to determine the number of clusters which are stable under sample fluctuations. The ERA principle with its approximation accuracy 'Y solves this problem by controlling the effective complexity of the hypothesis class. In spirit, this approach is similar to the Gibbs-algorithm presented for example in [3]. The Gibbs-algorithm samples a random hypothesis from the version space to predict the label of the 1 + lth data point Xl+!o The version space is defined as the set of hypotheses which are consistent with the first 1 given data points. In our approach we use an alternative definition of consistency, where all hypothesis in an appropriate neighborhood of the empirical minimizer define the version space (see also [4]). Averaging over this neighborhood yields a structure with risk equivalent to the expected risk obtained by random sampling from this set of hypotheses. There exists also a tight methodological relationship to [7] and [4] where learning curves for the learning of two class classifiers are derived using techniques from statistical mechanics. 2 The Empirical Risk Approximation Principle The data samples Z = {zr E 0, 1 ~ r ~ l} which have to be analyzed by the unsupervised learning algorithm are elements of a suitable object (resp. feature) space O. The samples are distributed according to a measure J.L which is not assumed to be known for the analysis.l A mathematically precise statement of the ERA principle requires several definitions which formalize the notion of searching for structure in the data. The quality of structures extracted from the data set Z is evaluated by the empirical risk R(a; Z) := 2:~=1 h(zr; a) of a structure a given the training set Z. The function h(z; a) is known as loss function in statistics. It measures the costs for processing a generic datum z with model a. Each value a E A parameterizes an individual loss function with A denoting the set of possible parameters. The loss function which minimizes the empirical risk is denoted by &1. := arg minaEA R( a; Z). The relevant quality measure for learning is the expected risk R(a) .h(z; a) dJ.L(z). The optimal structure to be inferred from the data is a1. .argminaEA R(a). The distribution J.L is assumed to decay sufficiently fast with bounded rth moments Ell {Ih(z; a) - R(a)IT} ~ rh? r - 2 V II {h(z; an, 'Va E A (r > 2). Ell {.} and VII {.} denote expectation and variance of a random variable, respectively. T is a distribution dependent constant. ERA requires the learning algorithm to determine a set hypotheses on the basis of the finest consistently learnable cover of the hypothesis class. Given a learning accuracy 'Y a subset of parameters A-y = {al,'" ,aIA-yI-l} U {&1.} can be defined such that the hypothesis class 1i is covered by the function balls with index sets B-y(a) := {a' : Ih(z; a') - h(z; a)1 dJ.L(z) ~ 'Y}, i. e. A C UaEA-y B-y(a). The em- t In In 1 Knowledge of covering numbers is required in the following analysis which is a weaker type of information than complete knowledge of the probability measure IL (see also [5]). 218 J. M Buhmann and M Held pirical minimizer &1. has been added to the cover to simplify bounding arguments. Large deviation theory is used to determine the approximation accuracy '1 for learning a hypothesis from the hypothesis class 11.. The expected risk of the empirical minimizer exceeds the global minimum of the expected risk R(01.) by faT with a probability bounded by Bernstein's inequality [8] ~ -21 (faT - 'Y)} l(f-'Y/ aT )2) _ < P { sup IR(o) - R(o)1 aEA-y < 21A')'1 exp ( - 8 + 4r (f _ 'Y/a T) = o. (1) The complexity IA')' I of the coarsened hypothesis class has to be small enough to guarantee with high confidence small f-deviations. 2 This large deviation inequality weighs two competing effects in the learning problem, i. e. the probability of a large deviation exponentially decreases with growing sample size I, whereas a large deviation becomes increasingly likely with growing cardinality of the 'Y-cover of the hypothesis class. According to (1) the sample complexity Io (-y, f, 0) is defined by to (f - '1/ aT) 2 2 log IA')'I - 8 + 4r (f _ 'Y/aT) + log "8 = o. (2) With probability 1 - 0 the deviation of the empirical risk from the expected risk is bounded by ~ (foPta T - '1) =: 'Yapp ? Averaging over a set of functions which exceed the empirical minimizer by no more than 2'Yapp in empirical risk yields an average hypothesis corresponding to the statistically significant structure in the data, i.e., R( 01.) - R( &1.) ~ R( 01. ) + 'Yapp - (R( &1.) - 'Yapp ) ~ 2'Y app since R( 01.) ~ R( &1.) by definition. The key task in the following remains to calculate the minimal precision f( '1) as a function of the approximation '1 and to bound from above the cardinality IA')' I of the 'Y-cover for specific learning problems. 3 Asymmetric clustering model The asymmetric clustering model was developed for the analysis resp. grouping of objects characterized by co-occurrence of objects and certain feature values [6]. Application domains for this explorative data analysis approach are for example texture segmentation, statistical language modeling or document retrieval. Denote by n = X x y the product space of objects Xi EX, 1 ~ i ~ nand features Yj E y, 1 ~ j ~ j. The Xi E X are characterized by observations Z = {zr} = {(Xi(r),Yj(r)) ,T = 1, ... ,l}. The sufficient statistics of how often the object-feature pair (Xi, Yj) occurs in the data set Z is measured by the set of frequencies {'f]ij : number of observations (Xi, Yj) /total number of observations}. Derived measurements are the frequency of observi~g object Xi, i. e. 'f]i = 2:;=1 'f]ij and the frequency of observing feature Yj given object Xi, i. e. 'f]jli = 'f]ij/'f]i. The asymmetric clustering model defines a generative model of a finite mixture of component probability distributions in feature space with cluster-conditional distributions q = (qjlv) ' 1 ~ j ~ j, 1 ~ v ~ k (see [6]). We introduce indicator variables M iv E {O, 1} for the membership of object Xi in cluster v E {I, ... ,k}. 2::=1 M iv = 1 Vi : 1 ~ i ~ n enforces the uniqueness constraint for assignments. 2The maximal standard deviation (1 T := sUPaEA-y y'V {h(z; a)} defines the scale to measure deviations of the empirical risk from the expected risk (see [2]). Model Selection in Clustering by Uniform Convergence Bounds 219 Using these variables the observed data Z are distributed according to the generative model over X x y: 1 k P {xi,YjIM,q} = - ~ (3) n L--v=1 Mivqjlv' For the analysis of the unknown data source - characterized (at least approximatively) by the empirical data Z - a structure 0: = (M, q) with M E {O, I} n x k has to be inferred. The aim of an ACM analysis is to group the objects Xi as coded by the unknown indicator variables M iv and to estimate for each cluster v a prototypical feature distribution qjlv' Using the loss function h(Xi' Yj; 0:) = logn - 2:~=1 M iv logqjlv the maximization of the likelihood can be formulated as minimization of the empirical risk: R(o:; Z) = 2:~=1 2:;=11}ij h(xi, Yj; 0:), where the essential quantity to be minimized is the expected risk: R(o:) = 2:~=1 2:;=1 ptrue {Xi, Yj} h(Xi' Yj; 0:). Using the maximum entropy principle the following annealing equations are derived [6]: 2:~1 (M iv )1}ij _ ~n wi=1 (M iv ) - L--.t=1 "n A qjlv exp (Miv )1}i "n (M hv )1}j1i, wh=1 (4) [.8 2:;=1 1}jli log Q]lv ] The critical temperature: Due to the limited precision of the observed data it is natural to study histogram clustering as a learning problem with the hypothesis class 1? = {-2:vMivlogqjlv :Miv E {0,1} /\ 2:vMiv = 1/\ Qjlv E H,t, .. ? ,1}/\ 2:j qjlv = I}. The limited number of observations results in a limited precision of the frequencies 1}jli' The value Q;lv = 0 has been excluded since it causes infinite expected risk for ptrue {Yj IXi} > O. The size of the regularized hypothesis class A-y can be upper bounded by the cardinality of the complete hypothesis class divided by the minimal cardinality of a 'Y-function ball centered at a function of the 'Y-cover A-y, i. e. IA-yl ~ 11?1/!llin IB-y(&)I. oEA'T The cardinality of a function ball with radius 'Y can be approximated by adopting for x ~ 0): techniques from asymptotic analysis [1] (8 (x) = g IB-y(5)1 = L 8 L ('Y - . } M { q,lo L i ?J' ~ptrue {Yj IXi} IIOg ~~I~(i) I) (6) %Im(t) and the entropy S is given by S(q,Q,x) = 'Yx - Lv Qv (L j qjlv .!.n L--, ~ ,log ~ exp L--p -1) + (-x L-~, ptrue J {Yj IXi} IIOg _Qjlp I). (7) %Im(i) The auxiliary variables Q = {Q v } ~=1 are Lagrange parameters to enforce the normalizations 2: j qjlv = 1. Choosing %10 = qjlm(i) Vm(i) = 0:, we obtain an approximation of the integral. The reader should note that a saddlepoint approximation in J. M Buhmann and M Held 220 the usual sense is only applicable for the parameter x but will fail for the q, Q parameters since the integrand is maximal at the non-differentiability point of the absolute value function. We, therefore, expand S (q, Q,x) up to linear terms 0 (q - q) and integrate piece-wise. Using the abbreviation Kill := Lj ptrue {Yj Ixd IIog qj~:~i) I the following saddle point approximation for the integral over x is obtained: 1 I: I: , = Pij.?Kjlj.? n .t=1 j.?=1 n k ? wIth Pia ( -XKia) = Lexp (~)" j.? exp -XKij.? (8) The entropy S evaluated at q = q yields in combination with the Laplace approximation [1] an estimate for the cardinality of the ,-cover log IA')' I = n (log k - S) + -21 I:.t,p KipP ip (I: II P illKill - KiP) x2 (9) where the second term results from the second order term of the Taylor expansion around the saddle point. Inserting this complexity in equation (2) yields an equation which determines the required number of samples 10 for a fixed precision f and confidence o. This equation defines a functional relationship between the precision f and the approximation quality, for fixed sample size 10 and confidence o. Under this assumption the precision f depends on , in a non-monotone fashion, i. e. (10) using the abbreviation C = log IA')' I + log~. The minimum of the function ?(,) defines a compromise between uncertainty originating from empirical fluctuations and the loss of precision due to the approximation by a ,-cover. Differentiating with respect to , and setting the result to zero (df(T)/d, = 0) yields as upper bound for the inverse temperature: ~ 1 10 ( 10+C7"2 x <- 7" + -;:;~:;;;==~iiT (1T 2n V21 0C + 7"2C2 )-1 (11) Analogous to estimates of k-means, phase-transitions occur in ACM while lowering the temperature. The mixture model for the data at hand can be partitioned into more and more components, revealing finer and finer details of the generation process. The critical xopt defines the resolution limit below which details can not be resolved in a reliable fashion on the basis of the sample size 10 . Given the inverse temperature x the effective cardinality of the hypothesis class can be upper bounded via the solution of the fix point equation (8). On the other hand this cardinality defines with (11) and the sample size lo an upper bound on x. Iterating these two steps we finally obtain an upper bound for the critical inverse temperature given a sample size 10. Empirical Results: For the evaluation of the derived theoretical result a series of Monte-Carlo experiments on artificial data has been performed for the asymmetric clustering model. Given the number of objects n = 30, the number of groups k = 5 and the size of the histograms f = 15 the generative model for this experiments was created randomly and is summarized in fig. 1. From this generative model sample sets of arbitrary size can be generated and the true distributions ptrue {Yj IXi} can be calculated. In figure 2a,b the predicted temperatures are compared to the empirically observed critical temperatures, which have been estimated on the basis of 2000 different samples of randomly generated co-occurrence data for each 10. The expected risk (solid) Model Selection in Clustering by Uniform Convergence Bounds v 1 2 3 4 5 m(i} 221 qjlv 0.11,0.01,0.11,0.07,0.08,0.04,0.06,0,0.13,0.07, 0.08, 0.1, 0, 0.11,0.031 0.18,0.1,0.09,0.02,0.05,0.09,0.08,0.03,0.06, 0.07, 0.03, 0.02, 0.07, 0.06, 0.05} 0.17,0.05,0.05,0.06,0.06,0.05,0.03,0.11,0.09,0, 0.02,0.1,0.03,0.07, 0.11} 0.15,0.07,0.1,0.03,0.09,0.03,0.04,0.05,0.06, 0.05,0.08,0.04,0.08,0.09, 0.04} 0.09,0.09,0.07,0.1,0.07,0.06,0.06,0.11,0.07,0.07, 0.1, 0.02,0.07,0.02, O} = (5,3,2,5,2,2,5,4,2,2,2,4,1,5,3,5,3,4,1 , 2,2,3,1,1,2, 5, 5, 2, 2, 1) Figure 1: Generative ACM model for the Monte-Carlo experiments. and empirical risk (dashed) of these 2000 inferred models are averaged. Overfitting sets in when the expected risk rises as a function of the inverse temperature x. Figure 2c indicates that on average the minimal expected risk is assumed when the effective number is smaller than or equal 5, i. e. the number of clusters of the true generative model. Predicting the right computational temperature, therefore, also enables the data analyst to solve the cluster validation problem for the asymmetric clustering model. Especially for 10 = 800 the sample fluctuations do not permit the estimate of five clusters and the minimal computational temperature prevents such an inference result. On the other hand for lo = 1600 and 10 = 2000 the minimal temperature prevents the algorithm to infer too many clusters, which would be an instance of overfitting. As an interesting point one should note that for an infinite number of observations the critical inverse temperature reaches a finite positive value and not more than the five effective clusters are extracted. At this point we conclude, that for the case of histogram clustering the Empirical Risk Approximation solves for realizable rules the problem of model validation, i. e. choosing the right number of clusters. Figure 2d summarizes predictions of the critical temperature on the basis of the empirical distribution 1]ij rather than the true distribution ptrue {Xi, Yj}. The empirical distribution has been generated by a training sample set with x of eq. (11) being used as a plug-in estimator. The histogram depicts the predicted inverse temperature for 10 = 1200. The average of these plug-in estimators is equal to the predicted temperature for the true distribution. The estimates of x are biased towards too small inverse temperatures due to correlations between the parameter estimates and the stopping criterion. It is still an open question and focus of ongoing work to rigorously bound the variance of this plug- in estimator. Empirically we observe a reduction of the variance of the expected risk occurring at the predicted temperature for higher sample sizes lo . 4 Conclusions The two conditions that the empirical risk has to uniformly converge towards the expected risk and that all loss functions within an 2,&PP -range of the global empirical risk minimum have to be considered in the inference process limits the complexity of the underlying hypothesis class for a given number of samples. The maximum entropy method which has been widely employed in deterministic annealing procedures for optimization problems is substantiated by our analysis. Solutions with too many clusters clearly overfit the data and do not generalize. The condition that the hypothesis class should only be divided in function balls of size , forces us to stop the stochastic search at the lower bound of the computational temperature. Another important result of this investigation is the fact that choosing the right stopping temperature for the annealing process not only avoids overfitting but also solves the cluster validation problem in the realizable case of ACM. A possible inference of too many clusters using the empirical risk functional is suppressed. 222 J. M Buhmann and M Held 80 1-' ,/ a) ) 78 6 - - _ " ; ' ,' \ \ \ 4 / - \ \ / ! / /' 80 b) _ ... lo 0 800emP ._yo 101200lmp 78 'o'2000mp / 76 / ~/ \ '"~74 \ \ \ \ 2 \ 72 \ , ~~ ~" \ \ "- \ 0 \ 680 11 '0 , '----------------------- ,5 20 inverse temperature c) 10 -~ .~, ," 9 ,I ill 8 ~ [ i . . ..... p 26 ~ <II .;, ~ <II 30 -.--~~ ,- 35 680~---~---~ 'O---~'~5---~2~ 0 ---~25~---3O~---~35? inverse temperature .. ~ ----- --.- ..... .i 7 '0 6 ~ -- ... _----- ...... - 70 ,._-_._-- 5 4 3 2 o 5 10 15 20 inverse temperature 25 30 35 ~ Figure 2: Comparison between the theoretically derived upper bound on x and the observed critical temperatures (minimum of the expected risk vs. x curve). Depicted are the plots for 10 = 800,1200,1600,2000. Vertical lines indicate the predicted critical temperatures. The average effective number of clusters is drawn in part c. In part d the distribution of the plug- in estimates is shown for La = 1200. References [1] N. G. De Bruijn. Asymptotic Methods in Analysis. North-Holland Publishing Co., (repr. Dover), Amsterdam, 1958, (1981) . [2] J . M. Buhmann. Empirical risk approximation. Technical Report IAI-TR 98-3, Institut fur Informatik III, Universitat Bonn, 1998. [3] D. Haussler, M. Kearns, and R. Schapire. Bounds on the sample complexity of Bayesian learning using information theory and the VC dimension. Machine Learning, 14(1) :83113, 1994. [4] D. Haussler, M. Kearns , H.S. Seung, and N. Tishby. Rigorous learning curve bounds from statistical mechanics. Machine Learning, 25:195- 236, 1997. [5] D. Haussler and M. Opper. Mutual information, metric entropy and cumulative relative entropy risk. Annals of Statistics, December 1996. [6] T . Hofmann, J. Puzicha, and M.I. Jordan. Learning from dyadic data. In M. S. Kearns , S. A. Solla, and D. A. Cohn, editors, Advances in Neural Information Processing Systems 11. MIT Press, 1999. to appear. [7] H. S. Seung, H. Sompolinsky, and N. Tishby. Statistical mechanics of learning from examples. Physical Review A, 45(8):6056-6091 , April 1992. [8] A. W . van der Vaart and J. A. Wellner. Weak Convergence and Empirical Processes. Springer-Verlag, New York, Berlin, Heidelberg, 1996. [9] V. N. Vapnik. Statistical Learning Theory. Wiley- Interscience, New York, 1998.
1674 |@word version:3 achievable:1 open:1 simulation:1 tr:1 solid:1 reduction:1 moment:1 series:1 denoting:1 document:1 protection:1 v21:1 finest:1 explorative:1 hofmann:1 enables:1 designed:2 plot:1 v:1 generative:6 dover:1 five:2 c2:1 interscience:1 introduce:1 theoretically:1 expected:14 mechanic:3 growing:2 cardinality:8 becomes:1 bounded:5 underlying:1 israel:1 gif:1 minimizes:1 minaea:1 developed:2 guarantee:2 fat:2 classifier:1 control:1 grant:1 pirical:1 appear:1 positive:1 limit:2 io:1 era:4 fluctuation:3 studied:1 co:4 limited:3 range:1 statistically:1 averaged:1 enforces:1 yj:15 kipp:1 procedure:2 xopt:1 empirical:27 revealing:1 confidence:3 selection:6 risk:34 equivalent:1 deterministic:2 missing:1 resolution:1 ixd:1 rule:1 estimator:3 haussler:3 searching:1 notion:1 jli:3 laplace:1 analogous:1 resp:2 annals:1 play:1 controlling:1 ixi:4 hypothesis:24 element:1 approximated:1 asymmetric:6 coarsened:1 role:1 observed:4 hv:1 calculate:1 sompolinsky:1 solla:1 decrease:1 complexity:9 seung:2 rigorously:1 tight:1 compromise:1 basis:4 resolved:1 iit:1 represented:1 substantiated:1 fast:1 effective:5 monte:3 artificial:1 neighborhood:3 choosing:3 widely:2 solve:1 statistic:3 vaart:1 ip:1 maximal:4 product:1 inserting:1 relevant:1 supposed:1 convergence:5 cluster:15 object:10 measured:1 ij:6 advocated:1 eq:1 solves:3 auxiliary:1 c:1 predicted:5 indicate:1 radius:1 stochastic:1 vc:1 centered:1 fix:1 investigation:1 mathematically:1 im:2 sufficiently:1 around:1 considered:1 exp:4 predict:1 uniqueness:1 applicable:1 label:1 qv:1 minimization:2 mit:1 clearly:1 aim:1 rather:1 derived:6 focus:1 yo:1 joachim:1 methodological:1 consistently:1 likelihood:1 indicates:1 fur:1 rigorous:1 sense:3 realizable:2 inference:7 dependent:1 stopping:2 membership:1 lj:1 nand:1 expand:1 originating:1 germany:1 arg:1 classification:1 ill:1 denoted:1 logn:1 development:1 ell:2 mutual:1 equal:2 sampling:1 unsupervised:4 minimized:1 jb:1 report:1 simplify:1 randomly:2 individual:1 phase:1 evaluation:1 analyzed:1 mixture:2 held:5 integral:2 institut:2 iv:6 taylor:1 weighs:1 theoretical:2 minimal:8 instance:1 modeling:1 cover:9 assignment:1 maximization:1 cost:1 pia:1 deviation:10 subset:1 uniform:4 too:4 tishby:2 universitat:1 teacher:1 yl:1 vm:1 return:2 de:2 summarized:1 north:1 mp:1 depends:2 vi:1 piece:1 performed:1 observing:1 sup:1 il:1 ir:1 accuracy:3 variance:3 yield:7 flir:1 generalize:1 bayesian:1 weak:1 informatik:2 carlo:3 finer:2 app:1 reach:1 definition:3 against:1 c7:1 frequency:4 pp:1 stop:1 wh:1 knowledge:2 segmentation:1 formalize:1 higher:1 supervised:1 maximally:1 iai:1 april:1 evaluated:2 correlation:1 overfit:1 hand:3 cohn:1 defines:6 quality:5 effect:1 true:4 excluded:1 covering:1 criterion:2 complete:2 performs:1 temperature:27 wise:1 recently:1 functional:2 empirically:2 regulates:1 physical:1 exponentially:1 discussed:1 rth:1 significant:1 measurement:1 gibbs:2 consistency:1 repr:1 language:1 dj:2 stable:1 supervision:1 verlag:1 certain:1 inequality:3 yi:1 der:1 minimum:4 employed:1 converge:1 determine:5 dashed:1 ii:5 infer:1 exceeds:1 technical:1 characterized:3 plug:4 bruijn:1 retrieval:1 divided:2 coded:1 a1:1 controlled:1 va:1 prediction:1 regression:1 expectation:1 df:1 metric:1 histogram:5 normalization:1 adopting:1 whereas:1 annealing:7 source:3 biased:1 member:1 december:1 spirit:1 coarsening:1 jordan:1 bernstein:2 iii:2 enough:1 exceed:1 competing:1 parameterizes:1 qj:1 wellner:1 aea:1 york:2 cause:1 iterating:1 covered:1 informally:1 differentiability:1 continuation:1 schapire:1 estimated:1 kill:1 group:2 key:1 drawn:1 lowering:1 monotone:1 inverse:10 parameterized:1 uncertainty:1 named:1 reader:2 summarizes:1 bound:14 datum:1 occur:1 constraint:1 x2:1 bonn:3 integrand:1 llin:1 argument:1 according:3 ball:4 combination:1 smaller:1 em:1 increasingly:1 suppressed:1 wi:1 partitioned:1 saddlepoint:1 erm:3 equation:5 remains:1 lexp:1 german:1 fail:1 available:1 permit:1 observe:1 appropriate:2 generic:1 enforce:1 occurrence:3 alternative:1 robustness:1 aia:1 clustering:14 publishing:1 yx:1 especially:1 added:1 quantity:1 occurs:1 question:1 usual:1 berlin:1 induction:2 marcus:1 analyst:1 index:1 relationship:2 statement:1 rise:1 unknown:2 upper:6 vertical:1 observation:5 finite:3 precise:1 arbitrary:1 inferred:4 pair:1 required:2 kip:1 distinction:1 lmp:1 usually:1 below:1 reliable:2 ia:6 suitable:2 critical:8 natural:1 force:1 regularized:1 predicting:1 buhmann:5 zr:3 indicator:2 miv:2 created:1 extract:2 review:1 literature:1 asymptotic:2 exemplarily:1 relative:1 loss:10 prototypical:1 generation:1 interesting:1 lv:3 validation:3 foundation:1 integrate:1 sufficient:1 consistent:2 pij:1 principle:9 editor:1 lo:5 supported:1 weaker:1 emp:1 differentiating:1 absolute:1 distributed:2 van:1 curve:3 calculated:1 dimension:1 transition:1 avoids:1 opper:1 cumulative:1 uni:1 global:2 overfitting:5 assumed:3 conclude:1 xi:15 alternatively:1 search:1 robust:1 heidelberg:1 expansion:1 necessarily:1 domain:1 rh:1 bounding:1 dyadic:1 fig:1 depicts:1 fashion:2 wiley:1 precision:7 xl:1 ib:2 specific:1 learnable:1 decay:1 grouping:1 exists:1 essential:1 ih:2 vapnik:2 texture:1 occurring:1 vii:1 entropy:7 depicted:1 likely:1 saddle:2 lagrange:1 prevents:2 amsterdam:1 approximatively:1 applies:1 ptrue:7 holland:1 springer:1 minimizer:5 determines:1 extracted:3 acm:5 conditional:1 lth:1 abbreviation:2 formulated:1 towards:2 typical:3 infinite:2 uniformly:1 averaging:2 kearns:3 total:1 la:1 puzicha:1 support:1 romerstrabe:1 ongoing:1 phenomenon:1 ex:1
738
1,675
Image Recognition in Context: Application to Microscopic Urinalysis XuboSong* Department of Electrical and Computer Engineering Oregon Graduate Institute of Science and Technology Beaverton, OR 97006 xubosong@ece.ogi.edu Joseph Sill Department of Computation and Neural Systems California Institute of Technology Pasadena, CA 91125 joe@busy.work.caltech.edu Yaser Abu-Mostafa Department of Electrical Engineering California Institute of Technology Pasadena, CA 91125 yase r@work.caltech.edu Harvey Kasdan International Remote Imaging Systems, Inc. Chatsworth, CA 91311 Abstract We propose a new and efficient technique for incorporating contextual information into object classification. Most of the current techniques face the problem of exponential computation cost. In this paper, we propose a new general framework that incorporates partial context at a linear cost. This technique is applied to microscopic urinalysis image recognition, resulting in a significant improvement of recognition rate over the context free approach. This gain would have been impossible using conventional context incorporation techniques. 1 BACKGROUND: RECOGNITION IN CONTEXT There are a number of pattern recognition problem domains where the classification of an object should be based on more than simply the appearance of the object itself. In remote sensing image classification, where each pixel is part of ground cover, a pixel is more likely to be a glacier if it is in a mountainous area, than if surrounded by pixels of residential areas. In text analysis, one can expect to find certain letters occurring regularly in particular arrangement with other letters(qu, ee,est, tion, etc.). The information conveyed by the accompanying entities is referred to as contextual information. Human experts apply contextual information in their decision making [2][ 6]. It makes sense to design techniques and algorithms to make computers aggregate and utilize a more complete set of information in their decision making the way human experts do. In pattern recognition systems, however, *Author for correspondence 964 X B. Song, J Sill, Y. Abu-Mostafa and H. Kasdan the primary (and often only) source of information used to identify an object is the set of measurements, or features , associated with the object itself. Augmenting this information by incorporating context into the classification process can yield significant benefits. Consider a set of N objects Ti , i = 1, ... N. With each object we associate a class label Ci that is a member of a label set n = {1 , ... , D} . Each object Ti is characterized by a set of measurements Xi E R P, which we call a feature vector. Many techniques [1][2][4J[6} incorporate context by conditioning the posterior probability of objects' identities on the joint features of all accompanying objects. i.e .? P(Cl, C2,??? , cNlxl , . . . , XN). and then maximizing it with respectto Cl, C2, . .. , CN . It can ""' (N\ given be shown thatp(cl,c2, . . . , cNlxl, . . . ,xN) ex p(cllxl) ... p(CNlxN) p(~ci 1 ?.. p CN certain reasonable assumptions. Once the context-free posterior probabilities p( Ci IXi) are known. e.g. through the use of a standard machine learning model such as a neural network, computing P(Cl, ... ,CNlxl, . . . ,XN) for all possible Cl, ... ,CN would entail (2N + 1)DN multiplications. and finding the maximum has complexity of DN. which is intractable for large Nand D. [2J Another problem with this formulation is the estimation of the high dimensional joint distribution p( Cl, ... , CN), which is ill-posed and data hungry. One way of dealing with these problems is to limit context to local regions. With this approach, only the pixels in a close neighborhood. or letters immediately adjacent are considered [4][6][7J. Such techniques may be ignoring useful information, and will not apply to situations where context doesn't have such locality, as in the case of microscopic urinalysis image recognition. Another way is to simplify the problem using specific domain knowledge [1], but this is only possible in certain domains. These difficulties motivate the efficient incorporation of partial context as a general framework, formulated in section 2. In section 3, we discuss microscopic urinalysis image recognition. and address the importance of using context for this application. Also in section 3, techniques are proposed to identify relevant context. Empirical results are shown in section 4. followed by discussions in section 5. 2 FORMULATION FOR INCORPORATION OF PARTIAL CONTEXT To avoid the exponential computational cost of using the identities of all accompanying objects directly as context, we use "partial context". denoted by A. It is called "partial" because it is derived from the class labels. as opposed to consisting of an explicit labelling of all objects. The physical definition of A depends on the problem at hand. In our application. A represents the presence or absence of certain classes. Then the posterior probability of an object Ti having class label Ci conditioned on its feature vector and the relevant context A is p(XiICi, A)P(Ci ; A) P(Xi ; A) We assume that the feature distribution of an object depends only on its own class. i.e., p(xilci, A) = p(xi lci) . This assumption is roughly true for most real world problems. Then. Image Recognition in Context: Application to Microscopic Urinalysis 965 ( .1 .)p(ciI A ) p(A)p(Xi) ( .1 . A) -- p(xilci)p(Ci; A) _ -pCtX t p(xijJ~IIA) p(Ci) P(Xi; A) pC~Xt, ()( p(cilxi) () = p(cilxi)P(Ci, A) Ci where p(Ci, A) = p~(~j~) is called the context ratio, through which context plays its role. The context-sensitive posterior probability p( Ci lXi, A) is obtained through the context-free posterior probability p(cilxi) modified by the context ratio P(Ci, A) . P The partial-context maximum likelihood decision rule chooses class label Ci for element i such that Ci = argmaxp(cilxi, A) (I) Cj A systematic approach to identify relevant context A is addressed in section 3.3. The partial-context approach treats each element in a set individually, but with additional information from the context-bearing factor A . Once p(cilxi) are known for all i = 1, ... , N, and the context A is obtained, to maximize p(cilxi, A) from D possible values that Ci can take on and for all i, the total number of multiplications is 2N, and the complexity for finding the maximum is N D. Both are linear in N. The density estimation part is also trivial since it is very easy to estimate p(cIA). 3 3.1 MICROSCOPIC URINALYSIS INTRODUCTION Urine is one of the most complex body fluid specimens: it potentially contains about 60 meaningful types of elements. Microscopic urinalysis detects the presence of elements that often provide early diagnostic information concerning dysfunction, infection, or inflammation of the kidneys and urinary tract. Thus this non-invasive technique can be of great value in clinical case management. Traditional manual microscopic analysis relies on human operators who read the samples visually and identify them, and therefore is time-consuming, labor-intensive and difficult to standardize. Automated microscopy of all specimens is more practical than manual microscopy, because it eliminates variation among different technologists. This variation becomes more pronounced when the same technologist examines increasing numbers of specimens. Also, it is less labor-intensive and thus less costly than manual microscopy. It also provides more consistent and accurate results. An automated urinalysis system workstation (The Y ellowI RI ST M, International Remote Imaging Systems, Inc.) has been introduced in numerous clinical laboratories for automated microscopy. Urine samples are processed and examined at lOOx (low power field) and 400x magnifications (high power field) with bright-field illumination. The Y ellowI RI ST M automated system collects video images of formed analytes in a stream of un centrifuged urine passing an optical assembly. Each image has one analyte in it. These images are given to a computer algorithm for automatic identification of analytes. Context is rich in urinalysis and plays a crucial role in analyte classification. Some combinations of analytes are more likely than others. For instance, the presence of bacteria indicates the presence of white blood cells, since bacteria tend to cause infection and thus trigger the production of more white blood cells. If amorphous crystals show up, they tend to show up in bunches and in all sizes. Therefore, if there are amorphous crystallook-alikes in various sizes, it is quite possible that they are amorphous crystals. Squamous epithelial cells can appear both flat or rolled up. If squamous epithelial cells in one form are detected, X B. Song, J Sill, Y. Abu-Mostafa and H. Kasdan 966 Table I: Features extracted from urine anylates images reature number reature desc:ription ( 2 4 9 10 tht: m~an or hluc distribution the mean of gn...-cn dislrihulmn 15 th paccnlile of ?ray level hislo?ram 85 th percenlile of gray level hislogmm lh~ standard devia.tion \11' gray level intensity energy of the (.aplacian lransl\)rmalion of grey level image II 12 13 14 IS 16 then it is likely that there are squamous epithelial cells in the other form. Utilizing such context is crucial for classification accuracy. The classes we are looking at are bacteria, calcium oxalate crystals, red blood cells, white blood cells, budding yeast, amorphous crystals, uric acid crystals, and artifacts. The task of automated microscopic urinalysis is, given a urine specimen that consists of up to a few hundred images of analytes, to classify each analyte into one of these classes. The automated urinalysis system we developed consists of three steps: image processing and feature extraction, learning and pattern recognition, and context incorporation. Figure 1 shows some example analyte images. Table 1 gives a list of features extracted from analyte images. 1 3.2 CONTEXT-FREE CLASSIFICATION The features are fed into a nonlinear feed-forward neural network with 16 inputs, 15 hidden units with sigmoid transfer functions, and 8 sigmoid output units. A cross-entropy error function is used in order to give the output a probability interpretation. Denote the input feature vector as x, the network outputs a D dimensional vector (D = 8 in our case) p = {p(dlx)} , d = 1, ... , D, where p(dlx) is p{dlx) = Prob( an analyte belongs to class dl feature x) The decision made at this stage is d{x) = argmax p(dlx) d 3.3 IDENTIFICATION OF RELEVANT PARTIAL CONTEXT Not all classes are relevant in terms of carrying contextual information. We propose three criteria based on which we can systematicalIy investigate the relevance of the class presence. To use these criteria, we need to know the folIowing distributions: the class prior distribution p(c) for c = 1, ... ,D; the conditional class distribution p{cIAd) for c = 1, ... ,D 1>'1 and >'2 are respectively the larger and the smaller eigenvalues of the second moment matrix of an image. 967 Image Recognition in Context: Application to Microscopic Urinalysis and d = 1, . .. ,D; and the class presence prior distribution p(Ad) for d = 1, . . . , D. Ad is a binary random variable indicating the presence of class d. Ad = 1 if class d is present, and Ad = 0 otherwise. All these distributions can be easily estimated from the database. The first criterion is the correlation coefficient between the presence of any two classes; the second one is the classical mutual information I(e; Ad) between the presence of a class Ad and the class probability pee), where I(e; Ad) is defined as I(e; Ad) = H(e) H(eIAd) where H(e) = 2:~1 p(e = i)ln(p(e = i)) is the entropy of the class priors and H(eIAd) = P(Ad = I)H(eIAd = 1)+P(Ad = O)H(eIAd = 0) is the conditional entropy of e conditioned on Ad. The third criterion is what we call the expected relative entropy D(eIIAd) between the presence ofa class Ad and the labeling probability pee) , which we define as D(eIIAd) = P(Ad = I)D(p(e)llp(eIAd = 1)) + P(Ad = O)D(p(e)llp(eIAd = 0)) where D(p(e)llp(eIAd 1)) 2:~lP(e = ilAd = l)ln(p(c;/l~t)=l)) and D(p(e)llp(eIAd = 0)) = 2:~1 p(e = ilAd = O)ln(p(C;/l~t)=O)) According to the first criterion, one type of analyte is considered relevant to another if the absolute value of their correlation coefficient is beyond a certain threshold. It shows that uric acid crystals, budding yeast and calcium oxalate crystals are not relevant to any other types even by a generous threshold of 0.10. Similarly, the bigger the mutual information between the presence of a class and the class distribution, the more relevant this class is. Ranking the analyte types in terms of I(e; Ad) in a descending manner gives rise to the following list: bacteria, amorphous crystals, red blood cells, white blood cells, uric acid crystals, budding yeast and calcium oxalate crystals. Once again, ranking the analyte types in terms of D(eIIAd) in a descending manner gives rise to the following list: bacteria, red blood cells, amorphous crystals, white blood cells, calcium oxalate crystals, budding yeast and uric acid crystals. All three criteria lead to similar conclusions regarding the relevance of class presence - bacteria, red blood cells, amorphous crystals, and white blood cells are relevant, while calcium oxalate crystals, budding yeast and uric acid crystals are not. (Baed on prior knowledge, we discard artifacts from the outset as an irrelevant class.) 3.4 ALGORITHM FOR INCORPORATING PARTIAL CONTEXT Once the M relevant classes are identified, the following algorithm is used to incorporate partial context. Step 0 Estimate the priors p(eIAd) and pee), for e E {I, 2, .. . , D} and d E {I, 2, ... , D}. Step 1 For a given Xi, compute p(edxi) for ei = 1,2, . .. , Dusing whichever base machine learning model is preferred ( in our case, a neural network). Step 2 Let the M relevant classes be R 1 , ..? , RM. According to the no-context p( ei IXi) and certain criteria for detecting the presence or absence of all the relevant classes, get A RI , ? ?? ,ARM' Step 3 Letp(ei lXi , Ao) = p(eilxi), where Ao is the null element. Incorporate context from each relevant class sequentially, i.e., for m = 1 to M, iteratively compute p(eilxi; Ao, .. . , ARm_I ' ARTn) = p(eilxi' Ao,.? . , ARTn_J p(ei IARTn)p(AR"J pee) Step 4 Recompute A RI , . . . ,ARM based on the new class labellings. Return to step 3 and repeat until algorithm converges. 2 2Hence, the algorithm has an E-M flavor, in that it goes back and forth between finding the most 968 X B. Song, J. Sill, Y. Abu-Mostafa and H Kasdan amorphous crystals artifacts calcium oxalate crystals hyaline casts Figure I: Example of some of the analyte images. Step 5 Label the objects according to the final context-contammg p(cilxi, ARI'?? ? ' ARM)' i.e., Ci = argmaxp(ciIXi, A R1 , ... , ARM) for i = 1, ... , N. Ci This algorithm is invariant with respect to the ordering of the M relevant classes in (Ai, ... , AM). The proof is omitted here. 4 RESULTS The algorithm using partia.1 context was tested on a database of 83 urine specimens, containing a total of 20,276 analyte images. Four classes are considered relevant according to the criteria described in section 3.3: bacteria, red blood cells, white blood cells and amorphous crystals. We measure two types of error: analyte-by-analyte error, and specimen diagnostic error. The average analyte-by-analyte error is reduced from 44.48% before using context to 36.66% after, resulting a relative error reduction of 17.6% (Table 2). The diagnosis for a specimen is either normal or abnormal. Tables 3 and 4 compare the diagnostic performance with and without using context, and Table 5 lists the relative changes. We can see using context significantly increases correct diagnosis for both normal and abnormal specimens, and reduces both false positives and false negatives. average element-by-element error without context 44.48 % with context 36.66 % Table 2: Comparison of using and not using contextual information for analyte-by-analyte error. probable class labels given the context and determining the context given the class labels. Image Recognition in Context: Application to Microscopic Urinalysis truly normal truly abnormal estimated normal 40.96 % 19.28 % 969 estimated abnormal 7.23 % 32.53 % Table 3: Diagnostic confusion matrix not using context truly normal truly abnormal estimated normal 42.17 % 16.87 % estimated abnormal 6.02% 34.94 % Table 4: Diagnostic confusion matrix using context truly normal truly abnormal estimated normal +2.95 % - 12.50 % estimated abnormal -16.73 % +7.41 % Table 5: Relative accuracy improvement (diagonal elements) and error reduction (off diagonal elements) in the diagnostic confusion matrix by using context. 5 CONCLUSIONS We proposed a novel framework that can incorporate context in a simple and efficient manner, avoiding exponential computation and high dimensional density estimation. The application of the partial context technique to microscopic urinalysis image recognition demonstrated the efficacy of the algorithm. This algorithm is not domain dependent, thus can be readily generalized to other pattern recognition areas. ACKNOWLEDGEMENTS The authors would like to thank Alexander Nicholson, Malik Magdon-Ismail, Amir Atiya at the Caltech Learning Systems Group for helpful discussions. References [I) Song, X.B . & SilU. & Abu-Mostafa & Harvey Kasdan, (1997) "Incorporating Contextual Information in White Blood Cell Identification", In M. Jordan, MJ. Kearns and S.A. Solla (eds.), Advances in Neural Information Processing Systems 7,1997, pp. 950-956. Cambridge, MA: MIT Press. [2] Song, Xubo (1999) "Contextual Pattern Recognition with Application to Biomedical Image Identification", Ph.D. Thesis, California Institute of Science and Technology. [3) Boehringer-Mannheim-Corporation, Urinalysis Today, Boehringer-Mannheim-Corporation, 1991. [4] Kittler, J.."Relaxation labelling", Pattern Recognition Theory and Applications, 1987, pp. 99108., Pierre A. Devijver and Josef Kittler, Editors, Springer-Verlag. [5] Kittler, J. & Illingworth, J., " Relaxation Labelling Algorithms - A Review", Image and Vision Computing, 1985, vol. 3, pp. 206-216. [6] Toussaint, G., "The Use of Context in Pattern Recognition", Pattern Recognition, 1978, vol. 10, pp. 189-204. [7] Swain, P. & Vardeman, S. & Tilton, J., "Contextual Classification of Multispectral Image Data", Pattern Recognition, 1981, Vol. 13, No.6, pp. 429-441.
1675 |@word grey:1 nicholson:1 reduction:2 moment:1 contains:1 efficacy:1 current:1 contextual:8 readily:1 amir:1 detecting:1 recompute:1 provides:1 dn:2 c2:3 consists:2 ray:1 manner:3 expected:1 roughly:1 detects:1 increasing:1 becomes:1 null:1 what:1 developed:1 finding:3 corporation:2 ti:3 ofa:1 rm:1 unit:2 appear:1 mountainous:1 before:1 positive:1 engineering:2 local:1 treat:1 limit:1 examined:1 collect:1 sill:4 graduate:1 practical:1 area:3 empirical:1 significantly:1 outset:1 get:1 close:1 operator:1 context:54 impossible:1 descending:2 conventional:1 demonstrated:1 maximizing:1 go:1 kidney:1 immediately:1 rule:1 examines:1 utilizing:1 tht:1 variation:2 play:2 trigger:1 today:1 ixi:2 associate:1 element:9 standardize:1 recognition:19 magnification:1 database:2 role:2 electrical:2 region:1 kittler:3 remote:3 ordering:1 solla:1 thatp:1 complexity:2 motivate:1 carrying:1 easily:1 joint:2 respectto:1 various:1 detected:1 labeling:1 aggregate:1 neighborhood:1 quite:1 posed:1 larger:1 otherwise:1 itself:2 final:1 eigenvalue:1 propose:3 relevant:15 forth:1 ismail:1 pronounced:1 r1:1 argmaxp:2 tract:1 converges:1 object:15 augmenting:1 technologist:2 correct:1 human:3 ao:4 probable:1 desc:1 accompanying:3 considered:3 ground:1 normal:8 visually:1 great:1 mostafa:5 early:1 generous:1 omitted:1 pee:4 estimation:3 epithelial:3 label:8 sensitive:1 individually:1 mit:1 modified:1 avoid:1 derived:1 improvement:2 likelihood:1 indicates:1 xubo:1 sense:1 am:1 helpful:1 dependent:1 nand:1 pasadena:2 hidden:1 josef:1 pixel:4 classification:8 ill:1 among:1 denoted:1 mutual:2 field:3 once:4 having:1 extraction:1 represents:1 others:1 simplify:1 few:1 argmax:1 consisting:1 investigate:1 rolled:1 truly:6 pc:1 accurate:1 partial:11 bacteria:7 lh:1 instance:1 classify:1 gn:1 inflammation:1 cover:1 ar:1 cost:3 hundred:1 swain:1 chooses:1 st:2 density:2 international:2 systematic:1 off:1 again:1 thesis:1 management:1 opposed:1 containing:1 expert:2 return:1 busy:1 coefficient:2 inc:2 oregon:1 ranking:2 depends:2 stream:1 ad:15 tion:2 red:5 multispectral:1 amorphous:9 bright:1 formed:1 accuracy:2 acid:5 who:1 yield:1 identify:4 identification:4 bunch:1 lci:1 manual:3 infection:2 ed:1 definition:1 energy:1 pp:5 invasive:1 associated:1 proof:1 workstation:1 gain:1 knowledge:2 cj:1 back:1 feed:1 formulation:2 stage:1 biomedical:1 correlation:2 until:1 hand:1 ei:4 nonlinear:1 artifact:3 gray:2 yeast:5 true:1 hence:1 read:1 laboratory:1 iteratively:1 white:8 ogi:1 adjacent:1 dysfunction:1 criterion:8 generalized:1 crystal:19 complete:1 confusion:3 hungry:1 analyte:17 image:24 novel:1 ari:1 sigmoid:2 physical:1 conditioning:1 interpretation:1 significant:2 measurement:2 cambridge:1 ai:1 automatic:1 iia:1 similarly:1 entail:1 etc:1 base:1 posterior:5 own:1 belongs:1 irrelevant:1 discard:1 verlag:1 certain:6 harvey:2 binary:1 caltech:3 additional:1 cii:1 specimen:8 maximize:1 ii:1 reduces:1 characterized:1 clinical:2 cross:1 concerning:1 bigger:1 vision:1 microscopy:4 cell:16 background:1 addressed:1 source:1 crucial:2 eliminates:1 tend:2 member:1 regularly:1 incorporates:1 jordan:1 call:2 ee:1 presence:13 easy:1 automated:6 identified:1 regarding:1 cn:5 intensive:2 song:5 yaser:1 passing:1 cause:1 useful:1 ph:1 processed:1 atiya:1 reduced:1 diagnostic:6 estimated:7 diagnosis:2 vol:3 abu:5 group:1 four:1 boehringer:2 threshold:2 blood:13 utilize:1 imaging:2 ram:1 relaxation:2 residential:1 prob:1 letter:3 reasonable:1 decision:4 abnormal:8 followed:1 correspondence:1 uric:5 letp:1 incorporation:4 ri:4 flat:1 optical:1 department:3 according:4 combination:1 smaller:1 lp:1 joseph:1 qu:1 making:2 labellings:1 invariant:1 ln:3 discus:1 know:1 fed:1 whichever:1 magdon:1 apply:2 reature:2 pierre:1 cia:1 lxi:2 kasdan:5 assembly:1 beaverton:1 classical:1 malik:1 arrangement:1 primary:1 costly:1 traditional:1 diagonal:2 microscopic:12 thank:1 entity:1 trivial:1 ratio:2 difficult:1 cilxi:7 potentially:1 negative:1 fluid:1 rise:2 design:1 calcium:6 budding:5 situation:1 looking:1 intensity:1 introduced:1 cast:1 california:3 address:1 beyond:1 llp:4 pattern:9 video:1 power:2 difficulty:1 squamous:3 arm:4 technology:4 numerous:1 analytes:4 text:1 prior:5 review:1 acknowledgement:1 multiplication:2 determining:1 relative:4 expect:1 toussaint:1 conveyed:1 consistent:1 editor:1 surrounded:1 production:1 repeat:1 free:4 institute:4 face:1 absolute:1 benefit:1 xn:3 world:1 rich:1 doesn:1 author:2 forward:1 made:1 preferred:1 dealing:1 mannheim:2 sequentially:1 consuming:1 xi:6 un:1 urine:6 table:9 mj:1 transfer:1 ca:3 ignoring:1 bearing:1 cl:6 complex:1 domain:4 body:1 referred:1 explicit:1 dlx:4 exponential:3 third:1 xt:1 specific:1 sensing:1 list:4 dl:1 incorporating:4 intractable:1 joe:1 false:2 importance:1 ci:17 labelling:3 conditioned:2 occurring:1 illumination:1 xubosong:2 flavor:1 locality:1 entropy:4 simply:1 appearance:1 likely:3 labor:2 springer:1 devijver:1 relies:1 extracted:2 ma:1 conditional:2 identity:2 formulated:1 absence:2 change:1 kearns:1 called:2 total:2 ece:1 est:1 meaningful:1 indicating:1 alexander:1 relevance:2 incorporate:4 tested:1 avoiding:1 ex:1
739
1,676
Recurrent cortical competition: Strengthen or weaken? Peter Adorjan*, Lars Schwabe, Christian Piepenbrock* , and Klaus Obennayer Dept. of Compo Sci., FR2-I, Technical University Berlin Franklinstrasse 28/29 10587 Berlin, Germany adorjan@epigenomics.com, {schwabe, oby} @cs.tu-berlin.de, piepenbrock@epigenomics.com http://www.ni.cs.tu-berlin.de Abstract We investigate the short term .dynamics of the recurrent competition and neural activity in the primary visual cortex in terms of information processing and in the context of orientation selectivity. We propose that after stimulus onset, the strength of the recurrent excitation decreases due to fast synaptic depression. As a consequence, the network shifts from an initially highly nonlinear to a more linear operating regime. Sharp orientation tuning is established in the first highly competitive phase. In the second and less competitive phase, precise signaling of multiple orientations and long range modulation, e.g., by intra- and inter-areal connections becomes possible (surround effects). Thus the network first extracts the salient features from the stimulus, and then starts to process the details. We show that this signal processing strategy is optimal if the neurons have limited bandwidth and their objective is to transmit the maximum amount of information in any time interval beginning with the stimulus onset. 1 Introduction In the last four decades there has been a vivid and highly polarized discussion about the role of recurrent competition in the primary visual cortex (VI) (see [12] for review). The main question is whether the recurrent excitation sharpens a weakly orientation tuned feedforward input, or the feed-forward input is already sharply tuned, hence the massive recurrent circuitry has a different function. Strong cortical recurrency implements a highly nonlinear mapping of the feed-forward input, and obtains robust and sharply tuned cortical response even if only a weak or no feed-forward orientation bias is present [6, 11, 2]. However, such a competitive network in most cases fails to process mUltiple orientations within the classical receptive field and may signal spurious orientations [7]. This motivates the concept that the primary visual cortex maps an already sharply orientation tuned feed-forward input in a less competitive (more linear) fashion [9, 13]. Although these models for orientation selectivity in VI vary on a wide scale, they have one common feature: each of them assumes that the synaptic strength is constant on the short time scale on which the network operates. Given the phenomenon of fast synaptic *Current address: Epigenomics GmbH, Kastanienallee 24,0-10435 Berlin, Germany P. Adorjan, L. Schwabe, C. Piepenbrock and K. Obermayer 90 dynamics this, however, does not need to be the case. Short term synaptic dynamics, e.g., of the recurrent excitatory synapses would allow a cortical network to operate in bothcompetitive and linear-regimes. We will show below (Section 2) that such a dynamic cortical amplifier network can establish sharp contrast invariant orientation tuning from a broadly tuned feed-forward input, while it is still able to respond correctly to mUltiple orientations. We then show (Section 3) that decreasing the recurrent competition with time naturally follows from functional considerations, i.e. from the requirement that the mutual information between stimuli and representations is maximal for any time interval beginning with stimulus onset. We consider a free-viewing scenario, where the cortical layer represents a series of static images that are flashed onto the retina for a fixation period (~T = 200 - 300 ms) between saccades. We also assume that the spike count in increasing time windows after stimulus onset carries the information. The key observations are that the signal-to-noise ratio of the cortical representation increases with time (because more spikes are available) and that the optimal strength of the recurrent connections (w.r.t. information transfer) decreases with the decreasing output noise. Consequently the model predicts that the information content per spike (or the SNR for ajixed sliding time window) decreases with time for a flashed static stimulus in accordance with recent experimental studies. The neural system thus adapts to its own internal changes by modifying its coding strategy, a phenomenon which one may refer to as "dynamic coding". 2 Cortical amplifier with fast synaptic plasticity To investigate our first hypothesis, we set up a model for an orientation-hypercolumn in the primary visual cortex with similar structure and parameters as in [7]. The important novel feature of our model is that fast synaptic depression is present at the recurrent excitatory connections. Neurons in the cortical layer receive orientation-tuned feed-forward input from the LGN and they are connected via a Mexican-hat shaped recurrent kernel in orientation space. In addition, the recurrent and feed-forward excitatory synapses exhibit fast depression due to the activity dependent depletion of the synaptic transmitter [1, 14]. We compare the response of the cortical amplifier models with and without fast synaptic plasticity at the recurrent excitatory connections to single and mUltiple bars within the classical receptive field. The membrane potential V (0, t) of a cortical cell tuned to an orientation 0 decreases due to the leakage and the recurrent inhibition, and increases due to the recurrent excitation T a at V(O, t) + V(O, t) (1) where T = 15 ms is the membrane time constant and ILGN (0, t) is the input received from the LGN. The recurrent excitatory and inhibitory cortical inputs are given by r:"(O, t) (2) where ~ (Of, 0) is a 1T periodic circular difference between the preferred orientations, JCX(O, Of , t) are the excitato~ and inhibitory connection strengths (with a E {exc, inh}, J~x~x = 0.2 m V /Hz and J:::ax = 0.8m V /Hz), and f is the presynaptic firing rate. The excitatory synaptic efficacy r xc is time dependent due to the fast synaptic depression, while the efficacy of inhibitory synapses Jinh is assumed to be constant. The recurrent excitation is sharply tuned (j exc = 7.5 0 , while the inhibition has broad tuning (jinh = 90 0 ? The mapping from the membrane potential to firing rate is approximated by a linear function with a threshold at 0 (f(O) = ,6max(O, V(O)),,6 = 15Hz/mV). Gaussian-noise with variances 91 Recurrent Cortical Competition: Strengthen or Weaken? Feedforward Input ,....., 1 >E ....::l g- Static N X '-' - n Q) rJJ '-' ,--,., , = o.. --- ,, 0 rJJ Q) ~ ~90 -45 45 90 0 Orientation [deg] Depressing 15 ~90 -45 (a) j! N X 15 '-' Q) rJJ = 0 0.. rJJ Q) ~ \ 0 45 90 Orientation [deg] ~90 -45 0 45 90 Orientation [deg] (b) (c) Figure 1: The feed-forward input (a), and the response ofthe cortical amplifier model with static recurrent synaptic strength (b), and a network with fast synaptic depression (c) if the stimulus is single bar with different stimulus contrasts (40%dotted; 60%dashed; 80%solid line). The cortical response is averaged over the first 100 illS after stimulus onset. of 6 Hz and 1.6 Hz is added to the input intensities and to the output of cortical neurons. The orientation tuning curves of the feed-forward input ILGN are Gaussians (O'"LGN = 18?) resting on a strong additive orientation independent component which would correspond to a geniculo-cortical connectivity pattern with an approximate aspect ratio of 1:2. Both, the orientation dependent and independent components increase with contrast. Considering a free-viewing scenario where the environment is scanned by saccading around and fixating for short periods of 200 - 300 illS we model stationary stimuli present for 300 illS. The stimuli are one or more bars with different orientations. Feed-forward and recurrent excitatory synapses exhibit fast depression. Fast synaptic deI'ression is modeled by the dynamics of the expected synaptic transmitter or "resource" R( t) for each synapse. The amount of the available transmitter decreases proportionally to the release probability p and to the presynaptic firing rate /, and it recovers exponentially (T~~N = 120 illS, Tr~~x = 850 illS, pLGN = 0.35 and pCtx = 0.55), 1 - R(t) Tree - /(t)p(t)R(t) = - Teff R(t) (f() ()) 1 +. t, P t Tree (3) The change of the membrane potential on the postsynaptic cell at time t is proportional to the released transmitter pR(t). The excitatory connectivity strength between neurons tuned to orientations 0 and 0' is expressed as j?xe(o , 0', t) = J:::~xpR991(t). Similarly this applies to the feed-forward synapses. Fast synaptic plasticity at the feed-forward synapses has been investigated in more detail in previous studies [3, 4]. In the following, we compare the predictions of the cortical amplifier model with and without fast synaptic depression at the recurrent excitatory connections. In both cases fast synaptic depression is present at the feed-forward connections limiting the duration of the effective feed-forward input to 200 - 400 illS. Figure 1 shows the orientation tuning curves at different stimulus contrasts. The feed-forward input is noisy and broadly tuned (Fig. la). Both models exhibit contrast invariant tuning (Fig. 1b, c). If fast synaptic depression is present at the recurrent excitation, the cortical network sharpens the broadly tuned feedforward input in the initial response phase. Once sharply tuned input is established, the tuning width does not change, only the response amplitude decreases in time. The predictions of the two models differ substantially if multiple orientations are present (Fig. 2). At first, we test the cortical response to two bars separated by 60? with different intensities (Figs. 2a, b). If the recurrent synaptic weights are static and strong enough (Fig. 2a), then only one orientation is signaled. The cortical network selects the orientation 92 P Adorjim. L. Schwabe. C. Piepenbrock and K. Obermayer Feedforward Input (a) f~1 I~I~--~_~~"~,------~ ~90 -45 0 45 Activity Profile Average Cortical Response g .: S ~ 90 Orientation [deg] ,' I .b~---~#"" A I .,g 90 0 -90 90 . ~ o ?c (b) .. ,----, ... '.-'--<" ',,--,'" ;;20 1~ .. .. ???INI 'i 110 :>E c: (c) - .. ~90 ,-,,, , .... -- '. a) 0 45 Orienretionldeg) 90 ~ -90 90 0 r:S:;;1 (: Ul -45 g .~ 0 ~ ''''' 0 , " --- /', -90 ~~~_ o '" ~90- -45 . 0 '. 45 '90 -90 (d) Orientation [deg] :;a????\~~~\$~* ~'> ~ \m0 '" s m?" : 0 ~ . ~-... 150 Time [ms] 300 Figure 2: The response of the cortical amplifier model with static (a,c) and fast depressing recurrent synapses (b, d). In both models the feed-forward synapses are fast depressing. In the left column the feed-forward input is shown, that is same for both models. Two types of stimuli were applied. The first stimulus consists of a stronger (a = -30?) and a weaker bar (a +30?) (a, b); the second stimulus consists of three equal intensity bars with orientations that are separated by 60? (c, d). In the middle column the cortical response is shown averaged for different time windows ([0 .. 30] dotted; [0 .. 80] dashed; [200 .. 300] solid line). In the right column the cortical activity profile is plotted as a function of time. Gray values indicate the activity with bright denoting high activities. = with the highest amplitude in a winner-take-all fashion. In contrast, if synaptic depression is present at the recurrent excitatory synapses, both bars are signaled in parallel (at low release probability, Fig. 2b) or after each other (high release probability, data not shown). First, those cells fire which are tuned to the orientation of the bar with the stronger intensity, and a sharply tuned response emerges at a single orientation-the network operates in a winner-take-all regime. The synapses of these highly active cells then become strongly depressed and cortical competition decreases. As the network is shifted to a more linear operation regime, the second orientation is signaled too. Note that this phenomenontogether with the observed contrast invariant tuning-cannot be reproduced by simply decreasing the static synaptic weights in the cortical amplifier model. The recurrent synaptic efficacy changes inhomogeneously in the network depending on the activity. Only the synapses of the highly active cells depress strongly, and therefore a sharply tuned response can be evoked by a bar with weak intensity. Fast synaptic depression thus behaves as a local self-regulation that modulates competition with a certain delay. This delay, and therefore the delay of the rise of the response to the second bar depends on the effective time constant reff(f(t),p) = rrec/(l + pf(t)rrec) of the synaptic depression at the recurrent connections. If the depression becomes faster due to an increase in the release probability p, then the delay decreases. The delay also scales with the difference between the bar intensities. The closer to each other they are, the shorter the delay will be. In Figs. 2c, d the cortical response to three bars with equal intensities is presented. Cells tuned to the presented three orientations respond in parallel if fast synaptic depression at the recurrent excitation is present (Figs. 2d). The cortical network with strong static recurrent synapses again fails to signal faithfully its feed-forward input. Additive noise on the 93 Recurrent Cortical Competition: Strengthen or Weaken? feed-forward input introduces a slight symmetry breaking and the network with static recurrent weights responds strongly at the orientation of only one of the presented bars (Fig. 2c). In summary, our simulations revealed that a recurrent network with fast synaptic depression is capable of obtaining robust sharpening of its feed-forward input and it also responds correctly to multiple orientations. Note that other local activity dependent adaptation mechanisms, such as slow potassium current, would have similar effects as the synaptic depression on the highly orientation specific excitatory connections. An experimentally testable prediction of our model is that the response to a flashed bar with lower contrast can be delayed by masking it with a second bar with higher contrast (Fig. 2b, right). We also suggest that long range integration from outside of the classical receptive field could emerge with a similar delay. In the initial phase of the cortical response, strong local features are amplified. In the longer, second phase, recurrent competition decreases and then weak modulatory recurrent or feed-forward input has a stronger relative effect. In the following, we investigate whether this strategy is favorable from the point of view of cortical encoding. 3 Dynamic coding In the previous section we have proposed that during cortical processing a highly nonlinear phase is followed by a more linear mode if we consider a short stimulus presentation or a fixation period. The simulations demonstrated that unless the recurrent competition is modulated in time, the network fails to account for more than one feature in its input. From a strictly functional point of view the question arises, why not to use weak recurrent competition during the whole processing period. We investigate this problem in an abstract signal-encoder framework i7 = g( i) + 1] , (4) where i is the input to the "cortical network", g(i) is a nonlinear mapping and-for the sake of simplicitY-1] is additive Gaussian noise. Naturally, in a real recurrent network output noise becomes input noise because of the feedback. Here we use the simplifying assumption that only output noise is present on the transformed input signal (input noise would lead to different predictions that should be further investigated). Output noise can be interpreted as a noisy channel that projects out from, e.g., the primary visual cortex. The nonlinear transformation g(i) here is considered as a functional description of a cortical amplifier network without analyzing how actually it is "implemented". Considering orientation selectivity, the signal i can be interpreted as a vector of intensities (or contrasts) of edges with different orientations. Edges which are not present have zero intensity. The coding capacity of a realistic neural network is limited. Among several other noise sources, this limitation could arise from imprecision in spike timing and a constraint on the maximal or average firi ng rate. The input-output mapping g( i) of a cortical amplifier network is approximated with the soft-max function (5) The f3 parameter can be interpreted as the level of recurrent competition. As f3 -+ 0 the network operates in a more linear mode, while f3 -+ 00 puts it into a highly nonlinear winner-take-all mode. In all cases the average activity in the network is constrained which has been suggested to minimize metabolic costs [5]. Let us consider a factorizing input distribution, 1 p( i) = Z IIi exp (_x ---t- a) for x ~ 0, (6) P. Adorjfm, L. Schwabe, C. Piepenbrock and K. Obermayer 94 8rr===~~----~----~ --- 0.5 1.0 0--0 6 'J:;-----Q- .c!1 - - ,,-s-' 00 0.05 0.1 Noise (stdev) 0.15 Figure 3: The optimal competition parameter j3 as a function of the standard deviation of the Gaussian output noise 'f}. The optimal j3 is calculated for highly super-Gaussian, Gaussian, and sub-Gaussian stimulus densities. The sparsity parameter a is indicated in the legend. where the exponent a detennines the sparsity of the probability density function, Z is a nonnalizing constant, and ~ detennines the variance. If a = 2, the input density is the positive half of a multivariate Gaussian distribution. With a > 2 the signal distribution becomes sub-Gaussian, and with a < 2 it becomes super-Gaussian. For optimal processing in time one needs to gain the maximal infonnation about the signal for any increasing time window. Let us assume that the stimulus is static and it is presented for a limited time. As time goes ahead after stimulus onset, the time window for the encoding and the read-out mechanism increases. During a longer period more samples of the noisy network output are available, and thus the output noise level decreases with time. We suggest that the optimal competition parameter j3opt_at which the mutual infonnation between input i and output if (Eq. 4) is maximized-<iepends on the noise level. As the noise decreases with time, j3 or the recurrent cortical competition should also change during cortical processing. To demonstrate this idea, the mutual infonnation is calculated numerically for a three-dimensional state space. One might expect that at higher noise levels the highest infonnation transfer can be obtained if the typical and salient features are strongly amplified. Note that this is only true if the standard deviation of the noise scales sub-linearly with activity, which is true for an additive noise process as well as Poisson firing. As noise decreases (e.g., with increasing the time window for estimation), the level of competition should decrease distributing the available resources (e.g., spikes) among more units and letting the network respond to finer details at the input. Investigating the level of optimal competition j3 as a function of the standard deviation of the output noise (Fig. 3) this intuition is indeed justified. The optimal j3 scales with the standard deviation of the additive noise process. Comparing signal distributions with the same variance but with different sparsity exponents a, we find that the sparser the signal distribution is, the higher the optimal competition becomes, because multiple features are unlikely to be present at the same time if the input distribution is sparse. By enforcing competition, the optimal encoding strategy also generates an activity distribution where only few units fire for a presented stimulus. Since edges with different orientations fonn a sparse distributed representation of natural scenes [8], our work suggests that a strongly competitive visual cortical network could achieve a better performance on our visual environment than a simple linear network would do. We can now interpret our simulation results presented in the Section 2 from a functional point of view and give a prediction for the dynamics of the recurrent cortical competition. Noting that the output noise is decreasing with increasing time-window for encoding, the cortical competition should also decrease following a similar trajectory as presented in Fig. 3. If competition is low and static, then the cumulative mutual infonnation between input and output would converge only slowly towards the overall infonnation that is available in the stimulus. If the competition is high during the whole observation period, then after a fast rise the cumulative mutual information would saturate well below the possible Recurrent Cortical Competition: Strengthen or Weaken? 95 maximum. If the level of competition is dynamic, and it decreases from an initially highly competitive state, then the network obtains maximal information transfer in time. One may argue that the valuable information about the signals mainly depends on the interest of the observer. Considering an encoding system for one variable it has been suggested that in a highly attentive state the recurrent competition increases [10]. In the view of our results we would refine this statement by suggesting that competition increases or decreases depending on the level of visual detail the observer pays attention to. Whenever representation of small details is also required, reducing competition is the optimal strategy given enough bandwidth. In summary, using a detailed model for an orientation hypercolumn in VI we have demonstrated that sharp contrast invariant tuning and faithful representation of multiple features can be achieved by a recurrent network if the recurrent competition decreases in time after stimulus onset. The model predicts that the cortical response to weak details in the stimulus emerges with a delay if a second stronger feature is also present. The modulation from, e.g., outside of the classical receptive field also has a delayed effect on cortical activity. Our study within an abstract framework revealed that weakening the recurrent cortical competition on a fast time scale is functionally advantageous, because a maximal amount of information can be transmitted in any time window after stimulus onset. Acknowledgments Supported by the Boehringer Ingelheim Fonds (C. P.), by the German Science Foundation (DFG grant GK 120-2) and by Wellcome Trust 0500801ZJ97. References [1] L. F. Abbott, J. A. Varela, K. Sen, and S. B. Nelson. Synaptic depression and cortical gain control. Science, 275:220-224,1997. [2] P. Adoljan, J.B. Levitt, J.S. Lund, and K. Obennayer. A model for the intracortical origin of orientation preference and tuning in macaque striate cortex. Vis. Neurosci .? 16:303-318. 1999. Contrast adaptation and info[3] P. AdOljan. C. Piepenbrock. and K. Obennayer. max in visual cortical neurons. Rev. Neurosci.. 10: 181-200, 1999. ftp://ftp.cs.tuberlin.de/pub/locaVnilpapers/adp99-contrast.ps.gz. [4] O. B. Artun, H. Z. Shouval, and L. N. Cooper. The effect of dynamic synapses on spatiotemporal receptive fields in visual cortex. Proc. Natl. Acad. Sci.? 95:11999-12003, 1998. [5] R. Baddeley. An efficient code in VI? Nature. 381 :560-561,1996. [6] R. Ben-Yishai, R. Lev Bar-Or. and H. Sompolinsky. Theory of orientation tuning in visual cortex. Proc. Natl. Acad. Sci.? 92:3844-3848,1995. [7] M. Carandini and D. L. Ringach. Predictions of a recurrent model of orientation selectivity. Vision Res., 37:3061-3071.1997. [8] D. J. Field. What is the goal of sensory coding. Neural Comput.? 6:559-601, 1994. [9] D. H. Hubel and T. N. Wiesel. Receptive fields. binocular interaction and functional architecture in cat's visual cortex. 1. Physiol., 165:559-568. 1962. [10] D. K. Lee. L. Itti, C. Kock. and 1. Braun. Attention activates winner-take-all competition among visual filters. Nat. Neurosci., 2:375-381,1999. [11] D. C. Somers, S. B. Nelson. and M. Sur. An emergent model of orientation selectiVity in cat visual cortical simple cells. 1. Neurosci., 15:5448-65, 1995. [12] H. Sompolinsky and R. Shapley. New perspectives on the mechanisms for orientation selectivity. Curro Op. in Neurobiol., 7:514-522, 1997. [13] T. W. Troyer, A. E. Krukowski. N. J. Priebe, and K. D. Miller. Contrast-invariant orientation tuning in visual cortex: Feedforward tuning and correlation-based intracortical connectivity. 1. Neurosci., 18:5908-5927. 1998. [14] M. V. Tsodyks and H. Markram. The neural code between neocortical pyramidal neurons depends on neurotransmitter release probability. Proc. Natl. Acad. Sci., 94:719-723. 1997.
1676 |@word middle:1 wiesel:1 sharpens:2 stronger:4 advantageous:1 simulation:3 simplifying:1 fonn:1 tr:1 solid:2 carry:1 initial:2 series:1 efficacy:3 pub:1 tuned:16 denoting:1 current:2 com:2 comparing:1 physiol:1 additive:5 realistic:1 plasticity:3 christian:1 piepenbrock:6 stationary:1 half:1 beginning:2 short:5 compo:1 preference:1 become:1 consists:2 fixation:2 shapley:1 inter:1 indeed:1 expected:1 decreasing:4 window:8 considering:3 increasing:4 becomes:6 pf:1 project:1 what:1 interpreted:3 substantially:1 neurobiol:1 sharpening:1 transformation:1 braun:1 control:1 unit:2 grant:1 positive:1 accordance:1 local:3 timing:1 consequence:1 acad:3 encoding:5 analyzing:1 lev:1 firing:4 modulation:2 might:1 evoked:1 suggests:1 limited:3 range:2 averaged:2 faithful:1 acknowledgment:1 implement:1 signaling:1 suggest:2 onto:1 cannot:1 put:1 context:1 www:1 map:1 demonstrated:2 go:1 attention:2 duration:1 simplicity:1 transmit:1 limiting:1 strengthen:4 massive:1 hypothesis:1 origin:1 fr2:1 approximated:2 predicts:2 observed:1 role:1 tsodyks:1 connected:1 sompolinsky:2 decrease:17 highest:2 valuable:1 intuition:1 environment:2 dynamic:10 weakly:1 emergent:1 cat:2 shouval:1 neurotransmitter:1 stdev:1 separated:2 fast:21 effective:2 oby:1 klaus:1 outside:2 encoder:1 noisy:3 reproduced:1 rr:1 sen:1 propose:1 interaction:1 maximal:5 adaptation:2 tu:2 achieve:1 adapts:1 amplified:2 description:1 competition:31 potassium:1 requirement:1 p:1 ben:1 ftp:2 depending:2 recurrent:46 op:1 received:1 eq:1 strong:5 implemented:1 c:3 indicate:1 differ:1 detennines:2 modifying:1 lars:1 filter:1 viewing:2 strictly:1 around:1 considered:1 exp:1 mapping:4 circuitry:1 m0:1 vary:1 released:1 favorable:1 estimation:1 proc:3 infonnation:6 faithfully:1 activates:1 gaussian:9 super:2 ax:1 release:5 transmitter:4 mainly:1 contrast:14 dependent:4 unlikely:1 weakening:1 initially:2 spurious:1 transformed:1 lgn:3 germany:2 selects:1 overall:1 adorjan:3 orientation:48 ill:6 among:3 exponent:2 constrained:1 integration:1 mutual:5 field:7 once:1 equal:2 shaped:1 ng:1 f3:3 represents:1 broad:1 saccading:1 stimulus:25 few:1 retina:1 delayed:2 dfg:1 phase:6 fire:2 amplifier:9 interest:1 investigate:4 highly:12 intra:1 circular:1 introduces:1 natl:3 yishai:1 edge:3 closer:1 capable:1 shorter:1 unless:1 tree:2 re:1 signaled:3 plotted:1 weaken:4 column:3 soft:1 cost:1 deviation:4 snr:1 delay:8 too:1 periodic:1 spatiotemporal:1 density:3 lee:1 connectivity:3 again:1 slowly:1 itti:1 fixating:1 potential:3 account:1 de:3 suggesting:1 intracortical:2 coding:5 mv:1 onset:8 vi:5 depends:3 view:4 observer:2 competitive:6 start:1 parallel:2 masking:1 minimize:1 bright:1 ni:1 variance:3 maximized:1 correspond:1 ofthe:1 miller:1 weak:5 trajectory:1 finer:1 synapsis:13 whenever:1 synaptic:28 attentive:1 naturally:2 recovers:1 static:11 gain:2 carandini:1 emerges:2 amplitude:2 actually:1 feed:21 higher:3 response:17 synapse:1 depressing:3 strongly:5 binocular:1 correlation:1 trust:1 nonlinear:6 mode:3 indicated:1 gray:1 effect:5 concept:1 true:2 hence:1 imprecision:1 read:1 flashed:3 ringach:1 during:5 width:1 self:1 excitation:6 m:3 ini:1 neocortical:1 demonstrate:1 image:1 consideration:1 novel:1 common:1 behaves:1 functional:5 winner:4 exponentially:1 slight:1 resting:1 numerically:1 interpret:1 functionally:1 refer:1 areal:1 surround:1 tuning:13 similarly:1 depressed:1 cortex:10 operating:1 inhibition:2 longer:2 nonnalizing:1 multivariate:1 own:1 recent:1 perspective:1 scenario:2 selectivity:6 certain:1 xe:1 transmitted:1 kock:1 converge:1 period:6 signal:12 dashed:2 sliding:1 multiple:8 technical:1 faster:1 long:2 prediction:6 j3:5 vision:1 poisson:1 kernel:1 achieved:1 cell:7 receive:1 addition:1 justified:1 interval:2 pyramidal:1 schwabe:5 source:1 operate:1 hz:5 legend:1 noting:1 feedforward:5 revealed:2 enough:2 iii:1 geniculo:1 architecture:1 bandwidth:2 idea:1 shift:1 i7:1 whether:2 depress:1 distributing:1 ul:1 peter:1 depression:17 modulatory:1 proportionally:1 detailed:1 amount:3 http:1 inhibitory:3 shifted:1 dotted:2 correctly:2 per:1 broadly:3 key:1 salient:2 four:1 threshold:1 boehringer:1 varela:1 abbott:1 respond:3 somers:1 vivid:1 polarized:1 curro:1 layer:2 pay:1 followed:1 refine:1 activity:12 strength:6 scanned:1 ahead:1 constraint:1 sharply:7 scene:1 sake:1 generates:1 aspect:1 membrane:4 postsynaptic:1 rev:1 teff:1 invariant:5 pr:1 depletion:1 wellcome:1 resource:2 count:1 mechanism:3 german:1 letting:1 available:5 gaussians:1 operation:1 recurrency:1 hat:1 assumes:1 xc:1 testable:1 establish:1 classical:4 leakage:1 objective:1 question:2 already:2 spike:5 added:1 strategy:5 primary:5 receptive:6 striate:1 responds:2 obermayer:3 exhibit:3 sci:4 berlin:5 capacity:1 nelson:2 exc:2 presynaptic:2 argue:1 enforcing:1 code:2 sur:1 modeled:1 ratio:2 regulation:1 statement:1 gk:1 obennayer:3 info:1 rise:2 priebe:1 motivates:1 observation:2 neuron:6 precise:1 inh:1 sharp:3 intensity:9 hypercolumn:2 required:1 connection:9 established:2 macaque:1 address:1 able:1 bar:16 suggested:2 below:2 pattern:1 lund:1 regime:4 sparsity:3 max:3 natural:1 dei:1 gz:1 extract:1 review:1 relative:1 expect:1 limitation:1 proportional:1 foundation:1 metabolic:1 excitatory:11 summary:2 supported:1 last:1 free:2 bias:1 allow:1 weaker:1 wide:1 markram:1 emerge:1 sparse:2 distributed:1 epigenomics:3 curve:2 feedback:1 cortical:48 calculated:2 cumulative:2 sensory:1 forward:21 approximate:1 obtains:2 preferred:1 deg:5 active:2 investigating:1 hubel:1 assumed:1 factorizing:1 decade:1 why:1 nature:1 channel:1 transfer:3 robust:2 symmetry:1 obtaining:1 investigated:2 troyer:1 main:1 linearly:1 neurosci:5 whole:2 noise:23 arise:1 profile:2 gmbh:1 fig:12 levitt:1 fashion:2 slow:1 cooper:1 fails:3 sub:3 rjj:4 comput:1 breaking:1 saturate:1 specific:1 ingelheim:1 modulates:1 nat:1 fonds:1 sparser:1 simply:1 visual:15 expressed:1 saccade:1 applies:1 goal:1 presentation:1 consequently:1 towards:1 content:1 change:5 experimentally:1 typical:1 operates:3 reducing:1 mexican:1 experimental:1 la:1 internal:1 modulated:1 arises:1 dept:1 baddeley:1 phenomenon:2
740
1,677
The Entropy Regularization Information Criterion Alex J. Smola Dept. of Engineering and RSISE Australian National University Canberra ACT 0200, Australia Alex.Smola@anu.edu.au John Shawe-Taylor Royal Holloway College University of London Egham, Surrey 1W20 OEX, UK john@dcs.rhbnc.ac.uk Bernhard Scholkopf Microsoft Research Limited St. George House, 1 Guildhall Street Cambridge CB2 3NH bsc@microsoft.com Robert C. Williamson Dept. of Engineering Australian National University Canberra ACT 0200, Australia Bob. Williamson @anu.edu.au Abstract Effective methods of capacity control via uniform convergence bounds for function expansions have been largely limited to Support Vector machines, where good bounds are obtainable by the entropy number approach. We extend these methods to systems with expansions in terms of arbitrary (parametrized) basis functions and a wide range of regularization methods covering the whole range of general linear additive models. This is achieved by a data dependent analysis of the eigenvalues of the corresponding design matrix. 1 INTRODUCTION Model selection criteria based on the Vapnik-Chervonenkis (VC) dimension are known to be difficult to obtain, worst case, and often not very tight. Yet they have the theoretical appeal of providing bounds, with few or no assumptions made. Recently new methods [8, 7, 6] have been developed which are able to provide a better characterization of the complexity of function classes than the VC dimension, and moreover, are easily obtainable and take advantage of the data at hand (i.e. they employ the concept of luckiness). These techniques, however, have been limited to linear functions or expansions of functions in terms of kernels as happens to be the case in Support Vector (SV) machines. In this paper we show that the previously mentioned techniques can be extended to expansions in terms of arbitrary basis functions, covering a large range of practical algorithms such as general linear models, weight decay, sparsity regularization [3], and regularization networks [4]. 343 The Entropy Regularization Information Criterion 2 SUPPORT VECTOR MACHINES Support Vector machines carry out an effective means of capacity control by minimizing a weighted sum of the training error (1) and a regularization term Q[J] = ~llwI12; i.e. they minimize the regularized risk functional Rreg[J] 1 := Remp[f] A m + AQ[f] = m ~ C(Xi, Yi, f(Xi)) + "2llwI12 . (2) t=l Here X := {Xl, ... Xm} C X denotes the training set, Y := {YI, ... Ym} C }j the corresponding labels (target values), X, }j the corresponding domains, A > a a regularization constant, C : X X }j X }j -+ JRt a cost function, and f : X -+ }j is given by f(x) := (x, w), or in the nonlinear case f(x) := (4l(x), w). (3) Here 4l : X -+ l' is a map into a feature space 1'. Finally, dot products in feature space can be written as (4l(x), 4l(X')) = k(x, x') where k is a so-called Mercer kernel. For n E N, ~n denotes the n-dimensional space of vectors x = (Xl, ... , Xn). We define spaces as follows: as vector spaces, they are identical to ~n, in addition, they are endowed with p-norms: f; We write fp = fr;:o Furthermore let Ue~ := {x: Ilxlle~ ::; fora < p < 00 forp = 00 I} be the unitf;-baU. For model selection purposes one wants to obtain bounds on the richness of the map Sx Sx : w f-t (f(xd, ... , f(xm)) = ((4l(xd, w), ... , (4l(xm), w)). (4) where w is restricted to an f2 unit ball of some radius A (this is equivalent to choosing an appropriate value of A - an increase in A decreases A and vice versa). By the "richness" of Sx specificaUy we mean the ?-covering numbers N( ?, SX (AUe;;, ), f1:J of the set Sx(AUlm). In the standard COLT notation, we mean p f: N(?, SX(AUl;;')' f:) := min { n There exists a set {Zl, ... zn} C F such that for all Z E Sx(AUem) we have min liz - zililm < ? p l::;i::;n 00 } - See [8] for further details. When carrying out model selection in this case, advanced methods [6] exploit the distribution of X mapped into feature space 1', and thus of the spectral properties of the operator Sx by analyzing the spectrum of the Gram matrix G = [gij]ij, where gij := k(Xi, Xj). All this is possible since k(Xi,Xj) can be seen as a dot product of Xi,Xj mapped into some feature space 1', i.e. k(Xi, Xj) = (4l(Xi), 4l(Xj )) . This property, whilst true for SV machines with Mercer kernels, does not hold in general case where f is expanded in terms of more or less arbitrary basis functions. 344 A. J. Smola. J. Shawe-Taylor, B. Sch61kopf and R. C. Williamson 3 THE BASIC PROBLEMS One basic problem is that when expanding 1 into n (5) i=l with Ii (x) being arbitrary functions, it is not immediately obvious how to regard 1 as a dot product in some feature space. One can show that the VC dimension of a set of n linearly independent functions is n. Hence one would intuitively try to restrict the class of admissible models by controlling the number of basis functions n in terms of which 1 can be expanded. Now consider an extreme case. In addition to the n basis functions Ii defined previously, we are given n further basis functions II, linearly independent of the previous ones, which differ from Ii only on a small domain X', i.e. Iilx\x = IIlx\xl. Since this new set of functions is linearly independent, the VC dimension of the joint set is given by 2n. On the other hand, if hardly any data occurs on the domain X', one would not notice the difference between Ii and II. In other words, the joint system of functions would behave as if we only had the initial system of n basis functions. 1 An analogous situation occurs if II = Ii + ?gi where ? is a small constant and gi was bounded, say, within [0, 1J. Again, in this case, the additional effect of the set offunctions II would be hardly noticable, but still, the joint set of functions would count as one with VC dimension 2n. This already indicates, that simply counting the number of basis functions may not be a good idea after all. .' ''~ Figure 1: From left to right: (a) initial set of functions h, ... , 15 (dots on the x-axis indicate sampling points); (b) additional set of functions IL ... , I~ which differ globally, but only by a small amount; (c) additional set offunctions IL ... , I~ which differ locally, however by a large amount; (d) spectrum of the corresponding design matrices - the bars denote the cases (a)-(c) in the corresponding order. Note that the difference is quite small. On the other hand, the spectra of the corresponding design matrices (see Figure 1) are very similar. This suggests the use of the latter for a model selection criterion. Finally we have the practical problem that capacity control, which in SV machines was carried out by minimizing the length of the "weight vector" w in feature space, cannot be done in an analogous way either. There are several ways to do this. Below we consider three that have appeared in the literature and for which there exist effective algorithms. Example 1 (Weight Decay) Define Q[IJ := ~ L:i ar . i.e. the coefficients ai of the junction expansion are constrained to an ?2 ball. In this case we can consider the following where operator S(1)? X . ?n 2 -t ?m 00' Sr): aM (f(xd, ... , I(x m )) = ((f(Xl), a), . .. , (f(Xm), a)) = Fa (6) Here I(x):= Ul(x) , .. ?In(x)), Fij := Ii(Xj), a'- (al, ... ,an ) and a E AUl'2for some A> O. 345 The Entropy Regularization Information Criterion Example 2 (Sparsity Regularization) In this case Q[J] := Li lail, i.e. the coefficients ai of the function expansion are constrained to an ?1 ball to enforce sparseness [3]. Thus sC;) :?1 -t ?~ with sC;) mapping a as in (6) except a E AUlI. This is similar to expansions encountered in boosting or in linear programming machines. Example 3 (Regularization Networks) Finally one could set Q[J] := ~a T Qa for some positive definite matrix Q. For instance, Qij could be obtainedfrom (Ph, P fj) where P is a regularization operator penalizing non-smooth functions [4J. In this case a lives inside some n-dimensional ellipsoid. By substituting a' := Q% a one can reduce this setting to the case of example 1 with a different set of basis functions (f'(x) = Q-% f(x)) and consider an evaluation operator s~) : ?2 s~): a' f-+ (f(xd, . .. , f(xm)) -t ?: given by = ((Q-% f(X1), a'), . .. , (Q-t f(xm), a')) = Q-t Fa' (7) where a' E AUl2 for some A> 0 and Fij = fi(xj) as in example 1. Example 4 (Support Vector Machines) An important special case of example 3 are Support Vector Machines where we have Qij = k(Xi,Xj) andfi(x) = k(Xi,X), henceQ = F. Hence the possible values generated by a Support Vector Machine can be written as s~): a' f-+ (f(X1), ... , f(xm)) where a' E AUl2 for some A = ((Q-% f(xd, a'), . .. , (Q-% f(xm), a')) = Ft a' (8) > o. 4 ENTROPY NUMBERS Covering numbers characterize the difficulty of learning elements of a function class. Entropy numbers of operators can be used to compute covering numbers more easily and more tightly than the traditional techniques based on VC-like dimensions such as the fat shattering dimension [1]. Knowing el (S x) = ? (see below for the definition) tells one that 10g:N(? , F,?~) ::; I, where F is the effective class of functions used by the regularised learning machines under consideration. In this section we summarize a few basic definitions and results as presented in [8] and [2]. The lth entropy number ?l (F) of a set F with a corresponding metric d is the precision up t~ whicI:! F can _be approximated by 1 elements of F; i.e. for all f E F there exists some fi E {h, ? ??, fd such that d(f, fi) ::; ?l. Hence ?1(F) is the functional inverse of the covering number of F. The entropy number of an bounded linear operator T: A -t B between normed linear spaces A and B is defined as ?1(T) := ?1(T(UA)) with the metric d being induced by II . liB. The dyadic entropy numbers el are defined by el := ?2'+1 (the latter quantity is often more convenient to deal with since it corresponds to the log of the covering number). We make use of the following three results on entropy numbers of the identity mapping from ?;1 into ?;2' diagonal operators, and products of operators. Let id;l ,P2 : ?;1 -t ?;2 ; id;1 ,P2 : x f-+ x The following result is due to Schlitt; the constants 9.94 and 1.86 were obtained in [9]. Proposition 1 (Entropy numbers for identity operators) Be mEN. Then el(id~,2) ::; 9.94 (t log (1 + T) ) 1 2 1 & el (id~,(xJ ::; 1.86 (t log (1 + T) )2 (9) A. J Smola, J Shawe-Taylor, B. SchOlkopfand R. C. Williamson 346 Proposition 2 (Carl and Stephani [2, p.11]) Let E, F, G be Banach spaces, R : F -+ G, and S: E -+ F. Then,forn, tEN, en+t-l (RS) ~ en(R)et(S), en(RS) ~ en (R)IISII and en(RS) ~ Note that the latter two inequalities follow directly from the fact that R: F -+ G by definition of the operator norm IIRII. Proposition 3 Let 0"1 ~ 0"2 ~ . .. ~ O"j ~ . .. ~ 0, 1 ~ p ~ 00 en(S)IIRII. ?l (R) (to) = IIRIlfor all and (11) for x = (Xl, X2, ... , Xj, . .. ) E f!p be the diagonal operator from f!p into itself, generated by the sequence (0" j ) j. Then for all n E N, 5 THE MAIN RESULT We can now state the main theorem which gives bounds on the entropy numbers of S~) for the first three examples of model selection described above (since Support Vector Machines are a special case of example 3 we will not deal with it separately). Proposition 4 Let! be expanded in a linear combination of basis functions as ! .L~=l adi and the coefficients a restricted to one of the convex sets as described in the examples 1 to 3. Moreover denote by Fij := !j(Xi) the design matrix on a particular sample X, and by Q the regularization matrix in the case of example 3. Then the following bound on Sx holds. 1. In the case of weight decay (ex. 1)(with h el(S~)) ~ + l2 ~ l + 1) 1.96 (llllog(1 +m/h))t 2. 1n the case of weight sparsity regularization (ex. 2) (with el(S~)) ~ 18.48 (lillog (1 + m/h)) t eI2(~)' h + l2 + l3 el 2 (~) (l3'llog (1 3. Finally, in the case of regularization networks (ex. 3) (with II el (Sr)) ~ 1.96 (lillog (1 (13) l + 2) + m/l3)) t. + l2 + m/h)) t el (~). 2 ~ ~ l (14) + 1) (15) Here ~ is a diagonal scaling operator (matrix) with (i, i) entries .j(ii and (.j(ii)i are the eigenvalues (sorted in decreasing order) of the matrix FFT in the case of examples 1 and 2, and FQ-l FT in the case of example 3. The entropy number of ~ is readily bounded in terms of (O"i)i by using (3). One can see that the first setting (weight decay) is a special case of the third one, namely when Q = 1, i.e. when Q is just the identity matrix. Proof The proofrelies on a factorization of S~) (i = 1,2,3) in the following way. First we consider the equivalent operator Sx mapping from f!~ to and perform a singular value decomposition [5] of the latter into Sx = V~W where V, W are operators of norm 1, and ~ contains the singular values of S~), i.e. the singular values of F and FQ- t f!r 347 The Entropy Regularization Information Criterion respectively. The latter, however, are identical to the square root of the eigenvalues of F FT or FQ-l FT. Consequently we can factorize S~) as in the diagram (16) Finally, in order to compute the entropy number of the overall operator one only has to use the factorization of Sx into S~) = id~oo VL:W for i E {1,3} and into S~) = id~oo VL:Wid~, 2 for example 2, and apply Proposition 2 several times. We also exploit l. ? the fact that for singular value decompositions IIVI\' IIWII s The present theorem allows us to compute the entropy numbers (and thus the complexity) of a class of functions on the current sample X. Going back to the examples of section 3, which led to large bounds on the VC dimension one can see that the new result is much less susceptible to such modifications: the addition of f{. ... f~ to h, ... f n does not change the eigenspectrum L: of the design matrix significantly (possibly only doubling the nominal value of the singular values), if the functions fi differ from fi only slightly. Consequently also the bounds will not change significantly even though the number of basis functions just doubled. Also note that the current error bounds reduce to the results of [6] in the SV case: here Q ij = Fij = k( Xi, X j) (both the design matrix F and the regularization matrix Q are determined by kernels) and therefore FQ-l F = Q. Thus the analysis of the singular values of FQ-l F leads to an analysis of the eigenvalues of the kernel matrix, which is exactly what is done when dealing with SV machines. 6 ERROR BOUNDS To use the above result we need a bound on the expected error of a hypothesis f in terms of the empirical error (training error) and the observed entropy numbers ?n(J'). We use [6, Theorem 4.1] with a small modification. Theorem 1 Let:1' be a set of linear junctions as described in the previous examples with en(Sx) as the corresponding bound on the observed entropy numbers of:1' on the dataset X. Moreover suppose thatforafixed threshold b E [?for some f E :1', sgn(f - b) correctly classifies the set X with a margin 'Y := minlSiSm If(Xi) - bl. t). Finally let U := min{ n E N with en(Sx) s 'Y /8.001} and a(U, <5) := 3.08(1 + bIn Then with confidence 1- <5 over X (drawn randomly from pm where P is some probability distribution) the expected error ofsgn(f - b) is boundedfrom above by ?(m,U,<5) =! (U(1+a(U,~)log(5t-m)log(17m)) + log (l6r)) . (17) The proof is essentially identical to that of [6, Theorem 4.1] and is omitted. [6] also shows how to compute en (S x) efficiently including an explicit formula for evaluating el (L:). 7 DISCUSSION We showed how improved bounds could be obtained on the entropy numbers of a wide class of popular statistical estimators ranging from weight decay to sparsity regularization 348 A. J Smola. J Shawe- Taylor, B. SchOllropf and R. C. Williamson (with SV machines being a special case thereof). The results are given in a way that is directly useable for practicioners without any tedious calculations of the VC dimension or similar combinatorial quantities. In particular, our method ignores (nearly) linear dependent basis functions automatically. Finally, it takes advantage of favourable distributions of data by using the observed entropy numbers as a base for stating bounds on the true entropy numbers with respect to the function class under consideration. Whilst this leads to significantly improved bounds (we achieved an improvement of approximately two orders of magnitude over previous VC-type bounds involving only the radius of the data R and the weight vector IIwll in the experiments) on the expected risk, the bounds are still not good enough to become predictive. This indicates that possibly rather than using the standard uniform convergence bounds (as used in the previous section) one might want to use other techniques such as a PAC-Bayesian treatment (as recently suggested by Herbrich and Graepel) in combination with the bounds on eigenvalues of the design matrix. Acknowledgements: This work was supported by the Australian Research Council and a grant of the Deutsche Forschungsgemeinschaft SM 62/1-1. References [1] N. Alon, S. Ben-David, N. Cesa-Bianchi, and D. Haussler. Scale-sensitive Dimensions, Uniform Convergence, and Learnability. 1. of the ACM, 44(4):615-631,1997. [2] B. Carl and I. Stephani. Entropy, compactness, and the approximation of operators. Cambridge University Press, Cambridge, UK, 1990. [3] S. Chen, D. Donoho, and M. Saunders. Atomic decomposition by basis pursuit. Technical Report 479, Department of Statistics, Stanford University, 1995. [4] F. Girosi, M. Jones, and T. Poggio. Regularization theory and neural networks architectures. Neural Computation, 7:219-269,1995. [5] R. A. Horn and C. R. Johnson. Matrix Analysis. Cambridge University Press, Cambridge, 1992. [6] B. Scholkopf, J. Shawe-Taylor, A. J. Smola, and R. C. Williamson. Generalization bounds via eigenvalues of the gram matrix. Technical Report NC-TR-99-035, NeuroColt2, University of London, UK, 1999. [7] J. Shawe-Taylor and R. C. Williamson. Generalization performance of classifiers in terms of observed covering numbers. In Proc. EUROCOLT'99, 1999. [8] R. C. Williamson, A. J. Smola, and B. Scholkopf. Generalization performance of regularization networks and support vector machines via entropy numbers of compact operators. NeuroCOLT NC-TR-98-019, Royal Holloway College, 1998. [9] R. C. Williamson, A. J. Smola, and B. SchOlkopf. A Maximum Margin Miscellany. Typescript, 1999.
1677 |@word rreg:1 norm:3 tedious:1 r:3 decomposition:3 tr:2 carry:1 initial:2 contains:1 chervonenkis:1 current:2 com:1 yet:1 written:2 readily:1 john:2 additive:1 girosi:1 offunctions:2 characterization:1 boosting:1 herbrich:1 become:1 scholkopf:4 qij:2 inside:1 expected:3 globally:1 decreasing:1 eurocolt:1 automatically:1 ua:1 lib:1 classifies:1 moreover:3 notation:1 bounded:3 deutsche:1 what:1 developed:1 whilst:2 act:2 xd:5 fat:1 exactly:1 classifier:1 uk:4 zl:1 control:3 unit:1 grant:1 positive:1 engineering:2 analyzing:1 id:6 approximately:1 might:1 au:2 suggests:1 limited:3 factorization:2 range:3 practical:2 horn:1 atomic:1 definite:1 useable:1 cb2:1 sch61kopf:1 empirical:1 significantly:3 convenient:1 word:1 confidence:1 doubled:1 cannot:1 selection:5 operator:17 risk:2 equivalent:2 map:2 forn:1 normed:1 convex:1 immediately:1 estimator:1 haussler:1 analogous:2 target:1 controlling:1 nominal:1 suppose:1 programming:1 carl:2 regularised:1 hypothesis:1 element:2 approximated:1 observed:4 ft:4 worst:1 richness:2 decrease:1 mentioned:1 complexity:2 carrying:1 tight:1 predictive:1 ei2:1 f2:1 basis:13 easily:2 joint:3 effective:4 london:2 sc:2 tell:1 choosing:1 saunders:1 quite:1 stanford:1 say:1 statistic:1 gi:2 itself:1 advantage:2 eigenvalue:6 sequence:1 product:4 fr:1 llwi12:2 convergence:3 ben:1 oo:2 alon:1 ac:1 stating:1 ij:3 p2:2 indicate:1 australian:3 differ:4 radius:2 fij:4 vc:9 wid:1 australia:2 sgn:1 bin:1 f1:1 generalization:3 proposition:5 hold:2 liz:1 guildhall:1 aul:2 mapping:3 substituting:1 omitted:1 purpose:1 proc:1 label:1 combinatorial:1 council:1 sensitive:1 vice:1 weighted:1 bsc:1 rather:1 improvement:1 indicates:2 fq:5 am:1 dependent:2 el:11 vl:2 compactness:1 stephani:2 going:1 overall:1 colt:1 constrained:2 special:4 sampling:1 identical:3 shattering:1 jones:1 nearly:1 report:2 few:2 employ:1 randomly:1 national:2 tightly:1 iivi:1 microsoft:2 practicioners:1 fd:1 evaluation:1 jrt:1 extreme:1 poggio:1 taylor:6 oex:1 forp:1 theoretical:1 instance:1 ar:1 zn:1 cost:1 entry:1 uniform:3 johnson:1 learnability:1 characterize:1 sv:6 st:1 ym:1 again:1 cesa:1 possibly:2 li:1 coefficient:3 try:1 root:1 minimize:1 il:2 square:1 largely:1 efficiently:1 bayesian:1 bob:1 definition:3 surrey:1 obvious:1 thereof:1 proof:2 dataset:1 treatment:1 popular:1 remp:1 lail:1 graepel:1 obtainable:2 back:1 follow:1 improved:2 done:2 though:1 furthermore:1 just:2 smola:8 hand:3 nonlinear:1 effect:1 concept:1 true:2 regularization:19 hence:3 deal:2 ue:1 covering:8 criterion:6 fj:1 ranging:1 consideration:2 recently:2 fi:5 functional:2 nh:1 banach:1 extend:1 typescript:1 cambridge:5 versa:1 ai:2 pm:1 shawe:6 aq:1 dot:4 had:1 l3:3 base:1 showed:1 inequality:1 life:1 yi:2 seen:1 george:1 additional:3 bau:1 ii:14 smooth:1 technical:2 calculation:1 involving:1 basic:3 essentially:1 metric:2 kernel:5 achieved:2 addition:3 want:2 separately:1 diagram:1 singular:6 sr:2 induced:1 counting:1 forschungsgemeinschaft:1 enough:1 fft:1 xj:10 architecture:1 restrict:1 reduce:2 idea:1 knowing:1 rhbnc:1 ul:1 hardly:2 amount:2 locally:1 ph:1 ten:1 exist:1 notice:1 correctly:1 write:1 threshold:1 drawn:1 penalizing:1 sum:1 inverse:1 scaling:1 bound:20 encountered:1 alex:2 x2:1 min:3 expanded:3 department:1 ball:3 combination:2 slightly:1 iiwii:1 modification:2 happens:1 iisii:1 intuitively:1 restricted:2 previously:2 count:1 junction:2 pursuit:1 endowed:1 noticable:1 apply:1 appropriate:1 egham:1 spectral:1 enforce:1 denotes:2 exploit:2 w20:1 forum:1 bl:1 already:1 quantity:2 occurs:2 neurocolt2:1 fa:2 traditional:1 diagonal:3 mapped:2 neurocolt:1 capacity:3 street:1 parametrized:1 eigenspectrum:1 length:1 ellipsoid:1 providing:1 minimizing:2 nc:2 difficult:1 susceptible:1 robert:1 design:7 perform:1 bianchi:1 sm:1 behave:1 situation:1 extended:1 dc:1 auli:1 arbitrary:4 david:1 namely:1 rsise:1 qa:1 able:1 bar:1 suggested:1 below:2 xm:8 fp:1 sparsity:4 appeared:1 summarize:1 royal:2 including:1 difficulty:1 regularized:1 advanced:1 axis:1 carried:1 literature:1 l2:3 acknowledgement:1 men:1 mercer:2 supported:1 wide:2 regard:1 dimension:10 xn:1 gram:2 evaluating:1 ignores:1 made:1 miscellany:1 compact:1 bernhard:1 dealing:1 xi:12 factorize:1 spectrum:3 expanding:1 adi:1 expansion:7 williamson:9 domain:3 iiwll:1 main:2 linearly:3 whole:1 dyadic:1 x1:2 canberra:2 en:9 precision:1 explicit:1 xl:5 house:1 third:1 admissible:1 theorem:5 luckiness:1 formula:1 pac:1 favourable:1 appeal:1 decay:5 exists:2 vapnik:1 magnitude:1 anu:2 sx:14 sparseness:1 margin:2 chen:1 entropy:23 led:1 simply:1 doubling:1 corresponds:1 acm:1 lth:1 identity:3 sorted:1 consequently:2 donoho:1 change:2 determined:1 except:1 llog:1 called:1 gij:2 holloway:2 college:2 support:9 latter:5 dept:2 ex:3
741
1,678
Information Factorization in Connectionist Models of Perception Javier R. Movellan Department of Cognitive Science Institute for Neural Computation University of California San Diego James L. McClelland Center for the Neural Bases of Cognition Department of Psychology Carnegie Mellon University Abstract We examine a psychophysical law that describes the influence of stimulus and context on perception. According to this law choice probability ratios factorize into components independently controlled by stimulus and context. It has been argued that this pattern of results is incompatible with feedback models of perception. In this paper we examine this claim using neural network models defined via stochastic differential equations. We show that the law is related to a condition named channel separability and has little to do with the existence of feedback connections. In essence, channels are separable if they converge into the response units without direct lateral connections to other channels and if their sensors are not directly contaminated by external inputs to the other channels. Implications of the analysis for cognitive and computational neurosicence are discussed. 1 Introduction We examine a psychophysical law, named the Morton-Massaro law, and its implications to connectionist models of perception and neural information processing. For an example of the type of experiments covered by the Morton-Massaro law consider an experiment by Massaro and Cohen (1983) in which subjects had to identify synthetic consonant sounds presented in the context of other phonemes. There were two response alternatives, seven stimulus conditions, and four context conditions. The response alternatives were /1/ and /r/, the stimuli were synthetic sounds generated by varying the onset frequency of the third formant, followed by the vowel /i/. Each of the 7 stimuli was placed after each offour different context consonants, /v/, /s/, /p/, and /t/. Morton (1969) and Massaro independently showed that in a remarkable range of experiments of this type, the influence of stimulus and context on response probabilities can be accounted for with a factorized version of Luce's strength model (Luce, 1959) P(R = k I S = i, C = j) Tls(i, k) Tlc(j, k) 2: I TIs C1" l) TIc), C l)' . . for (l,,),k) E S x ex 'R. (1) Here S, C and R are random variables representing the stimulus, context and the subject's response, S, C and 'R are the set of stimulus, context and response al- J. R. Movellan and J. L. McClelland 46 ternatives, l1s(i, k) > 0 represents the support of stimulus i for response k, and l1c(j, k) > 0 the support of context j for response k. Assuming no strength parameter is exactly zero, (1) is equivalent to P(R P(R = k I S = i,e = j) = '(l1S(i,k)) = II S = i, e = j) l1s(i, l) (l1c(j,k)) , for all (i,j,k) E S x ex R. l1c(j, l) (2) This says that response probability ratios factorize into two components, one which is affected by the stimulus but unaffected by the context and one affected by the context but unaffected by the stimulus. 2 Diffusion Models of Perception Massaro (1989) conjectured that the Morton-Massaro law may be incompatible with feedback models of perception. This conjecture was based on the idea that in networks with feedback connections the stimulus can have an effect on the context units and the context can have an effect on the stimulus units making it impossible to factorize the influence of information sources. In this paper we analyze such a conjecture and show that, surprisiQ.gly, the Morton-Massaro law has little to do with the existence of feedback and lateral connections. We ground our analysis on continuous stochastic versions of recurrent neural networks 1. We call these models diffusion (neural) networks for they are stochastic diffusion processes defined by adding Brownian motion to the standard recurrent neural network dynamics. Diffusion networks are defined by the following stochastic differential equation dYi(t) = JLi(Y(t), X) dt + dBi(t) (J for i E {I, ... , n}, (3) where Yi(t) is a random variable representing the internal potential at time t of the unit, Y(t) = (Yl(t),??? ,Yn(t))', X represents the external input, which consists of stimulus and context, and Bi is Brownian motion, which acts as a stochastic driving term. The constant (J > 0, known as the dispersion, controls the amount of noise injected onto each unit. The function JLi, known as the drift, determines the average instantaneous change of activation and is borrowed from the standard recurrent neural network literature: this change is modulated by a matrix w of connections between units, and a matrix v that controls the influence of the external inputs onto each unit. ith JLi(Yi(t), X) 1 - = ~i(Yi(t)) (Yi(t) - Yi(t)), for all i E {I,??? , n}, (4) where 1/ ~i is a positive function, named the capacitance, controlling the speed of processing and Yi(t) = L j Zj(t) Wi,j Zj(t) + LVi,kXk, for alli E {I, .. ? ,n}, (5) k = CPi(}j(t)) = CP(O!i }j(t)) = 1/(1 + e- a ? Y;(t)). (6) Here Wi ,j, an element of the connection matrix w, is the weight from unit j to unit i, Vi,k is an element of the matrix v, cP is the logistic activation function and the O!i > 0 terms are gain parameters, that control the sharpness of the activation functions. For large values of O!i the activation function of unit i converges to a step function. The variable Zj(t) represents a short-time mean firing rate (the activation) of unit lFor an analysis grounded on discrete time networks with binary states see McClelland (1991). 47 Information Factorization j scaled in the (0,1) range. Intuition for equation (4) can be achieved by thinking of it as a the limit of a discrete time difference equation, in such case Y(t + ~t) = Yi(t) + J.'i (Yi (t), X)~t + (rli5:tNi (t), (7) where the Ni(t) are independent standard Gaussian random variables. For a fixed state at time t there are two forces controlling the change in activation: the drift, which is deterministic, and the dispersion which is stochastic. This results in a distribution of states at time t + ~t. As ~t goes to zero, the solution to the difference equation (7) converges to the diffusion process defined in (4). In this paper we focus on the behavior of diffusion networks at stochastic equilibrium, i.e., we assume the network is given enough time to approximate stochastic equilibrium before its response is sampled. 3 Channel Separability In this section we show that the Morton-Massaro is related to an architectural constraint named channel separability, which has nothing to do with the existence of feedback connections. In order to define channel separability it is useful to characterize the function of different units using the following categories: 1) Response specification units: A unit is a response specification unit, if, when the state of all the other units in the network is fixed, changing the state of this unit affects the probability distribution of overt responses. 2) Stimulus units: A unit belongs to the stimulus channel if: a) it is not a response unit, and b) when the state of the response units is fixed, the probability distribution of the activations of this unit is affected by the stimulus. 3) Context units: A unit belongs to the context channel if: a) it is not a response unit, and b) when the states of the response units are fixed, the probability distribution of the activations of this unit can be affected by the context. Given the above definitions, we say that a network has separable stimulus and context channels if the stimulus and context units are disjoint: no unit simultaneously belongs to the stimulus and context channels. In essence, channels are structurally separable if they converge into the response units without direct lateral connections to other channels and if their sensors are not directly contaminated by external inputs to the other channels (see Figure 1). In the rest of the paper we show that if a diffusion network is structurally separable the Morton-Massaro law can be approximated with arbitrary precision regardless of the existence of feedback connections. For simplicity we examine the case in which the weight matrix is symmetric. In such case, each state has an associated goodness function that greatly simplifies the analysis. In a later section we discuss how the results generalize to the non-symmetric case. Let y E IRn represent the internal potential of a diffusion network. Let Zi = cp(aiYi) for i = 1,??? , n represent the firing rates corresponding to y. Let zS, ZC and zr represent the components of z for the units in the stimulus channel, context channel and response specification module. Let x be a vector representing an input and let x S , XC be the components of x for the external stimulus and context. Let a = (a1,??? , an) be a fixed gain vector and ZO/(t) a random vector representing the firing rates at time t of a network with gain vector a. Let = limt-+oo (t), represent the firing rates at stochastic equilibrium. In Movellan (1998) it is shown that if the weights are symmetric i.e., W = w' and l/Ki(x) = dcpi(X)/dx then the equilibrium probability density of is as follows za za za PZQlx(zs,zc,zr I XS,X C ) = K ( 1 a Xs,Xc ) exp((2/0'2) Ga(zs,zr I XS,X C )) , (8) 48 J. R. Movellan and J. L. McClelland /CoDtut SdmU~ Input Figure 1: A network with separable context and stimulus processing channels. The stimulus sensor and stimulus relay units make up the stimulus channel units, and the context sensor and context channel units make up the context channel units. Note that any of the modules can be empty except the response module. where Ka(x s , xc) = / exp((2/(72) Ga(z I Xs , xc)) dz, (9) n Ga(z I x) = H(z I x) - L Sa; (Zi), (10) i=l H(z I x) Sa; (Zi) = z' w z/2 + z' V x, (11) = ai (IOg(Zi) + log(1 - Zi)) + ~i (Zi log(zi) + (1 - Zi) log(1 - Zi)) . (12) Without loss of generality hereafter we set (72 = 2. When there are no direct connections between the stimulus and context units there are no terms in the goodness function in which XS or ZS occur jointly with XC or ZC. Because of this, the goodness can be separated into three additive terms, that depend on x S , XC and a third term which depends on the response units: Ga(z\zc,zr I XS,X C) = G~(zs,zr I X + G~(zr,zc I XC) + G~(zr) S) where G~(ZS, zr I XS) = (zs),w s,szs/2 + (zS)'ws,rzr + (ZS)'vs,sx s + (zr),vr,sx s - , (13) L S(zt) , i (14) G~(ZC, zr I XS ) = (ZC)'wc,czc /2 + (zc),wc,rzr + (zc),vc,cx c + (zr),vr,cx c - L S(zf) , i (15) (16) 49 Information Factorization where ws,r is a submatrix of w connecting the stimulus and response units. Similar notation is used for the other submatrices of wand v. It follows that we can write the ratio of the jOint probability density' of two states z and z as follows: PZ.. lx(zs,zc,zr I XS,X C ) exp(G~(zS,zr I x s ) + G~(zc,zr I XC) + G~(zr? (17) pZ.. lx(zS,zC,zrlxs,x c ) - exp(G~(zS,zrlxs)+G~(zC,zrlxC)+G~(zr?' which factorizes as desired. To get probability densities for the response units, we integrate over the states of all the other units PZ;;IX(zr I XS,X C) = / / PZ.. lx(zs,zc,zr I XS,X C) dz s dz c , (18) and after rearranging terms pZ;;IX(zr I XS,X C) = Kcr(:s,x C) ( / exp( Gz(zs,zr I XS) + Gr(zr? dZ S) (19) ( / exp( G c(ZC, zr I xc? dZ C) , which also factorizes. All is left is mapping continuous states of the response units to discrete external responses. To do so we partition the space of the response specification units into discrete regions. The probability of a response becomes the integral of the probability density over the region corresponding to that response. The problem is that the integral of probability densities does not necessarily factorize even though the densities factorize at every point. Fortunately there are two important cases for which the law holds, at least as a good approximation. The first case is when the response regions are small and thus we can approximate the integral over that region by the density at a point times the volume of the region. In such a case the ratio of the integrals can be approximated by the ratio of the probability densities of those individual states. The second case applies to models, like McClelland and Rumelhart's (1981) interactive activation model, in which each response is associated with a distinct response unit. These models typically have negative connections amongst the response units so that at equilibrium one unit tends to be active while the others are inactive. In such a case a common response policy picks the response corresponding to the active unit. We now show that such a policy can approximate the Morton-Massaro law to an arbitrary level of precision as the gain parameter of the response units is increased. Let z represent the joint state of a network and let the first r components of z be the states of the response specification units. Let z(1) = (1,0,0, ... ,0)', Z(2) = (0,1,0,??? ,0)' be two r-dimensional vectors representing states of the response specification units. For i E {1,2} and ~ E (0,1) let z~) = (1 - Z(i?~ R~) = {x E IRr + (z(i?(l - : Xj ~), E ((1- ~)Z~i), ~ (20) + (1 - ~)Z~i?, for j = 1,??? , r}. (21) The sets R~) and R~) are regions of the [O,l]r space mapping into two distinct external responses. We now investigate the convergence of the probability ratio of these two responses as we let ~ 4 0, i.e., as the response regions collapse into corners of [0, l]r. I X = x) J PzrlX(u I x)du = = lim Rt . " A-+O P(Z~ E R~) I X = x) A-+O JR~) PZ;;lx(u I x)du . ~rpZ;;IX(z~) I x) . J J eG~(z~),z?,ze I z}dz dz c hm (1) = hm A-+O ~rpZ;;IX(zA I x) A-+O J J eG .. (zt. ,Z',ze I z)dz s dz c lim p(zr E cr R(2) A (2) (22) 8 (1) ? (23) J. R. Movellan and J. L. McClelland 50 Table 1: Predictions by the Morton-Massaro law (left side) versus diffusion network (square brackets) for subject 7 of Massaro and Cohen (1983) Experiment 2. Each prediction of the diffusion network is based on 100 random samples. Context Stimulus 0 1 2 3 4 5 6 0.0017 0.0126 0.1105 0.5463 0.9827 0.9999 0.9999 P S V 0.01 0.00 0.19 0.54 1.00 1.00 1.00 0.0000 0.0000 0.0008 0.0079 0.2756 0.9924 0.9924 0.00 0.00 0.00 0.00 0.30 0.99 1.00 0.0152 0.1008 0.5208 0.9133 0.9980 0.9999 0.9999 0.03 0.10 0.45 0.91 1.00 1.00 1.00 T 0.9000 0.9849 0.9984 0.9998 0.9999 1.0000 1.0000 0.91 0.97 1.00 1.00 1.00 1.00 1.00 Now note that r Go(z~), Z8, ZC I x) = H(z~), Z8, ZC I x) - L So; (Z~!i) - L So; (zt) - L SOj (zj), i=1 i j (24) and since E~=l So; (Z~!i) = E;=1 So; (Z~!i)' it follows that . P(Z~ E R~) I X = x) _ hm (1) ~-+o P(Z~ E R~ I X = x) J J eH(z~),z?,ze I x)-E;Sa;(zi)-E j Saj(zj)dz J J eH(za.. ,Z',ze I x)-E; Sai(zt)-E j Saj(Zj)dz 8 dz c 8 dzc (1) ? (25) It is easy to show that this ratio factorizes. Moreover, for all .6. 0:1 = ... = O:r = 0:, where 0: > 0 then lim P(Z~ E [.6.,1 - .6.t) 0-+00 = 0, > 0 if we let (26) since as the gain of the response units increases So; decreases very fast at the corners of (0, 1y. Thus as 0: -4 00 the random variable Z~ converges in distribution to a discrete random variable with mass at the corner of the [0, It hypercube and with factorized probability ratios as expressed on (25). Since the indexing ofthe response units is arbitrary the argument applies to all the responses. o 4 Discussion Our analysis establishes that in diffusion networks the Morton-Massaro law is not incompatible with the presence of feedback and lateral connections. Surprisingly, even though in diffusion networks with feedback connections stimulus and context units are interdependent, it is still possible to factorize the effect of stimulus and context on response probabilities. The analysis shows that the Morton-Massaro can be arbitrarily approximated as the sharpness of the response units is increased. In practice we have found very good approximations with relatively small values of the sharpness parameter (see Table 1 for an example). The analysis assumed that the weights were symmetric. Mathematical analysis of the general case with non-symmetric weights is difficult. Information Factorization 51 However useful approximations exist (Movellan & McClelland, 1995) showing that if the noise parameter (7 is relatively small or if the activation function c.p is approximately linear, symmetric weights are not needed to exhibit the Morton-Massaro law. The analysis presented here has potential applications to investigate models of perception and the functional architecture of the brain. For example the interactive activation model of word perception has a separable architecture and thus, diffusion versions of it adhere to the Morton Massaro law. The analysis also points to potential applications in computational neuroscience. It would be of interest to study whether the Morton-Massaro holds at the level of neural responses. For example, we may excite a neuron with two different sources of information and observe its short term average response to combination of stimuli. If the observed distribution of responses exhibits the Morton-Massaro law, this would be consistent with the existence of separable channels converging into that neuron. Otherwise, it would indicate that the channels from the two input areas to the response may not be structurally separable. References Luce, R. D. (1959). Individual choice behavior. New York: Wiley. Massaro, D. W. (1989). Testing between the TRACE Model and the fuzzy logical model of speech perception. Cognitive Psychology, 21, 398-42l. Massaro, D. W. (1998). Perceiving Talking Faces. Cambridge, Massachusetts: MIT Press. Massaro, D. W. & Cohen, M. M. (1983a). Phonological constraints in speech perception. Perception and Psychophysics, 94, 338-348. McClelland, J. L. (1991). Stochastic interactive activation and the effect of context on perception. Cognitive Psychology, 29, 1-44. Morton, J. (1969). The interaction of information in word recognition. Psychological Review, 76, 165-178. Movellan, J. R. (1998). A Learning Theorem for Networks at Detailed Stochastic Equilibrium. Neural Computation, 10(5), 1157-1178. Movellan, J. R. & McClelland, J. L. (1995) . Stochastic interactive processing, channel separability and optimal perceptual inference: an examination of Morton's law. Technical Report PDP.CNS.95A, Available at http://cnbc.cmu.edu, Carnegie Mellon University.
1678 |@word version:3 pick:1 hereafter:1 rpz:2 kcr:1 ka:1 activation:12 dx:1 additive:1 partition:1 v:1 ith:1 short:2 lx:4 mathematical:1 direct:3 differential:2 consists:1 cnbc:1 behavior:2 examine:4 brain:1 szs:1 little:2 becomes:1 notation:1 moreover:1 factorized:2 mass:1 tic:1 z:15 offour:1 fuzzy:1 every:1 sai:1 ti:1 act:1 interactive:4 exactly:1 scaled:1 control:3 unit:53 yn:1 positive:1 before:1 tends:1 limit:1 firing:4 approximately:1 alli:1 factorization:4 collapse:1 range:2 bi:1 testing:1 practice:1 movellan:8 area:1 submatrices:1 word:2 get:1 onto:2 ga:4 context:31 influence:4 impossible:1 equivalent:1 deterministic:1 center:1 dz:12 go:2 regardless:1 independently:2 sharpness:3 simplicity:1 dbi:1 jli:3 diego:1 controlling:2 massaro:21 element:2 rumelhart:1 approximated:3 ze:4 recognition:1 observed:1 module:3 region:7 decrease:1 intuition:1 dynamic:1 depend:1 joint:2 zo:1 separated:1 distinct:2 fast:1 say:2 otherwise:1 formant:1 jointly:1 gly:1 interaction:1 tni:1 convergence:1 empty:1 converges:3 oo:1 recurrent:3 soj:1 borrowed:1 sa:3 indicate:1 stochastic:12 vc:1 argued:1 hold:2 ground:1 exp:6 equilibrium:6 cognition:1 mapping:2 claim:1 driving:1 relay:1 overt:1 establishes:1 mit:1 sensor:4 gaussian:1 cr:1 varying:1 factorizes:3 morton:17 focus:1 greatly:1 inference:1 typically:1 w:2 irn:1 lfor:1 psychophysics:1 phonological:1 represents:3 thinking:1 connectionist:2 stimulus:32 contaminated:2 others:1 report:1 simultaneously:1 individual:2 cns:1 vowel:1 interest:1 investigate:2 bracket:1 dyi:1 implication:2 integral:4 desired:1 psychological:1 increased:2 goodness:3 gr:1 characterize:1 synthetic:2 density:8 yl:1 connecting:1 cognitive:4 external:7 corner:3 potential:4 onset:1 vi:1 depends:1 later:1 analyze:1 square:1 ni:1 phoneme:1 identify:1 ofthe:1 generalize:1 unaffected:2 za:5 definition:1 frequency:1 james:1 associated:2 gain:5 sampled:1 massachusetts:1 logical:1 lim:3 javier:1 dt:1 irr:1 response:49 though:2 generality:1 logistic:1 effect:4 symmetric:6 eg:2 essence:2 motion:2 cp:3 aiyi:1 instantaneous:1 common:1 functional:1 cohen:3 volume:1 discussed:1 mellon:2 cambridge:1 ai:1 had:1 specification:6 base:1 brownian:2 showed:1 conjectured:1 belongs:3 binary:1 arbitrarily:1 yi:8 fortunately:1 converge:2 ii:1 sound:2 technical:1 a1:1 controlled:1 iog:1 prediction:2 converging:1 cmu:1 represent:5 limt:1 grounded:1 achieved:1 c1:1 adhere:1 source:2 rest:1 subject:3 call:1 presence:1 enough:1 easy:1 affect:1 xj:1 psychology:3 zi:10 architecture:2 idea:1 simplifies:1 luce:3 inactive:1 whether:1 tlc:1 speech:2 york:1 useful:2 covered:1 detailed:1 amount:1 mcclelland:9 category:1 http:1 exist:1 zj:6 neuroscience:1 disjoint:1 carnegie:2 discrete:5 write:1 affected:4 four:1 changing:1 diffusion:13 wand:1 injected:1 named:4 architectural:1 incompatible:3 submatrix:1 ki:1 followed:1 strength:2 occur:1 constraint:2 wc:2 speed:1 argument:1 separable:8 relatively:2 conjecture:2 department:2 according:1 combination:1 jr:1 describes:1 separability:5 wi:2 making:1 indexing:1 equation:5 discus:1 needed:1 available:1 observe:1 alternative:2 existence:5 xc:9 hypercube:1 psychophysical:2 capacitance:1 rt:1 exhibit:2 amongst:1 lateral:4 seven:1 assuming:1 ratio:8 difficult:1 trace:1 negative:1 zt:4 policy:2 zf:1 neuron:2 dispersion:2 pdp:1 arbitrary:3 drift:2 connection:13 california:1 perception:12 pattern:1 force:1 eh:2 examination:1 zr:23 representing:5 gz:1 hm:3 l1s:3 review:1 literature:1 interdependent:1 law:17 loss:1 versus:1 remarkable:1 integrate:1 consistent:1 accounted:1 placed:1 saj:2 surprisingly:1 zc:17 side:1 institute:1 face:1 feedback:9 z8:2 san:1 approximate:3 active:2 assumed:1 consonant:2 excite:1 factorize:6 continuous:2 table:2 channel:23 rearranging:1 du:2 necessarily:1 lvi:1 noise:2 nothing:1 cpi:1 tl:1 vr:2 wiley:1 precision:2 structurally:3 perceptual:1 third:2 ix:4 theorem:1 showing:1 x:13 pz:6 adding:1 sx:2 cx:2 kxk:1 expressed:1 talking:1 applies:2 determines:1 change:3 except:1 perceiving:1 internal:2 support:2 modulated:1 ex:2
742
1,679
Modeling High-Dimensional Discrete Data with Multi-Layer Neural Networks Samy Bengio * IDIAP CP 592, rue du Simplon 4, 1920 Martigny, Switzerland bengio@idiap.ch Yoshua Bengio Dept.IRO Universite de Montreal Montreal, Qc, Canada, H3C 317 bengioy@iro.umontreal.ca Abstract The curse of dimensionality is severe when modeling high-dimensional discrete data: the number of possible combinations of the variables explodes exponentially. In this paper we propose a new architecture for modeling high-dimensional data that requires resources (parameters and computations) that grow only at most as the square of the number of variables, using a multi-layer neural network to represent the joint distribution of the variables as the product of conditional distributions. The neural network can be interpreted as a graphical model without hidden random variables, but in which the conditional distributions are tied through the hidden units. The connectivity of the neural network can be pruned by using dependency tests between the variables. Experiments on modeling the distribution of several discrete data sets show statistically significant improvements over other methods such as naive Bayes and comparable Bayesian networks, and show that significant improvements can be obtained by pruning the network. 1 Introduction The curse of dimensionality hits particularly hard on models of high-dimensional discrete data because there are many more possible combinations of the values of the variables than can possibly be observed in any data set, even the large data sets now common in datamining applications. In this paper we are dealing in particular with multivariate discrete data, where one tries to build a model of the distribution of the data. This can be used for example to detect anomalous cases in data-mining applications, or it can be used to model the class-conditional distribution of some observed variables in order to build a classifier. A simple multinomial maximum likelihood model would give zero probability to all of the combinations not encountered in the training set, i.e., it would most likely give zero probability to most out-of-sample test cases. Smoothing the model by assigning the same non-zero probability for all the unobserved cases would not be satisfactory either because it would not provide much generalization from the training set. This could be obtained by using a multivariate multinomial model whose parameters B are estimated by the maximum a-posteriori (MAP) principle, i.e., those that have the greatest probability, given the training data D, and using a diffuse prior PCB) (e.g. Dirichlet) on the parameters. A graphical model or Bayesian network [6, 5) represents the joint distribution of random variables Zl ... Zn with n P(ZI ... Zn) = II P(ZiIParentsi) i=l ?Part of this work was done while S.B. was at CIRANO, Montreal, Qc. Canada. Modeling High-Dimensional Discrete Data with Neural Networks 401 where Parentsi is the set of random variables which are called the parents of variable i in the graphical model because they directly condition Zi, and an arrow is drawn, in the graphical model, to Zi, from each of its parents. A fully connected "left-to-right" graphical model is illustrated in Figure 1 (left), which corresponds to the model n P(ZI . .. Zn) = II P(ZiIZl ... Zi-r) . (1) i= l Figure 1: Left: a fully connected "left-to-right" graphical model. Right: the architecture of a neural network that simulates a ful1y connected "left-to-right" graphical model. The observed values Zi = Zi are encoded in the corresponding input unit group. hi is a group of hidden units. gi is a group of output units, which depend on Zl ... Zi - l , representing the parameters of a distribution over Zi. These conditional probabilities P(ZiIZl . . . Zi-r) are multiplied to obtain the joint distribution. Note that this representation depends on the ordering of the variables (in that all previous variables in this order are taken as parents). We call each combination of the values of Parentsi a context. In the "exact" model (with the full table of all possible contexts) all the orders are equivalent, but if approximations are used, different predictions could be made by different models assuming different orders. In graphical models, the curse of dimensionality shows up in the representation of conditional distributions P(Zi IParentsi) where Zi has many parents. If Zj E Parentsi can take nj values, there are TI j nj different contexts which can occur in which one would like to estimate the distribution of Zi. This serious problem has been addressed in the past by two types of approaches, which are sometimes combined: 1. Not modeling all the dependencies between all the variables: this is the approach mainly taken with most graphical models or Bayes networks [6, 5] . The set of independencies can be assumed using a-priori or human expert knowledge or can be learned from data. See also [2] in which the set Parentsi is restricted to at most one element, which is chosen to maximize the correlation with Zi. 2 . Approximating the mathematicalform of the joint distribution with a form that takes only into account dependencies of lower order, or only takes into account some of the possible dependencies, e.g., with the Rademacher-Walsh expansion or multi-binomial [1,3], which is a low-order polynomial approximation of a full joint binomial distribution (and is used in the experiments reported in this paper). The approach we are putting forward in this paper is mostly of the second category, although we are using simple non-parametric statistics of the dependency between pairs of variables to further reduce the number of required parameters. In the multi-binomial model [3], the joint distribution of a set of binary variables is approximated by a polynomial. Whereas the "exact" representation of P( Zl = Z l , ... Zn = zn) as a function of Z l . . . Zn is a polynomial of degree n, it can be approximated with a lower Y. Bengio and S. Bengio 402 degree polynomial, and this approximation can be easily computed using the RademacherWalsh expansion [1] (or other similar expansions, such as the Bahadur-Lazarsfeld expansion [1]) . Therefore, instead of having 2 n parameters, the approximated model for P(Zl , . . . Zn) only requires O(nk) parameters. Typically, order k = 2 is used. The model proposed here also requires O(n 2 ) parameters, but it allows to model dependencies between tuples of variables, with more than 2 variables at a time. In previous related work by Frey [4], a fully-connected graphical model is used (see Figure 1, left) but each of the conditional distributions is represented by a logistic, which take into account only first-order dependency between the variables: 1 P(Zi = llZl ... Zi-d = ( L j<i Wj Z j )' 1 + exp -Wo In this paper, we basically extend Frey's idea to using a neural network with a hidden layer, with a particular architecture, allowing multinomial or continuous variables, and we propose to prune down the network weights . Frey has named his model a Logistic Autoregressive Bayesian Network or LARC . He argues that the prior variances on the logistic weights (which correspond to inverse weight decays) should be chosen inversely proportional to the number of conditioning variables (i.e. the number of inputs to the particular output neuron). The model was tested on a task of learning to classify digits from 8x8 binary pixel images. Models with different orderings of the variables were compared and did not yield significant differences in performance. When averaging the predictive probabilities from 10 different models obtained by considering 10 different random orderings, Frey obtained small improvements in likelihood but not in classification. The model performed better or equivalently to other models tested: CART, naive Bayes, K-nearest neighbors, and various Bayesian models with hidden variables (Helmholtz machines). These results are impressive, taking into account the simplicity of the LARC model. 2 Proposed Architecture The proposed architecture is a "neural network" implementation of a graphical model where all the variables are observed in the training set, with the hidden units playing a significant role to share parameters across different conditional distributions. Figure 1 (right) illustrates the model in the simpler case of a fully connected (Ieft-to-right) graphical model (Figure 1, left) . The neural network represents the parametrized function jo(zt, . .. , zn) = log(?O(Zl = Zl,? ? ., Zn = zn)) (2) approximating the joint distribution of the variables, with parameters 0 being the weights of the neural network. The architecture has three layers, with each layer organized in groups associated to each of the variables. The above log-probability is computed as the sum of conditional log-probabilities n jO(Zl , . .. , zn) = L 109(P(Zi = zilgi(zl, .. . , Zi-l))) i=l where gi( Zt, . .. , zi-d is the vector-valued output of the i-th group of output units, and it gives the value of the parameters of the distribution of Zi when Zl = Zl , Z2 = Z2, .. . , Zi-l = Zi-l' For example, in the ordinary discrete case, gi may be the vector of probabilities associated with each of the possible values of the multinomial random variable Zi . In this case, we have P(Zi = i'lgi) = gi ,i' In this example, a softmax output for the i-th group may be used to force these parameters to be positive and sum to 1, i.e., gi ,i' = Lil e g' i , i' Modeling High-Dimensional Discrete Data with Neural Networks 403 where g~ i' are linear combinations of the hidden units outputs, with i' ranging over the number of elements of the parameter vector associated with the distribution of Zi (for a fixed value of Zl ... Zi-l). To guarantee that the functions gi(Zl, ... , Zi-l) only depend on Zl ... Zi-l and not on any of Zi ... Zn, the connectivity struture of the hidden units must be constrained as follows: g~,i' = bi,i' + 2: j~i mj 2: Wi,i' ,j,j' hj,j' j'=1 where the b's are biases and the w's are weights of the output layer, and the hj,j' is the output of the j'-th unit (out of mj such units) in the j-th group of hidden layer nodes. It may be computed as follows: hj ,j' = tanh(cj,j' + 2: nk 2: Vj ,j' ,k ,k' Zk ,k') k<j k'=l where the c's are biases and the v's are the weights of the hidden layer, and Zk,k' is k'-th element of the vectorial input representation of the value Zk = Zk. For example, in the binary case (Zi = 0 or 1) we have used only one input node, i.e., Zi binomial -t Zi,O = Zi and in the multinomial case we use the one-hot encoding, Zi E {O, 1, ... ni - I} -t Zi ,i' = 8Zi ,i' where 8i ,i' = 1 if i = i' and 0 otherwise. The input layer has n - 1 groups because the value Zn = Zn is not used as an input. The hidden layer also has n - 1 groups corresponding to the variables j = 2 to n (since P(Z.) is represented unconditionally in the first output group, its corresponding group does not need any hidden units or inputs, but just has biases). 2.1 Discussion The number of free parameters of the model is O(n 2 H) where H = maXi mj is the maximum number of hidden units per hidden group (i.e., associated with one of the variables). This is basically quadratic in the number of variables, like the multi-binomial approximation that uses a polynomial expansion of the joint distribution. However, as H is increased, representation theorems for neural networks suggest that we should be able to approximate with arbitrary precision the true joint distribution. Of course the true limiting factor is the amount of data, and H should be tuned according to the amount of data. In our experiments we have used cross-validation to choose a value of mj = H for all the hidden groups. In this sense, this neural network representation of P(ZI ... Zn) is to the polynomial expansions (such as the multi-binomial) what ordinary multilayer neural networks for function approximation are to polynomial function approximators. It allows to capture high-order dependencies, but not all of them. It is the number of hidden units that controls "how many" such dependencies will be captured, and it is the data that "chooses" which of the actual dependencies are most useful in maximizing the likelihood. Unlike Bayesian networks with hidden random variables, learning with the proposed architecture is very simple, even when there are no conditional independencies. To optimize the parameters we have simply used gradient-based optimization methods, either using conjugate or stochastic (on-line) gradient, to maximize the total log-likelihood which is the sum of values of f (eq. 2) for the training examples. A prior on the parameters can be incorporated in the cost function and the MAP estimator can be obtained as easily, by maximizing the total log-likelihood plus the log-prior on the parameters. In our experiments we have used a "weight decay" penalty inspired by the analysis of Frey [4], with a penalty proportional to the number of weights incoming into a neuron. Y. Bengio and S. Bengio 404 However, it is not so clear how the distribution could be generally marginalized, except by summing over possibly many combinations of the values of variables to be integrated. Another related question is whether one could deal with missing values: if the total number of values that the missing variables can take is reasonably small, then one can sum over these values in order to obtain a marginal probability and maximize this probability. If some variables have more systematically missing values, they can be put at the end of the variable ordering, and in this case it is very easy to compute the marginal distribution (by taking only the product of the output probabilities up to the missing variables). Similarly, one can easily compute the predictive distribution of the last variable given the first n - 1 variables. The framework can be easily extended to hybrid models involving both continuous and discrete variables. In the case of continuous variables, one has to choose a parametric form for the distribution of the continuous variable when all its parents (i.e., the conditioning context) are fixed. For example one could use a normal, log-normal, or mixture of normals. Instead of having softmax outputs, the i-th output group would compute the parameters of this continuous distribution (e.g., mean and log-variance). Another type of extension allows to build a conditional distribution, e.g., to model P(ZI ... ZnlXl ... Xm). One just adds extra input units to represent the values of the conditioning variables Xl ... X m . Finally, an architectural extension that we have implemented is to allow direct input-tooutput connections (still following the rules of ordering which allow gi to depend only on Zl ... Zi-l). Therefore in the case where the number of hidden units is 0 (H = 0) we obtain the LARC model proposed by Frey [4]. 2.2 Choice of topology Another type of extension of this model which we have found very useful in our experiments is to allow the user to choose a topology that is not fully connected (Ieft-to-right). In our experiments we have used non-parametric tests to heuristically eliminate some of the connections in the network, but one could also use expert or prior knowledge, just as with regular graphical models, in order to cut down on the number of free parameters. In our experiments we have used for a pairwise test of statistical dependency the Kolmogorov-Smirnov statistic (which works both for continuous and discrete variables) . The statistic for variables X and Y is Jl sup IP(X :::; Xi, Y :::; Yi) - P(X :::; Xi)P(Y :::; Yi) I i where l is the number of examples and P is the empirical distribution (obtained by counting s = over the training data). We have ranked the pairs according to their value of the statistic s, and we have chosen those pairs for which the value of statistic is above a threshold value s*, which was chosen by cross-validation. When the pairs {(Zi' Zj)} are chosen to be part of the model, and assuming without loss of generality that i < j for those pairs, then the only connections that are kept in the network (in addition to those from the k-th hidden group to the k-th output group) are those from hidden group i to output group j, and from input group i to hidden group j, for every such (Zi' Zj) pair. 3 Experiments In the experiments we have compared the following models: ? Naive Bayes: the likelihood is obtained as a product of multinomials (one per variable). Each multinomial is smoothed with a Dirichlet prior. ? Multi-Binomial (using Rademacher-Walsh expansion of order 2) [3]. Since this only handles the case of binary data, it was only applied to the DNA data set. ? A simple graphical model with the same pairs of variables and variable ordering as selected for the neural network, but in which each of the conditional distribution is modeled Modeling High-Dimensional Discrete Data with Neural Networks 405 by a separate multinomial for each of the conditioning context. This works only if the number of conditioning variables is small so in the Mushroom, Audiology, and Soybean experiments we had to reduce the number of conditioning variables (following the order given by the above tests). The multinomials are also smoothed with a Dirichlet prior. ? Neural network: the architecture described above, with or without hidden units (i.e., LARC), with or without pruning. 5-fold cross-validation was used to select the number of hidden units per hidden group and the weight decay for the neural network and LARC. Cross-validation was also used to choose the amount of pruning in the neural network and LARC, and the amount of smoothing in the Dirichlet priors for the muItinomials of the naive Bayes model and the simple graphical model. 3.1 Results All four data sets were obtained on the web from the VCI Machine Learning and STATLOG databases. Most of these are meant to be for classification tasks but we have instead ignored the classification and used the data to learn a probabilistic model of all the input features. ? DNA (from STATLOG): there are 180 binary features. 2000 cases were used for training and cross-validation, and 1186 for testing. ? Mushroom (from VCI): there are 22 discrete features (taking each between 2 and 12 values). 4062 cases were used for training and cross-validation, and 4062 for testing. ? Audiology (from VCI): there are 69 discrete features (taking each between 2 and 7 values). 113 cases are used for training and 113 for testing (the original train-test partition was 200 + 26 and we concatenated and re-split the data to obtain more significant test figures). ? Soybean (from VCI): there are 35 discrete features (taking each between 2 and 8 values). 307 cases are used for training and 376 for testing. Table 1 clearly shows that the proposed model yields promising results since the pruned neural network was superior to all the other models in all 4 cases, and the pairwise differences with the other models are statistically significant in all 4 cases (except Audiology, where the difference with the network without hidden units, LARC, is not significant). 4 Conclusion In this paper we have proposed a new application of multi-layer neural networks to the modelization of high-dimensional distributions, in particular for discrete data (but the model could also be applied to continuous or mixed discrete / continuous data). Like the polynomial expansions [3] that have been previously proposed for handling such high-dimensional distributions, the model approximates the joint distribution with a reasonable (O( n 2 )) number of free parameters but unlike these it allows to capture high-order dependencies even when the number of parameters is small. The model can also be seen as an extension of the previously proposed auto-regressive logistic Bayesian network [4], using hidden units to capture some high-order dependencies. Experimental results on four data sets with many discrete variables are very encouraging. The comparisons were made with a naive Bayes model, with a multi-binomial expansion, with the LARC model and with a simple graphical model, showing that a neural network did significantly better in terms of out-of-sample log-likelihood in all cases. The approach to pruning the neural network used in the experiments, based on pairwise statistical dependency tests, is highly heuristic and better results might be obtained using approaches that take into account the higher order dependencies when selecting the conditioning variables. Methods based on pruning the fully connected network (e.g., with a "weight elimination" penalty) should also be tried. Also, we have not tried to optimize Y Bengio and S. Bengio 406 naive Bayes multi-Binomial order 2 ordinary graph. model LARC prunedLARC full-conn. neural net. pruned neural network naive Bayes multi-Binomial order 2 ordinary graph. model LARC prunedLARC full-conn. neural net. pruned neural network DNA mean (stdev) p-value 100.4 (.18) <le-9 117.8(.01) <le-9 108.1 (.06) <le-9 83.2 (.24) 7e-5 91.2(.15) <le-9 120.0 (.02) <le-9 82.9 (.21) Audiology mean (stdev) p-value 36.40 (2 .9) <le-9 16.56 (.48) 17.69 (.65) 16.69 (.41) 17 .39 (.58) 16.37 (.45) 6.8e-4 <le-9 0.20 <le-9 Mushroom mean (stdev) p-value 47 .00 (.29) <le-9 44.68 (.26) <le-9 42.51 (.16) <le-9 43.87 (.13) < le-9 33.58 (.01) <le-9 31.25 (.04) Soybean mean (stdev) p-value 34.74 (1.0) <le-9 43 .65 (.07) 16.95 (.35) 19.06 (.43) 21.65 (.43) 16.55 (.27) <le-9 5.5e-4 <le-9 <le-9 Table 1: Average out-of-sample negative log-likelihood obtained with the various models on four data sets (standard deviations of the average in parenthesis and p-value to test the null hypotheses that a model has same true generalization error as the pruned neural network). The pruned neural network was better than all the other models in in all cases, and the pair-wise difference is always statistically significant (except with respect to the pruned LARC on Audiology). the order of the variables, or combine different networks obtained with different orders, like [4] . References [1] RR Bahadur. A representation of the joint distribution of responses to n dichotomous items. In ed. H. Solomon, editor, Studies in Item Analysis and Predictdion, pages 158-168. Stanford University Press, California, 1961. [2] c.K. Chow. A recognition method using neighbor dependence. IRE Trans. Elec. Comp., EC-l1 :683-690, October 1962. [3] RO. Duda and P.E. Hart. Pattern Classification and Scene Analysis. Wiley, New York, 1973. [4] B. Frey. Graphical models for machine learning and digital communication. MIT Press, 1998. [5] Steffen L. Lauritzen. The EM algorithm for graphical association models with missing data. Computational Statistics and Data Analysis, 19:191-201,1995. [6] Judea Pearl. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann, 1988.
1679 |@word polynomial:8 duda:1 smirnov:1 heuristically:1 tried:2 selecting:1 tuned:1 past:1 z2:2 assigning:1 mushroom:3 must:1 partition:1 selected:1 item:2 regressive:1 ire:1 node:2 simpler:1 direct:1 combine:1 pairwise:3 multi:11 steffen:1 inspired:1 actual:1 curse:3 encouraging:1 considering:1 null:1 what:1 interpreted:1 unobserved:1 nj:2 guarantee:1 every:1 ti:1 ro:1 classifier:1 hit:1 zl:14 unit:19 control:1 positive:1 frey:7 encoding:1 might:1 plus:1 walsh:2 bi:1 statistically:3 testing:4 digit:1 empirical:1 significantly:1 regular:1 suggest:1 put:1 context:5 optimize:2 equivalent:1 map:2 missing:5 maximizing:2 qc:2 simplicity:1 estimator:1 rule:1 his:1 handle:1 limiting:1 user:1 exact:2 us:1 samy:1 hypothesis:1 element:3 helmholtz:1 approximated:3 particularly:1 recognition:1 cut:1 database:1 observed:4 role:1 capture:3 wj:1 connected:7 ordering:6 depend:3 predictive:2 easily:4 joint:11 represented:2 various:2 kolmogorov:1 stdev:4 train:1 elec:1 lazarsfeld:1 whose:1 encoded:1 heuristic:1 valued:1 stanford:1 plausible:1 otherwise:1 statistic:6 gi:7 h3c:1 ip:1 rr:1 net:2 propose:2 product:3 parent:5 rademacher:2 montreal:3 nearest:1 lauritzen:1 eq:1 implemented:1 idiap:2 switzerland:1 stochastic:1 human:1 elimination:1 generalization:2 statlog:2 extension:4 normal:3 exp:1 tanh:1 mit:1 clearly:1 always:1 hj:3 improvement:3 likelihood:8 mainly:1 detect:1 sense:1 posteriori:1 tooutput:1 inference:1 typically:1 integrated:1 eliminate:1 chow:1 hidden:26 pixel:1 classification:4 priori:1 smoothing:2 softmax:2 constrained:1 marginal:2 having:2 represents:2 yoshua:1 intelligent:1 serious:1 mining:1 highly:1 severe:1 mixture:1 re:1 increased:1 classify:1 modeling:8 zn:15 ordinary:4 cost:1 deviation:1 reported:1 dependency:15 combined:1 chooses:1 probabilistic:2 connectivity:2 jo:2 solomon:1 choose:4 possibly:2 soybean:3 expert:2 account:5 de:1 depends:1 performed:1 try:1 sup:1 bayes:8 square:1 ni:1 variance:2 kaufmann:1 correspond:1 yield:2 bayesian:6 basically:2 lgi:1 comp:1 ed:1 universite:1 associated:4 judea:1 knowledge:2 dimensionality:3 organized:1 cj:1 higher:1 response:1 done:1 generality:1 just:3 correlation:1 web:1 vci:4 logistic:4 true:3 satisfactory:1 illustrated:1 deal:1 argues:1 cp:1 l1:1 reasoning:1 image:1 ranging:1 wise:1 umontreal:1 common:1 superior:1 multinomial:9 conditioning:7 exponentially:1 modelization:1 jl:1 extend:1 he:1 approximates:1 association:1 significant:8 similarly:1 had:1 impressive:1 add:1 multivariate:2 binary:5 approximators:1 yi:2 seen:1 captured:1 morgan:1 prune:1 maximize:3 ii:2 full:4 cross:6 hart:1 parenthesis:1 prediction:1 anomalous:1 involving:1 multilayer:1 represent:2 sometimes:1 whereas:1 addition:1 addressed:1 grow:1 extra:1 unlike:2 explodes:1 cart:1 simulates:1 call:1 counting:1 bengio:9 easy:1 split:1 zi:41 architecture:8 topology:2 reduce:2 idea:1 whether:1 wo:1 penalty:3 york:1 ieft:2 ignored:1 useful:2 generally:1 clear:1 amount:4 category:1 dna:3 bahadur:2 zj:3 estimated:1 per:3 discrete:17 group:21 conn:2 four:3 putting:1 threshold:1 independency:2 drawn:1 kept:1 graph:2 sum:4 inverse:1 audiology:5 named:1 reasonable:1 architectural:1 comparable:1 layer:11 hi:1 fold:1 quadratic:1 encountered:1 occur:1 vectorial:1 scene:1 diffuse:1 dichotomous:1 pruned:7 according:2 combination:6 conjugate:1 across:1 em:1 wi:1 restricted:1 taken:2 resource:1 previously:2 end:1 multiplied:1 datamining:1 original:1 binomial:10 dirichlet:4 graphical:18 marginalized:1 concatenated:1 build:3 approximating:2 question:1 parametric:3 dependence:1 gradient:2 separate:1 parametrized:1 iro:2 assuming:2 modeled:1 equivalently:1 mostly:1 october:1 pcb:1 negative:1 martigny:1 implementation:1 zt:2 lil:1 allowing:1 neuron:2 extended:1 incorporated:1 communication:1 smoothed:2 arbitrary:1 canada:2 pair:8 required:1 connection:3 california:1 learned:1 pearl:1 trans:1 able:1 pattern:1 xm:1 greatest:1 simplon:1 hot:1 ranked:1 force:1 hybrid:1 representing:1 inversely:1 x8:1 naive:7 unconditionally:1 auto:1 prior:8 fully:6 loss:1 mixed:1 proportional:2 validation:6 digital:1 degree:2 principle:1 editor:1 systematically:1 playing:1 share:1 course:1 last:1 free:3 bias:3 allow:3 neighbor:2 taking:5 autoregressive:1 forward:1 made:2 ec:1 cirano:1 pruning:5 approximate:1 dealing:1 incoming:1 summing:1 assumed:1 tuples:1 xi:2 continuous:8 table:3 promising:1 mj:4 zk:4 reasonably:1 ca:1 learn:1 expansion:9 du:1 rue:1 vj:1 did:2 arrow:1 wiley:1 precision:1 xl:1 bengioy:1 tied:1 down:2 theorem:1 showing:1 maxi:1 decay:3 illustrates:1 nk:2 simply:1 likely:1 ch:1 corresponds:1 conditional:11 hard:1 except:3 averaging:1 called:1 total:3 experimental:1 select:1 meant:1 dept:1 tested:2 handling:1
743
168
410 NEURAL CONTROL OF SENSORY ACQUISITION: THE VESTIBULO-OCULAR REFLEX. Michael G. Paulin, Mark E. Nelson and James M. Bower Division of Biology California Institute of Technology Pasadena, CA 91125 ABSTRACT We present a new hypothesis that the cerebellum plays a key role in actively controlling the acquisition of sensory infonnation by the nervous system. In this paper we explore this idea by examining the function of a simple cerebellar-related behavior, the vestibula-ocular reflex or VOR, in which eye movements are generated to minimize image slip on the retina during rapid head movements. Considering this system from the point of view of statistical estimation theory, our results suggest that the transfer function of the VOR, often regarded as a static or slowly modifiable feature of the system, should actually be continuously and rapidly changed during head movements. We further suggest that these changes are under the direct control of the cerebellar cortex and propose experiments to test this hypothesis. 1. INTRODUCTION A major thrust of research in our laboratory involves exploring the way in which the nervous system actively controls the acquisition of infonnation about the outside world. This emphasis is founded on our suspicion that the principal role of the cerebellum, through its influence on motor systems, is to monitor and optimize the quality of sensory information entering the brain. To explore this question, we have undertaken an investigation of the simplest example of a cerebellar-related motor activity that results in improved sensory inputs, the vestibulo-ocular reflex (VOR). This reflex is responsible for moving the eyes to compensate for rapid head movements to prevent retinal image slip which would otherwise significantly degrade visual acuity (Carpenter, 1977). 2. VESTIBULO-OCULAR REFLEX (VOR) The VOR relies on the vestibular apparatus of the inner ear which is an inertial sensor that detects movements of the head. Vestibular output caused by head movements give rise to compensatory eye movements through an anatomically well described neural pathway in the brain stem (for a review see Ito, 1984). Visual feedback also makes an important contribution to compensatory eye movements during slow head movements, Neural Control of Sensory Acquisition but during rapid head movements with frequency components greater than about 1Hz, the vestibular component dominates (Carpenter, 1977). A simple analysis of the image stabilization problem indicates that during head rotation in a single plane, the eyes should be made to rotate at equal velocity in the opposite direction. This implies that, in a simple feedforward control model, the VOR transfer function should have unity gain and a 1800 phase shift. This would assure stabilized retinal images of distant objects. It turns out, however, that actual measurements reveal the situation is not this simple. Furman, O'Leary and Wolfe (1982), for example, found that the monkey VOR has approximately unity gain and 1800 phase shift only in a narrow frequency band around 2Hz. At 4Hz the gain is too high by a factor of about 30% (fig. 1). 1.2 ~ -< C) 1.0 j IHf ,...., ~ 5 0 - I til ff1d 1tHf~ff 1\ lIJ (f) ~O n.. -5 0.8 2 3 4 FREQUENCY (Hz) 5 3 2 4 5 FREQUENCY (Hz) Figure 1: Bode gain and phase plots for the transfer function of the horizontal component of the VOR of the alert Rhesus monkey at high frequencies (Data from Furman et al. (1982?. Given the expectation of unity gain, one might be tempted to conclude from the monkey data that the VOR simply does not perform well at high frequencies. But 4Hz is not a very high frequency for head movements, and perhaps it is not the VOR which is performing poorly, but the simplified analysis using classical control theory. In this paper, we argue that the VOR uses a more sophisticated strategy and that the "excessive" gain in the system seen at higher frequencies actually improves VOR performance. 3. OPTIMAL ESTIMATION In order to understand the discrepancy between the predictions of simple control theory models and measured VOR dynamics, we believe it is necessary to take into account more of the real world conditions under which the VOR operates. Examples include noisy head velocity measurements, conduction delays and multiple, possibly conflicting, measurements of head velocity, acceleration, muscle contractions, etc., generated by different sensory modalities. The mathematical framework that is appropriate for analyz- 411 412 Paulin, Nelson and Bower ing problems of this kind is stochastic state-space dynamical systems theory (Davis and Vinter, 1985). This framework is an extension of classical linear dynamical systems theory that accommodates multiple inputs and outputs, nonlinearities, time-varying dynamics, noise and delays. One area of application of the state space theory has been in target tracking, where the basic principle involves using knowledge of the dynamics of a target to estimate its most probable trajectory given imprecise data. The VOR can be viewed as a target tracking system whose target is the "world", which moves in head coordinates. We have reexamined the VOR from this point of view. The Basic VOR. To begin our analysis of the VOR we have modeled the eye-head-neck system as a damped inverted pendulum with linear restoring forces (fig. 2) where the model system is driven by random (Gaussian white) torque. Within this model, we want to predict the correct compensatory "eye" movements during "head" movements to stabilize the direction in which the eye is pointing. Figure 2 shows the amplitude spectrum of head velocity for this model. In this case, the parameters of the model result in a system that has a natural resonance in the range of 1 to 2 Hz and attenuates higher frequencies. 20 ? ? ??? S I S 0.1 z , s. u ? '. ? ?? i. I 1.0 FREQUENCY Figure 2: Amplitude spectrum of model head velocity. We provide noisy measurements of "head" velocity and then ask what transfer function, or filter, will give the most accurate "eye" movement compensation? This is an estimation problem and, for Gaussian measurement error, the solution was discovered by Kalman and Bucy (1961). The optimal fIlter or estimator is often called the KalmanBucy filter. The gain and phase plots of the optimal filter for tracking movements of the inverted pendulum model are shown in figure 3. It can be seen that the gain of the optimal estimator for this system peaks near the maximum in the spectrum of "head-neck" velocity (fig. 2). This is a general feature of optimal filters. Accordingly, to accurately compensate for head movement in this system, the VOR would need to have a frequency dependent gain. Neural Control of Sensory Acquisition - - ~ 20 ~ 0 bO 0 So j 0 5 ~ ~ ~ tI) -20 ~90 0.1 1.0 .0 10.0 Figure 3: Bode gain plot (left) and phase plot (right) of an optimal estimator for tracking the inverted pendulum using noisy data. Time Varying dynamics and the VOR So far we have considered our model for VOR optimization only in the simple case of a constant head-neck velocity power spectrum. Under natural conditions, however, this spectrum would be expected to change. For example, when gait changes from walking to running, corresponding changes in the VOR transfer function would be necessary to maintain optimal performance. To explore this, we added a second inverted pendulum to our model to simulate body dynamics. We simulated changes in gait by changing the resonant frequency of the trunk. Figure 4 compares the spectra of head-neck velocity with two different trunk parameters. As in the previous example, we then computed transfer functions of the optimal filters for estimating head velocity from noisy measurements in these two cases. The gain and phase characteristics of these filters are also shown in Figure 5. These plots demonstrate that significant changes in the transfer function of the VOR would be necessary to maintain visual acuity in our model system under these different conditions. Of course, in the real situation head-neck dynamics will change rapidly and continuously with changes in gait, posture, substrate, etc. requiring rapid continuous changes in VOR dynamics rather than the simple switch implied here. -~ ~ ? 5 HEAD 20 0 .......----~~-~ ~ ~ -20 0.1 1.0 FREQUENCY 10.0 Figure 4: Head velocity spectrum during "walking" (light) and "running" (heavy). 413 414 Paulin, Nelson and Bower ~ ~ 0 ? -.8- 1----""'::3I~ 5 ~ -20 ? ? ?? , Ci. .1 D ?? 1.0 10.0 .1 1.0 10.0 Figure 5: Bode gain plots (left) and phase plots (right) for optimal estimators of head angular velocity during "walking" (light) and "running" (heavy). 4. SIGNIFICANCE TO THE REAL VOR Our results show that the optimal VOR transfer function requires a frequency dependent gain to accurately adjust to a wide range of head movements under real world conditions. Thus, the deviations from unity gain seen in actual measurements of the VOR may not represent poor, but rather optimal, performance. Our modeling similarly suggests that several other experimental results can be reinterpreted. For example, localized peaks or valleys in the VOR gain function can be induced experimentally through prolonged sinusoidal oscillations of subjects wearing magnifying or reducing lenses. However, this "frequency selectivity" is not thought to occur naturally and has been interpreted to imply the existence of frequency selective channels in the VOR control network (Lisberger, Miles and Optican, 1983). In our view there is no real distinction between this phenomenon and the "excessive" gain in normal monkey VOR; in each case the VOR optimizes its response for the particular task which it has to solve. This is testable. If we are correct, then frequency selective gain changes will occur following prolonged narrow-band rotation in the light without wearing lenses. In the classical framework there is no reason for any gain changes to occur in this situation. Another phenomenon which has been observed experimentally and that the current modeling sheds new light on is referred to as "pattern storage". After single-frequency sinusoidal oscillation on a turntable in the light for several hours, rabbits will continue to produce oscillatory eye movements when the lights are extinguished and the turntable stops. Trained rabbits also produce eye oscillations at the training frequency when oscillated in the dark at a different frequency (Collewijn, 1985). In this case the sinusoidal pattern seems to be "stored" in the nervous system. However, the effect is naturally accounted for by our optimal estimator hypothesis without relying on an explicit "pattern storage mechanism". An optimal estimator works by matching its dynamics to the dynamics of the signal generator, and in effect it tries to force an internal model to mimic the signal generator by comparing actual and expected patterns of sensory inputs. When Neural Control of Sensory Acquisition no data is available, or the data is thought to be very unreliable, an optimal estimator relies completely, or almost completely, on the model. In cases where the signal is patterned the estimator will behave as though it had memorized the pattern. Thus, if we hypothesize that the VOR is an optimal estimator we do not need an extra hypothesis to explain pattern storage. Again, our hypothesis is testable. If we are correct, then repeating the pattern storage experiments using rotational velocity waveforms obtained by driving a frequency-tuned oscillator with Gaussian white noise will produce identical dynamical effects in the VOR. There is no sinusoidal pattern in the stimulus, but we predict that the rabbits can be induced to generate sinusoidal eye movements in the dark after this training. The modeling results shown in figures 4 and 5 represent an extension of our ideas into the area of gait (or more generally "context") dependent changes in VOR which has not been considered very much in VOR research. In fact, VOR experimental paradigms, in general, are explicitly set up to produce the most stable VOR dynamics possible. Accordingly, little work has been done to quantify the short term changes in VOR dynamics that must occur in response to changes in effective head-neck dynamics. Experiments of this type would be valuable and are no more difficult technically than experiments which have already been done. For example, training an animal on a turntable which can be driven randomly with two distinct velocity power spectra, i.e. two "gaits", and providing the animal with external cues to indicate the gait would, we predict, result in an animal that could use the cues to switch its VOR dynamics. A more difficult but also more compelling demonstration would be to test VOR dynamics with impulsive head accelerations in different natural situations, using an unrestrained animal. s. SENSOR FUSION AND PREDICTION To this point, we have discussed compensatory eye movements by treating the VOR as a single input, single output system. This allowed us to concentrate on a particular aspect of VOR control: tracking a time-varying dynamical system (the head) using noisy data. In reality there are a number of other factors which make control of compensatory eye movements a somewhat more complex task than it appears to be when it is modeled using classical control theory. For example, a variety of vestibular as well as non-vestibular signals (e.g. visual, proprioceptive) relating to head movements are transmitted to the compensatory eye movement control network (Ito, 1984). This gives rise to a "sensor fusion" problem where data from different sources must be combined. The optimal solution to this problem for a multiple input - multiple output, time-varying linear, stochastic system is also given by the Kalman-Bucy filter (Davis and Vinter, 1985). Borah, Young and Curry (1988) have demonstrated that a Kalman-Bucy filter model of visualvestibular sensor fusion is able to account for visual-vestibular interactions in motion perception. Oman (1982) has also developed a Kalman-Bucy filter model of visualvestibular interactions. Their results show that the optimal estimation approach is useful 415 416 Paulin, Nelson and Bower for analyzing multivariate aspects of compensatory eye movement control, and complement our analysis of dynamical aspects. Another set of problems arises in the VOR because of small time delays in neural transmission and muscle activation. To optimize its response, the mammalian VOR needs to make up for these delays by predicting head movements about lOmsec in advance (ret). Once the dynamics of the signal generator have been identified, prediction can be performed using model-based estimation (Davis and Vinter, 1985). A neural analog of a Taylor series expansion has also been proposed as a model of prediction in the VOR (pellionisz and LUnas, 1979), but this mec.hanism is extremely sensitive to noise in the data and was abandoned as a practical technique for general signal prediction several decades ago in favor of model-based techniques (Wiener, 1948). The later approach may be more appropriate for analyzing neural mechanisms of prediction (Arbib and Amari, 1985). An elementary description of optimal estimation theory for target tracking, and its possible relation to cerebellar function, is given by Paulin (1988). 6. ROLE OF CEREBELLAR CORTEX IN VOR CONTROL To this point we have presented a novel characterization of the problem of compensatory eye movement control without considering the physical circuitry which implements the behavior. However, there are two parts to the optimal estimation problem. At each instant it is necessary to (a) filter the data using the optimal transfer function to drive the desired response and (b) determine what transfer function is optimal at that instant and adjust the filtering network accordingly. The first problem is fairly straightforward, and existing models of VOR demonstrate how a network of neurons based on known brains tern circuitry can implement a particular transfer function (Cannon and Robinson, 1985). The second problem is more difficult because requires continuous monitoring of the context in which head movements occur using a variety of sources of relevant data to tune the optimal filter for that context. We speculate that the cerebellar cortex performs this task. First, the cortex of the vestibulo-cerebellum is in a position to mflke , the required computation, since it receives detailed information from multiple sensory modalities that provide information on the state of the motor system (Ito, 1985). Second, the cerebellum projects to and appears to modulate the brain stem compensatory eye movement control network (Mackay and Murphy, 1979). We predict that the cerebellar cortex is necessary to produce rapid, context-dependent optimal state dependent changes in VOR transfer function which we have discussed. This speculation can be tested with turntable experiments similar to those described in section 4 above in the presence and absence of the cerebellar cortex. Neural Control of Sensory Acquisition 7. THE GENERAL FUNCTION OF CEREBELLAR CORTEX According to our hypothesis, the cerebellar cortex is required for making optimal compensatory eye movements during head movements. This is accomplished by continuously modifying the dynamics of the underlying control network in the brainstem, based on current sensory information. The function of the cerebellar cortex in this case can then be seen in a larger context as using primary sensory information (vestibular, visual) to coordinate the use of a motor system (the extraoccular eye muscles) to position a sensory array (the retina) to optimize the quality of sensory information available to the brain. We believe that this is the role played by the rest of the cerebellum for other sensory systems. Thus, we suspect that the hemispheres of the rat cerebellum, with their peri-oral tactile input (Bower et al., 1983), are involved in controlling the optimal use of these tactile surfaces in sensory exploration through the control of facial musculature. Similarly, the hemispheres of the primate cerebellum, which have hand and finger tactile inputs (Ito, 1984), may be involved in an analogous exploratory task in primates. These tactile sensory-motor systems are difficult to analyze, and we are currently studying a functionally analogous but more accessible model system, the electric sense of weakly electric fish (cf Rasnow et al., this volume). 8.CONCLUSION Our view of the cerebellum assigns it an important dynamic role which contrasts markedly with the more limited role it was assumed to have in the past as a learning device (Marr, 1969; Albus, 1971; Robinson, 1976). There is evidence that cerebellar cortex has some learning abilities (Ito, 1984), but it is recognized that cerebellar cortex has an important dynamic role in motor control. However, there are widely differing opinions as to the nature of that role (Ito, 1985; Miles and Lisberger, 1981; Pellionisz and Llinas, 1979). Our proposal, that the VOR is a neural analog of an optimal estimator and that the cerebellar cortex monitors context and sets reflex dynamics accordingly, should not be interpreted as a claim that the nervous system actually implements the computations which are involved in applied optimal estimation, such as the KalmanBucy filter. Understanding the neural basis of cerebellar function will require the combined power of a number of experimental, theoretical and modeling approaches (cf Wilson et al., this volume). We believe that analyses of the kind presented here have an important role in characterizing behaviors controlled by the cerebellum. Acknowledgments This work was supported by the NIH (BNS 22205), the NSF (EET-8700064), and the Joseph Drown Foundation. References Arbib M.A. and Amari S. 1985. Sensori-moto Transformations in the Brain (with a critique of the tensor theory of the cerebellum). J. Theor. BioI. 112:123-155 417 418 Paulin, Nelson and Bower Borah J., Young L.R. and Curry, R.E. 1988. Optimal Estimator Model for Human Spatial Orientation. In: Proc. N.Y. Acad. Sci. B. Cohen and V. Henn (eds.). In Press. Bower lM. and Woolston D.C. 1983. The Vertical Organization of Cerebellar Cortex. J. Nemophysiol. 49: 745-766. Carpenter R.H.S. 1977. Movements of the Eyes. Pion, London. Davis M.B.A. and Vinter R.B. 1985. Stochastic Modelling and Control. Chapman and Hall, NY. Funnan J.M., O'Leary D.P. and Wolfe lW. 1982. Dynamic Range of the Frequency Response of the Horizontal Vestibulo-Ocular Reflex of the Alert Rhesus Monkey. Acta Otolaryngol. 93: 81 Ito, M. 1984. The Cerebellum and Neural Control. Raven Press, NY. Kalman R.E. 1960. A New Approach to Linear Filtering and Prediction Problems. l Basic Eng., March 1960. Kalman R.E. and Bucy R.S. 1961. New Results in Linear Filtering and Prediction Theory. 1. Basic Eng., March 1961. Lisberger, S.G. 1988. The Nemal Basis for Learning of Simple Motor Skills. Science, 242:728735. Lisberger S.G. , Miles F.A. and Optican L.M. 1983. Frequency Selective Adaptation: Evidence for Channels in the Vestibulo-Ocular Reflex. J. Neurosci. 3:1234-1244 Mackay W.A. and Murphy J.T. 1979. Cerebellar Modulation of reflex Gain. Prog. Neurobiol. 13:361-417. Oman C.M. 1982. A heuristic mathematical Model for the Dynamics of Sensory Conflict and Motion Sickness. Acta Oto-Laryngol. S392. Paulin M.G. 1988. A Kalman Filter Model of the Cerebellum. In: Dynamic Interactions in Nemal Networks: Models and Data. M. Arbib and S. Amari (eds). Springer-Verlag, NY. pp239-261. Pellionisz A. and Llinas R. 1979. Brain Modelling by Tensor Network Theory and Computer Simulation. The Cerebellum: Distributed Processor for Predictive Coordination. Nemoscience 4:323-348. Robinson D.A. 1976. Adaptive Control of the Vestibulo-Ocular Reflex by the Cerebellum. J. Nemophys.36:954-969. Robinson D.A. 1981. The Use of Control Systems Analysis in the Neurophysiology of Eye Movements. Ann. Rev. Neurosci. 4:463-503. Wiener, N. 1948. Cybernetics: Communication and Control in the Animal and the Machine. MIT Press, Boston.
168 |@word neurophysiology:1 seems:1 simulation:1 rhesus:2 contraction:1 eng:2 series:1 tuned:1 optican:2 past:1 existing:1 current:2 comparing:1 activation:1 must:2 vor:49 distant:1 thrust:1 motor:7 hypothesize:1 plot:7 treating:1 nemal:2 cue:2 device:1 nervous:4 accordingly:4 plane:1 paulin:7 short:1 characterization:1 mathematical:2 alert:2 direct:1 pathway:1 expected:2 rapid:5 behavior:3 brain:7 torque:1 detects:1 relying:1 prolonged:2 actual:3 little:1 considering:2 begin:1 estimating:1 project:1 underlying:1 what:2 kind:2 interpreted:2 neurobiol:1 monkey:5 developed:1 ret:1 differing:1 transformation:1 ti:1 shed:1 control:27 apparatus:1 acad:1 analyzing:2 critique:1 modulation:1 approximately:1 might:1 emphasis:1 acta:2 suggests:1 patterned:1 limited:1 range:3 practical:1 responsible:1 acknowledgment:1 restoring:1 implement:3 area:2 significantly:1 thought:2 matching:1 imprecise:1 suggest:2 valley:1 storage:4 context:6 influence:1 optimize:3 demonstrated:1 straightforward:1 rabbit:3 oscillated:1 assigns:1 estimator:11 array:1 regarded:1 marr:1 exploratory:1 coordinate:2 analogous:2 target:5 controlling:2 play:1 substrate:1 us:1 slip:2 hypothesis:6 velocity:14 assure:1 wolfe:2 walking:3 mammalian:1 observed:1 role:9 vestibula:1 movement:32 valuable:1 dynamic:22 trained:1 weakly:1 oral:1 predictive:1 technically:1 division:1 completely:2 basis:2 finger:1 distinct:1 effective:1 london:1 outside:1 whose:1 heuristic:1 larger:1 solve:1 widely:1 otherwise:1 amari:3 favor:1 ability:1 noisy:5 propose:1 gait:6 interaction:3 adaptation:1 relevant:1 rapidly:2 poorly:1 albus:1 description:1 transmission:1 produce:5 object:1 measured:1 involves:2 implies:1 indicate:1 quantify:1 direction:2 waveform:1 concentrate:1 correct:3 filter:14 stochastic:3 modifying:1 stabilization:1 brainstem:1 exploration:1 human:1 opinion:1 memorized:1 require:1 investigation:1 probable:1 elementary:1 theor:1 exploring:1 extension:2 around:1 considered:2 drown:1 normal:1 hall:1 predict:4 claim:1 pointing:1 circuitry:2 driving:1 major:1 lm:1 estimation:8 proc:1 currently:1 infonnation:2 coordination:1 sensitive:1 mit:1 sensor:4 gaussian:3 rather:2 cannon:1 varying:4 wilson:1 acuity:2 modelling:2 indicates:1 unrestrained:1 contrast:1 sickness:1 sense:1 dependent:5 pasadena:1 relation:1 selective:3 orientation:1 resonance:1 animal:5 spatial:1 fairly:1 mackay:2 equal:1 once:1 chapman:1 biology:1 identical:1 excessive:2 discrepancy:1 mimic:1 stimulus:1 extinguished:1 mec:1 retina:2 randomly:1 murphy:2 phase:7 maintain:2 organization:1 reinterpreted:1 adjust:2 light:6 damped:1 accurate:1 necessary:5 facial:1 taylor:1 desired:1 theoretical:1 modeling:4 compelling:1 impulsive:1 deviation:1 delay:4 examining:1 too:1 stored:1 conduction:1 bucy:5 combined:2 peri:1 peak:2 accessible:1 michael:1 continuously:3 leary:2 again:1 ear:1 possibly:1 slowly:1 luna:1 external:1 til:1 actively:2 account:2 nonlinearities:1 sinusoidal:5 retinal:2 speculate:1 stabilize:1 caused:1 explicitly:1 performed:1 view:4 try:1 later:1 furman:2 analyze:1 pendulum:4 ihf:1 contribution:1 minimize:1 wiener:2 characteristic:1 sensori:1 accurately:2 trajectory:1 monitoring:1 drive:1 cybernetics:1 ago:1 processor:1 oscillatory:1 explain:1 ed:2 acquisition:7 frequency:23 ocular:7 james:1 involved:3 naturally:2 woolston:1 static:1 gain:19 stop:1 ask:1 knowledge:1 improves:1 amplitude:2 inertial:1 actually:3 sophisticated:1 appears:2 higher:2 response:5 improved:1 llinas:2 done:2 though:1 angular:1 hand:1 receives:1 horizontal:2 quality:2 reveal:1 perhaps:1 curry:2 believe:3 effect:3 requiring:1 entering:1 laboratory:1 proprioceptive:1 mile:3 white:2 oto:1 cerebellum:14 during:9 davis:4 rat:1 demonstrate:2 performs:1 motion:2 image:4 novel:1 nih:1 rotation:2 physical:1 cohen:1 volume:2 discussed:2 analog:2 relating:1 functionally:1 measurement:7 significant:1 similarly:2 had:1 moving:1 stable:1 cortex:13 surface:1 etc:2 multivariate:1 optimizes:1 driven:2 hemisphere:2 selectivity:1 verlag:1 continue:1 accomplished:1 muscle:3 inverted:4 seen:4 transmitted:1 greater:1 somewhat:1 vinter:4 recognized:1 determine:1 paradigm:1 signal:6 multiple:5 stem:2 ing:1 compensate:2 controlled:1 prediction:8 basic:4 expectation:1 cerebellar:17 represent:2 proposal:1 want:1 source:2 modality:2 extra:1 rest:1 markedly:1 hz:7 induced:2 subject:1 suspect:1 near:1 presence:1 feedforward:1 switch:2 variety:2 arbib:3 identified:1 opposite:1 inner:1 idea:2 shift:2 tactile:4 generally:1 useful:1 detailed:1 tune:1 turntable:4 repeating:1 dark:2 band:2 simplest:1 generate:1 nsf:1 stabilized:1 fish:1 modifiable:1 key:1 monitor:2 changing:1 prevent:1 undertaken:1 musculature:1 prog:1 almost:1 resonant:1 oscillation:3 collewijn:1 played:1 activity:1 occur:5 aspect:3 simulate:1 extremely:1 bns:1 performing:1 according:1 march:2 poor:1 bode:3 unity:4 joseph:1 rev:1 making:1 primate:2 anatomically:1 trunk:2 turn:1 mechanism:2 studying:1 available:2 appropriate:2 existence:1 abandoned:1 running:3 include:1 cf:2 instant:2 testable:2 classical:4 implied:1 move:1 tensor:2 question:1 added:1 posture:1 already:1 strategy:1 primary:1 simulated:1 accommodates:1 sci:1 nelson:5 degrade:1 argue:1 reason:1 kalman:7 modeled:2 rotational:1 providing:1 demonstration:1 difficult:4 rise:2 attenuates:1 perform:1 vertical:1 neuron:1 compensation:1 behave:1 situation:4 communication:1 head:34 discovered:1 complement:1 required:2 speculation:1 compensatory:10 conflict:1 california:1 distinction:1 narrow:2 conflicting:1 hour:1 vestibular:7 robinson:4 able:1 dynamical:5 pattern:8 perception:1 power:3 natural:3 force:2 predicting:1 technology:1 eye:22 imply:1 suspicion:1 thf:1 lij:1 review:1 understanding:1 filtering:3 localized:1 generator:3 foundation:1 vestibulo:7 principle:1 heavy:2 course:1 changed:1 accounted:1 supported:1 understand:1 institute:1 wide:1 characterizing:1 magnifying:1 distributed:1 feedback:1 world:4 sensory:19 made:1 adaptive:1 simplified:1 founded:1 far:1 skill:1 eet:1 unreliable:1 conclude:1 assumed:1 spectrum:8 continuous:2 decade:1 reality:1 channel:2 transfer:12 nature:1 ca:1 expansion:1 complex:1 electric:2 significance:1 neurosci:2 noise:3 allowed:1 carpenter:3 body:1 fig:3 referred:1 oman:2 ff:1 reexamined:1 slow:1 ny:3 position:2 explicit:1 bower:7 lw:1 ito:7 young:2 rasnow:1 otolaryngol:1 dominates:1 fusion:3 evidence:2 raven:1 ci:1 boston:1 simply:1 explore:3 visual:6 henn:1 tracking:6 bo:1 reflex:10 springer:1 lisberger:4 relies:2 bioi:1 modulate:1 viewed:1 acceleration:2 ann:1 tempted:1 oscillator:1 absence:1 change:15 experimentally:2 operates:1 reducing:1 principal:1 called:1 lens:2 neck:6 experimental:3 internal:1 mark:1 tern:1 rotate:1 arises:1 wearing:2 tested:1 phenomenon:2
744
1,680
Broadband Direction-Of-Arrival Estimation Based On Second Order Statistics Justinian Rosca Joseph 6 Ruanaidh Alexander Jourjine Scott Rickard {rosca,oruanaidh,jourjine,rickard}@scr.siemens.com Siemens Corporate Research, Inc. 755 College Rd E Princeton, NJ 08540 Abstract N wideband sources recorded using N closely spaced receivers can feasibly be separated based only on second order statistics when using a physical model of the mixing process. In this case we show that the parameter estimation problem can be essentially reduced to considering directions of arrival and attenuations of each signal. The paper presents two demixing methods operating in the time and frequency domain and experimentally shows that it is always possible to demix signals arriving at different angles. Moreover, one can use spatial cues to solve the channel selection problem and a post-processing Wiener filter to ameliorate the artifacts caused by demixing. 1 Introduction Blind source separation (BSS) is capable of dramatic results when used to separate mixtures of independent signals. The method relies on simultaneous recordings of signals from two or more input sensors and separates the original sources purely on the basis of statistical independence between them. Unfortunately, BSS literature is primarily concerned with the idealistic instantaneous mixing model. In this paper, we formulate a low dimensional and fast solution to the problem of separating two signals from a mixture recorded using two closely spaced receivers. Using a physical model of the mixing process reduces the complexity of the model and allows one to identify and to invert the mixing process using second order statistics only. We describe the theoretical basis of the new approach, and then focus on two algorithms, which were implemented and successfully applied to extensive sets of real-world data. In essence, our separation architecture is a system of adaptive directional receivers designed using the principles ofBSS. The method bears resemblance to methods in beamforming [8] in that it works by spatial filtering. Array processing techniques [2] reduce noise by separating signal space from noise space, which necessitates more receivers than emitters. The main differences are that standard beamforming and array processing techniques [8, 2] are generally strictly concerned with processing directional narrowband signals. The difference with BSS [7, 6] is that our approach is model-based and therefore the elements of the mixing matrix are highly constrained: a feature that aids in the robust and reliable identification of the mixing process. J. Rosca, J. 776 6 Ruanaidh, A. Jourjine and S. Rickard The layout of the paper is as follows. Sections 2 and 3 describe the theoretical foundation of the separation method that was pursued. Section 4 presents algorithms that were developed and experimental results. Finally we summarize and conclude this work. 2 Theoretical foundation for the BSS solution As a first approximation to the general multi-path model, we use the delay-mixing model. In this model, only direct path signal components are considered. Signal components from one source arrive with a fractional delay between the time of arrivals at two receivers. By fractional delays, we mean that delays between receivers are not generally integer multiples of the sampling period. The delay depends on the position of the source with respect to the receiver axis and the distance between receivers. Our BSS algorithms demix by compensating for the fractional delays. This, in effect, is a form of adaptive beamforming with directional notches being placed in the direction of sources of interference [8]. A more detailed account of the analytical structure of the solutions can be found in [1]. Below we address the case of two inputs and two outputs but there is no reason why the discussion cannot be generalized to multiple inputs and multiple outputs. Assume a linear mixture of two sources, where source amplitude drops off in proportion to distance: Xi(t) = 1 - S I (t - R-I _Z ) Ril 1 + -S2(t - C Ri2 R-2 (1) _Z ) C j = 1, 2, where c is the speed of wave propagation, and Rij indicates the distance from receiver i to source j. This describes signal propagation through a uniform non-dispersive medium. In the Fourier domain, Equation 1 results in a mixing matrix A( w) given by: A(w) = [~lle-jW~ ~12e-jW~ 1 -jw~ R21e c 1 R 22 e _jw!!JJ.. 1 (2) c It is important to note that the columns can be scaled arbitrarily without affecting separation of sources because rescaling is absorbed into the sources. This implies that row scaling in the demixing matrix (the inverse of A( is arbitrary. w? Using the Cosine Rule, Rij can be expressed in terms of the distance Rj of source j to the midpoint between two receivers, the direction of arrival of source j, and the distance between receivers, d, as follows: = [HJ + (~)' + 2(-1)' R;j m r 1 Hj COS OJ (3) Expanding the right term above using the binomial expansion and preserving only zeroth and first order terms, we can express distance from the receivers to the sources as: Rij = ( Rj + 8~j) + (_l)i (~) cosOj (4) This approximation is valid within a 5% relative error when d ::; ~. With the substitution for Rij and with the redefinition of source j to include the delay due to the term within brackets in Equation 4 divided by c, Equation 1 becomes: Xi(t) = ~ ~ij .Sj (t+(-l)i?(:c).cosOj ) , i= 1,2 (5) J In the Fourier domain, equation 5 results in the simplification to the mixing matrix A( w): A(w) = [ _1_ e-jwo1 R Il ? . _1_ eJW01 R21? _1_ e-jw02 ] Rl2 . _1_ ejw02 R 22' (6) 777 Broadband DOA Estimation Based on Second Order Statistics Here phases are functions of the directions of arrival ()j (defined with respect to the midpoint between receivers), the distance between receivers d, and the speed of propagation c: Oi 2dc cos ()i ,i 1, 2. Rij are unknown, but we can again redefine sources so diagonal elements are unity: = = (7) where c), C2 are two positive real numbers. In wireless communications sources are typically distant compared to antenna distance. For distant sources and a well matched pair of receivers c) ~ C2 ~ 1. Equation 7 describes the mixing matrix for the delay model in the frequency domain, in terms of four parameters, 0) ,02, c), C2. The corresponding ideal demixing matrix W(w), for each frequency w, is given by: W(w) _) 1 = [A(w) ] = detA(w) [e jW02 (8) -c2 .ejwol The outputs, estimating the sources, are: ] _ W w [X)(W) ] _ 1 [ [ z)(w) Z2(W) () X2(W) - detA(w) _c)e- jW02 ] [ e-; WO l x)(w) ] X2(W) (9) Making the transition back to the time domain results in the following estimate of the outputs: (10) where @ is convolution, and (11) Formulae 9 and 10 form the basis for two algorithms to be described next, in the time domain and the frequency domains. The algorithms have the role of determining the four unknown parameters. Note that the filter corresponding to H (w, 0) , 02, C), C2) should be applied to the output estimates in order to map back to the original inputs. 3 Delay and attenuation compensation algorithms The estimation of the four unknown parameters 0), 02, C), C2 can be carried out based on second order criteria that impose the constraint that outputs are decorrelated ([9, 4, 6, 5]). 3.1 Time and frequency domain approaches The time domain algorithm is based on the idea of imposing the decorrelation constraint (Z) (t), Z2(t)} 0 between the estimates ofthe outputs, as a function of the delays D) and D2 and scalar coefficients c) and C2. This is equivalent to the following criterion: = (12) where F(.) measures the cross-correlations between the signals given below, representing filtered versions of the differences of fractionally delayed measurements: J Rosca. J 778 = h(t, D), D2, e), e2) 0 Z2(t) = h(t, D) , D2, e) , e2) 0 Z)(t) 6 Ruanaidh. A. Jourjine and S. Rickard + D2) - e)X2(t?) (e2X) (t + D2) - X2(i?) (X)(t F(D), D2, e), e2) (13) = (Z)(t), Z2(t)} In the frequency domain, the cross-correlation of the inputs is expressed as follows: RX(w) = A(w)Rs (w)AH(w) ( 14) The mixing matrix in the frequency domain has the form given in Equation 7. Inverting this cross correlation equation yields four equations that are written in matrix form as: ( 15) Source orthogonality implies that the off-diagonal terms in the covariance matrix must be zero: RT2(W) =0 Rf)(w) = 0 (16) For far field conditions (i.e. the distance between the receivers is much less than the distance from sources) one obtains the following equations: = The terms a e- jw1h and b = e- jwoz are functions of the time delays. Note that there is a pair of equations of this kind for each frequency. In practice, the unknowns should be estimated from data at all available frequencies to obtain a robust estimate. 3.2 Channel selection Up to this point, there was no guarantee that estimated parameters would ensure source separation in some specific order. We could not decide a priori whether estimated parameters for the first output channel correspond to the first or second source. However, the dependence of the phase delays on the angles of arrival suggests a way to break the permutation symmetry in source estimation, that is to decide precisely which estimate to present on the first channel (and henceforth on the second channel as well). The core idea is that directionality and spatial cues provide the information required to break the symmetry. The criterion we use is to sort sources in order of increasing delay. Note that the correspondence between delays and sources is unique when sources are not symmetrical with respect to the receiver axis. When sources are symmetric there is no way of distinguishing between their positions because the cosine of the angles of arrival, and hence the delay, is invariant to the sign of the angle. 4 Experimental results A robust implementation of criterion 12 averages cross-correlations over a number of windows, of given size. More precisely F is defined as follows: F( 0),02) = L Blocks I(Z) (t), Z2(t)W ( 18) 779 Broadband DOA Estimation Based on Second Order Statistics Normally q = 1 to obtain a robust estimate. Ngo and Bhadkamkar [5] suggest a similar criterion using q = 2 without making use of the determinant of the mixing matrix. After taking into account all terms from Equation 18, including the determinant of the mixing matrix A, we obtain the function to be used for parameter estimation in the frequency domain: F(01,02) = ~ ~ I I q 1 2 ? -b a Rl1 x (W) - -R22(W) b x x (w) - -bRI2(w) 1 x - abR21 (19) w { det A} + TJ a a where TJ is a (Wiener Filter-like) constant that helps prevent singularities and q is normally set to one. Computing the separated sources using only time differences leads to highpass filtered outputs. In order to implement exactly the theoretical demixing procedure presented one has to divide by the determinant of the mixing matrix. Obviously one could filter using the inverse of the determinant to obtain optimal results. This can be implemented in the form of a Wiener filter. The Wiener filter requires knowledge both ofthe signal and noise power spectral densities. This information is not available to us but a reasonable approximation is to assume that the (wideband) sources have a flat spectral density and the noise corrupting the mixtures is white. In this case, the Wiener Filter becomes: H w _ ( ( )- {detA(W)}2) { det A (w )} 2 + TJ 1 det A (w ) (20) where the parameter TJ has been empirically set to the variance of the mixture. Applying this choice of filter usually dramatically improves the quality of the separated outputs. The technique of postprocessing using the determinant of the mixing matrix is perfectly general and applies equally well to demixtures computed using matrices of FIR filters. The quality of the result depends primarily on the care with which the inverse filter is implemented. It also depends on the accuracy of the estimate for the mixing parameters. One should avoid using the Wiener filter for near-degenerate mixtures. The proof of concept for the theory outlined above was obtained using speech signals which if anything pose a greater challenge to separation algorithms because of the correlation structure of speech. Two kinds of data are considered in this paper: synthetic direct propagation delay data and synthetic mUlti-path data. Data can be characterized along two dimensions of difficulty: synthetic vs. real-world, and direct path vs. multi-path. Combinations along these dimensions represented the main type of data we used. The value of distance between receivers dictates the order of delays that can appear due to direct path propagation, which is used by the demixing algorithms. Data was generated synthetically employing fractional delays corresponding to the various positions of the sources [3]. We modeled multi-path by taking into account the decay in signal amplitude due to propagation distance as well as the absorption of waves. Only the direct path and one additional path were considered. The algorithms developed proved successful for separation of two voices from direct path mixtures, even where the sources had very similar spectral power characteristics, and for separation of one source for multi-path mixtures. Moreover, outputs were free from artifacts and were obtained with modest computational requirements. Figure 1 presents mean separation results of the first and second channels, which correspond to the first and second sources, for various synthetic data sets. Separation depends on the angles of arrival. Plots show no separation in the degenerate case of equal or closeby angles of arrival, but more than lOdB mean separation in the anechoic case and 5dB in the mUlti-path case. 780 J. Rosca, J. 50 ,. .. . Ruanaidh, A. Jourjine and S. Rickard i f .. I I \/ i" I ,. .~ .. I" ,," I~ so / 1.." . :-..\ " '" f. I t -1?0 6 ') Doma1_ AnechoicT~ Anechoic F '00 50 .... -",so .. ., =~H "-1'- ... 210 I -6. .... 50 '00 ,. ,. i I -"- If ill ... ... ... ,,., .... .... :-' .... i,. i" I. 1. I j" :\ " f? I ,. , =t~Oomal so '00 .... -"'so 210 ... " I? ' .... =t~ so '00 ... ... -"'50 Figure 1: Two sources were positioned at a relatively large distance from a pair of closely spaced receivers. The first source was always placed at zero degrees whilst the second source was moved uniformly from 30 to 330 degrees in steps of 30 degrees. The above shows mean separation and standard deviation error bars of first and second sources for six synthetic delay mixtures or synthetic mUlti-path data mixtures using the time and frequency domain algorithms. 5 Conclusions The present source separation approach is based on minimization of cross-correlations of the estimated sources, in the time or frequency domains, when using a delay model and explicitly employing dirrection of arrival. The great advantage of this approach is that it reduces source separation to a decorrelation problem, which is theoretically solved by a system of equations. Although the delay model used generates essentially anechoic time delay algorithms, the results of this work show systematic improvements even when the algorithms are applied to real multi-path data. In all cases separation improvement is robust with respect to the power ratios of sources. Acknowledgments We thank Radu Balan and Frans Coetzee for useful discussions and proofreading various versions of this document and our collaborators within Siemens for providing extensive data for testing. Broadband DOA Estimation Based on Second Order Statistics 781 References [1] A. Jourjine, S. Rickard, J. 6 Ruanaidh, and J. Rosca. Demixing of anechoic time delay mixtures using second order statistics. Technical Report SCR-99-TR-657, Siemens Corporate Research, 755 College Road East, Princeton, New Jersey, 1999. [2] Hamid Krim and Mats Viberg. Two decades of array signal processing research. IEEE Signal Processing Magazine, 13(4), 1996. [3] Tim Laakso, Vesa Valimaki, Matti Karjalainen, and Unto Laine. Splitting the unit delay. IEEE Signal Processing Magazine, pages 30-60,1996. [4] L. Molgedey and H.G. Schuster. Separation of a mixture of independent signals using time delayed correlations. Phys.Rev.Lett., 72(23):3634-3637, July 1994. [5] T. J. Ngo and N.A. Bhadkamkar. Adaptive blind separation of audio sources by a physically compact device using second order statistics. In First International Workshop on leA and BSS, pages 257-260, Aussois, France, January 1999. [6] Lucas Parra, Clay Spence, and Bert De Vries. Convolutive blind source separation based on multiple decorrelation. In NNSP98, 1988. [7] K. Torkolla. Blind separation for audio signals: Are we there yet? In First International Workshop on Independent component analysis and blind source separation, pages 239244, Aussois, France, January 1999. [8] V. Van Veen and Kevin M. Buckley. Beamforrning: A versatile approach to spatial filtering. IEEE ASSP Magazine, 5(2), 1988. [9] E. Weinstein, M. Feder, and A. Oppenheim. Multi-channel signal separation by decorrelation. IEEE Trans. on Speech and Audio Processing, 1(4):405-413, 1993.
1680 |@word ruanaidh:5 determinant:5 version:2 proportion:1 d2:6 r:1 covariance:1 dramatic:1 tr:1 versatile:1 substitution:1 document:1 com:1 z2:5 yet:1 written:1 must:1 distant:2 designed:1 drop:1 plot:1 v:2 cue:2 pursued:1 device:1 core:1 filtered:2 along:2 c2:7 direct:6 weinstein:1 frans:1 redefine:1 theoretically:1 multi:9 compensating:1 window:1 considering:1 increasing:1 becomes:2 estimating:1 moreover:2 matched:1 medium:1 kind:2 r21:1 developed:2 whilst:1 nj:1 guarantee:1 attenuation:2 exactly:1 scaled:1 normally:2 unit:1 appear:1 positive:1 path:14 zeroth:1 suggests:1 co:2 wideband:2 emitter:1 unique:1 acknowledgment:1 testing:1 spence:1 practice:1 block:1 implement:1 procedure:1 veen:1 ri2:1 dictate:1 road:1 suggest:1 cannot:1 selection:2 applying:1 equivalent:1 map:1 layout:1 formulate:1 splitting:1 closeby:1 rule:1 array:3 rl1:1 magazine:3 distinguishing:1 element:2 role:1 rij:5 solved:1 complexity:1 purely:1 molgedey:1 basis:3 necessitates:1 represented:1 various:3 jersey:1 separated:3 fast:1 describe:2 kevin:1 solve:1 statistic:8 antenna:1 obviously:1 advantage:1 analytical:1 mixing:16 degenerate:2 anechoic:4 moved:1 requirement:1 demix:2 help:1 tim:1 pose:1 rt2:1 ij:1 implemented:3 implies:2 direction:5 closely:3 filter:11 collaborator:1 hamid:1 absorption:1 singularity:1 parra:1 strictly:1 considered:3 great:1 idealistic:1 estimation:8 successfully:1 minimization:1 sensor:1 always:2 avoid:1 hj:2 focus:1 improvement:2 indicates:1 typically:1 france:2 ill:1 priori:1 lucas:1 spatial:4 constrained:1 field:1 equal:1 sampling:1 report:1 feasibly:1 primarily:2 delayed:2 phase:2 highly:1 mixture:12 bracket:1 tj:4 capable:1 modest:1 divide:1 theoretical:4 column:1 deviation:1 uniform:1 delay:24 successful:1 synthetic:6 density:2 international:2 krim:1 systematic:1 off:2 again:1 recorded:2 fir:1 henceforth:1 rescaling:1 account:3 de:1 coefficient:1 inc:1 caused:1 explicitly:1 blind:5 depends:4 vesa:1 break:2 wave:2 sort:1 il:1 oi:1 accuracy:1 wiener:6 variance:1 characteristic:1 spaced:3 identify:1 ofthe:2 yield:1 directional:3 correspond:2 identification:1 rx:1 ah:1 simultaneous:1 oppenheim:1 phys:1 decorrelated:1 frequency:12 e2:3 proof:1 coetzee:1 proved:1 knowledge:1 fractional:4 dispersive:1 improves:1 clay:1 amplitude:2 positioned:1 back:2 jw:4 correlation:7 propagation:6 quality:2 artifact:2 resemblance:1 effect:1 concept:1 bhadkamkar:2 hence:1 ril:1 symmetric:1 white:1 essence:1 anything:1 cosine:2 criterion:5 generalized:1 scr:2 narrowband:1 postprocessing:1 instantaneous:1 physical:2 empirically:1 measurement:1 imposing:1 rd:1 outlined:1 had:1 doa:3 operating:1 viberg:1 deta:3 arbitrarily:1 preserving:1 greater:1 care:1 impose:1 additional:1 period:1 signal:20 july:1 multiple:4 corporate:2 rj:2 reduces:2 technical:1 characterized:1 cross:5 divided:1 post:1 equally:1 redefinition:1 essentially:2 physically:1 invert:1 lea:1 affecting:1 source:45 recording:1 db:1 beamforming:3 integer:1 ngo:2 near:1 ideal:1 synthetically:1 concerned:2 independence:1 architecture:1 perfectly:1 reduce:1 idea:2 det:3 whether:1 six:1 notch:1 feder:1 wo:1 speech:3 jj:1 buckley:1 generally:2 dramatically:1 detailed:1 useful:1 reduced:1 sign:1 estimated:4 r22:1 mat:1 aussois:2 express:1 four:4 fractionally:1 rosca:6 prevent:1 laine:1 angle:6 inverse:3 ameliorate:1 arrive:1 reasonable:1 decide:2 separation:22 scaling:1 simplification:1 correspondence:1 precisely:2 orthogonality:1 constraint:2 x2:4 flat:1 generates:1 fourier:2 speed:2 proofreading:1 relatively:1 radu:1 combination:1 describes:2 unity:1 joseph:1 rev:1 making:2 invariant:1 interference:1 equation:12 available:2 spectral:3 rl2:1 voice:1 original:2 binomial:1 include:1 ensure:1 dependence:1 diagonal:2 distance:13 separate:2 jourjine:6 separating:2 thank:1 reason:1 modeled:1 ratio:1 providing:1 unfortunately:1 e2x:1 implementation:1 unknown:4 karjalainen:1 convolution:1 compensation:1 january:2 communication:1 assp:1 dc:1 highpass:1 bert:1 arbitrary:1 princeton:2 inverting:1 pair:3 required:1 extensive:2 trans:1 address:1 bar:1 below:2 usually:1 scott:1 convolutive:1 summarize:1 challenge:1 rf:1 reliable:1 oj:1 including:1 power:3 decorrelation:4 difficulty:1 representing:1 axis:2 carried:1 literature:1 determining:1 relative:1 bear:1 permutation:1 filtering:2 foundation:2 degree:3 principle:1 corrupting:1 row:1 balan:1 placed:2 wireless:1 free:1 arriving:1 lle:1 taking:2 midpoint:2 van:1 bs:6 dimension:2 world:2 valid:1 transition:1 lett:1 unto:1 adaptive:3 far:1 employing:2 sj:1 obtains:1 compact:1 receiver:19 symmetrical:1 conclude:1 xi:2 decade:1 why:1 channel:7 matti:1 robust:5 expanding:1 symmetry:2 expansion:1 domain:14 main:2 s2:1 noise:4 arrival:10 broadband:4 aid:1 position:3 formula:1 specific:1 decay:1 demixing:7 workshop:2 rickard:6 vries:1 absorbed:1 expressed:2 scalar:1 applies:1 relies:1 experimentally:1 directionality:1 uniformly:1 experimental:2 siemens:4 east:1 college:2 alexander:1 audio:3 schuster:1
745
1,681
Dual Estimation and the Unscented Transformation EricA. Wan ericwan@ece.ogi.edu Rudolph van der Merwe rudmerwe@ece.ogi.edu Alex T. Nelson atneison@ece.ogi.edu Oregon Graduate Institute of Science & Technology Department of Electrical and Computer Engineering 20000 N.W. Walker Rd., Beaverton, Oregon 97006 Abstract Dual estimation refers to the problem of simultaneously estimating the state of a dynamic system and the model which gives rise to the dynamics. Algorithms include expectation-maximization (EM), dual Kalman filtering, and joint Kalman methods. These methods have recently been explored in the context of nonlinear modeling, where a neural network is used as the functional form of the unknown model. Typically, an extended Kalman filter (EKF) or smoother is used for the part of the algorithm that estimates the clean state given the current estimated model. An EKF may also be used to estimate the weights of the network. This paper points out the flaws in using the EKF, and proposes an improvement based on a new approach called the unscented transformation (UT) [3]. A substantial performance gain is achieved with the same order of computational complexity as that of the standard EKF. The approach is illustrated on several dual estimation methods. 1 Introduction We consider the problem of learning both the hidden states Xk and parameters w of a discrete-time nonlinear dynamic system, F(Xk , Vk, w) H(xk, nk, w), (1) (2) where Yk is the only observed signal. The process noise Vk drives the dynamic system, and the observation noise is given by nk. Note that we are not assuming additivity of the noise sources. A number of approaches have been proposed for this problem. The dual EKF algorithm uses two separate EKFs: one for signal estimation, and one for model estimation. The states are estimated given the current weights and the weights are estimated given the current states. In the joint EKF, the state and model parameters are concatenated within a combined state vector, and a single EKF is used to estimate both quantities simultaneously. The EM algorithm uses an extended Kalman smoother for the E-step, in which forward and 667 Dual Estimation and the Unscented Transformation backward passes are made through the data to estimate the signal. The model is updated during a separate M-step. For a more thorough treatment and a theoretical basis on how these algorithms relate, see Nelson [6]. Rather than provide a comprehensive comparison between the different algorithms, the goal of this paper is to point out the assumptions and flaws in the EKF (Section 2), and offer a improvement based on the unscented transformation/filter (Section 3). The unscented filter has recently been proposed as a substitute for the EKF in nonlinear control problems (known dynamic model) [3]. This paper presents new research on the use of the UF within the dual estimation framework for both state and weight estimation. In the case of weight estimation, the UF represents a new efficient "second-order" method for training neural networks in general. 2 Flaws in the EKF Assume for now that we know the model (weight parameters) for the dynamic system in Equations 1 and 2. Given the noisy observation Yk, a recursive estimation for Xk can be expressed in the form, Xk = (optimal prediction ofxk) + Gk x [Yk - (optimal prediction ofYk)] (3) This recursion provides the optimal MMSE estimate for Xk assuming the prior estimate Xk and current observation Yk are Gaussian. We need not assume linearity of the model. The optimal terms in this recursion are given by yl: = E[H(xl:, nk)], (4) where the optimal prediction xl: is the expectation of a nonlinear function of the random variables Xk-l and Vk-l (similar interpretation for the optimal prediction of Yk). The optimal gain term is expressed as a function of posterior covariance matrices (with Yk = Yk - Yl:) ? Note these terms also require taking expectations of a nonlinear function of the prior state estimates. The Kalman filter calculates these quantities exactly in the linear case. For nonlinear models, however, the extended KF approximates these as: YI: = H(xl:,fl), (5) where predictions are approximated as simply the function of the prior mean value for estimates (no expectation taken). The covariance are determined by linearizing the dynamic equations (Xk+l ~ AXk + BVk, Yk ~ CXk + Dnk), and then determining the posterior covariance matrices analytically for the linear system. As such, the EKF can be viewed as providing "first-order" approximations to the optimal terms (in the sense that expressions are approximated using a first-order Taylor series expansion of the nonlinear terms around the mean values). While "second-order" versions of the EKF exist, their increased implementation and computational complexity tend to prohibit their use. 3 The Unscented TransformationIFilter The unscented transformation (UT) is a method for calculating the statistics of a random variable which undergoes a nonlinear transformation [3] . Consider propagating a random variable a (dimension L) through a nonlinear function, (3 = g( a). Assume a has mean ct and covariance P Q . To calculate the statistics of {3, we form a matrix X of 2L + 1 sigma vectors Xi, where the first vector (Xo) corresponds to ct, and the rest are computed from the mean (+ )plus and (-)minus each column of the matrix square-root of P Q . These sigma E. A. Wan, R. v. d. Merwe and A. T. Nelson 668 vectors are propagated through the nonlinear function, and the mean and covariance for [3 are approximated using a weighted sample mean and covariance, /3 ~ - 1? {~g(xo) + -21 I:9(Xi)} +~ Pp" , (6) i=l L:~ {~[g(XO) - il][g(Xo) - il)T + ~ ~[g(X') - il)[g(X,) - ilf} (7) where ~ is a scaling factor. Note that this method differs substantially from general "sampling" methods (e.g., Monte-Carlo methods and particle filters [1]) which require orders of magnitude more sample points in an attempt to propagate an accurate (possibly nonGaussian) distribution of the state. The UT approximations are accurate to the third order for Gaussian inputs for all nonlinearities. For non-Gaussian inputs, approximations are accurate to at least the second-order, with the accuracy determined by the choice of ~ [3]. A simple example is shown in Figure 1 for a 2-dimensional system: the left plots shows the true mean and covariance propagation using Monte-Carlo sampling; the center plots show the performance of the UT (note only 5 sigma points are required); the right plots show the results using a linearization approach as would be done in the EKF. The superior performance of the UT is clear. Actual (sampling) mean (3 I = g(o) 1 UT Linearized (EKF) .11"=- 0 -I (3 Yi = g(X i ) 1 = g(o) P(:! = ATPaA l Figure 1: Example of the UT for mean and covariance propagation. a) actual, b) UT, c) first-order linear (EKF). The unscented filter (UF) [3] is a straightforward extension of the UT to the recursive estimation in Equation 3, where we set 0: = Xk, and denote the corresponding sigma matrix as X(klk). The UF equations are given on the next page. It is interesting to note that no explicit calculation of lacobians or Hessians are necessary to implement this algorithm. The total number of computations is only order ?2 as compared to ?3 for the EKF. I 4 Application to Dual Estimation This section shows the use of the UF within several dual estimation approaches. As an application domain for comparison, we consider modeling a noisy time-series as a nonlinear INote that a matrix square-root using the Cholesky factorization is of order L 3 /6. However, the covariance matrices are expressed recursively, and thus the square-root can be computed in only order L2 by performing a recursive update to the Cholesky factorization. 669 Dual Estimation and the Unscented Transformation UF Equations Wo = K/(L + K) WI . .. W2L = 1/2(L + fl.) , X(klk - 1) = F[X(k - 11k - 1), P!lv2 ] x"k = 2:~!o WiXi(klk - 1) P"k = 2:~!o WdXi(klk - 1) - x"k][Xi(klk - 1) - x"kf Y(klk - 1) = H[X(klk - 1), P~;] Y"k = 2:~!o WiYi(klk - 1) P XkYk = 2::!o Wi[Yi(klk 2::!o Wi[Xi(klk - Xk = x"k + PXkYkP~:h (n - P hh = 1) - Y"kJ[Yi(klk - 1) - 1) - x"k][Yi(klk - 1) - Yk"f Y"kf Y"k) P k = P"k - PX"Yk(P~:yJTP!'kYk autoregression: Xk = f(Xk-l, ... Xk-M, Yk = Xk w) + Vk + nk, Vk E {l. . . N} (8) The underlying clean signal Xk is a nonlinear function of its past M values, driven by white Gaussian process noise Vk with variance 11;. The observed data point Yk includes the additive noise nk, which is assumed to be Gaussian with variance 11;. The corresponding state-space representation for the signal Xk is given by: + B? Vk-I Yk = [1 (9) (10) 0 In this context, the dual estimation problem consists of simultaneously estimating the clean signal Xk and the model parameters w from the noisy data Yk. 4.1 Dual EKF I Dual UF One dual estimation approach is the dual extended Kalman filter developed in [8, 6]. The dual EKF requires separate state-space representation for the signal and the weights. A state-space representation for the weights is generated by considering them to be a stationary process with an identity state transition matrix, driven by process noise Uk: + Uk Wk = Wk-l Yk = f(Xk-I,Wk) +Vk +nk? (11) (12) The noisy measurement Yk has been rewritten as an observation on w. This allows the use of an EKF for weight estimation (representing a "second-order" optimization procedure) [7]. Two EKFs can now be run simultaneously for signal and weight estimation. At every time-step, the current estimate of the weights is used in the signal-filter, and the current estimate of the signal-state is used in the weight-filter. 670 E. A. Wan, R. v. d. Merwe and A. T Nelson The dual UFIEKF algorithm is formed by simply replacing the EKF for state-estimation with the UF while still using an EKF for weight-estimation. In the dual UF algorithm both state- and weight-estimation are done with the UF. Note that the state-transition is linear in the weight filter, so the nonlinearity is restricted to the measurement equation. Here, the UF gives a more exact measurement-update phase of estimation. The use of the UF for weight estimation in general is discussed in further detail in Section 5. 4.2 Joint EKF I Joint UF An alternative approach to dual estimation is provided by the joint extended Kalman filter [4,5]. In this framework the signal-state and weight vector are concatenated into a single, joint state vector: Zk = The estimation of Zk can be done recursively by writing the state-space equations for the joint state as: [xf wfV. (13) and running an EKF on the joint state-space to produce simultaneous estimates of the states Xk and w . As discussed in [6], the joint EKF provides approximate MAP estimates by maximizing the joint density of the signal and weights given the noisy data. Again, our approach in this paper is to use the UF instead of the EKF to provide more accurate estimation of the state, resulting in the joint UF algorithm. 4.3 EM - Unscented Smoothing A somewhat different iterative approach to dual estimation is given by the expectationmaximization (EM) algorithm applied to nonlinear dynamic systems [2]. In each iteration, the conditional expectation of the signal is computed, given the data and the current estimate of the model (E-step). Then the model is found that maximizes a function of this conditional mean (M-step). For linear models, the M-step can be solved in closed form . The E-step is computed with a Kalman smoother, which combines the forward-time estimated mean and covariance (x{ ,pt) of the signal given past data, with the backward-time predicted mean and covariance (xf ,pf) given the future data, producing the following smoothed statistics given all the data: (14) (15) When a MLP neural network model is used, the M-step can no longer be computed in closed-form, and a gradient-based approach is used instead. The resulting algorithm is usually referred to as generalized EM (GEM) 2. The E-step is typically approximated by an extended Kalman smoother, wherein a linearization of the model is used for backward propagation of the state estimates. We propose improving the E-step of the EM algorithm for nonlinear models by using a UP instead of an EKF to compute both the forward and backward passes in the Kalman smoother. Rather than linearize the model for the backward pass, as in [2], a neural network is trained on the backward dynamics (as well as the forward dynamics). This allows for a more exact backward estimation phase using the UF, and enables the development of an unscented smoother (US). 2 An exact M-step is possible using RBF networks [2]. Dual Estimation and the Unscented Transformation 671 4.4 Experiments We present results on two simple time-series to provide a clear illustration of the use of the UP over the EKE The first series is the Mackey-Glass chaotic series with additive WGN (SNR ~ 3dB). The second time series (also chaotic) comes from an autoregressive neural network with random weights driven by Gaussian process noise and also corrupted by additive WGN (SNR ~ 3dB). A standard 5-3-1 MLP with tanh hidden activation functions and a linear output layer was used in all the filters. The process and measurement noise variances were assumed to be known. Results on training and testing data, as well as training curves for the different dual estimation methods are shown below. The quoted numbers are normalized (clean signal variance) mean-square estimation and prediction errors. The superior performance of the UT based algorithms (especially the dual UF) are clear. Note also the more stable learning curves using the UF approaches. These improvements have been found to be consistent and statistically significant on a number of additional experiments. Mackey-Glass Algorithm Dual EKF Dual UF/EKF Dual UF Joint EKF Joint UF Train Est. Pred. 0.20 0.50 0.19 0.50 0.15 0.45 0.22 0.53 0.19 0.50 Test Est. Pred. 0.21 0.54 0.19 0.53 0.14 0.48 0.22 0.56 0.18 0.53 Chaotic AR-NN Algorithm . Dual EKF Dual UF/EKF Dual UF Joint EKF Joint UF Test Est. Pred. 0.69 0.36 0.28 0.69 0.27 0.63 0.34 0.72 0.30 0.67 Train Est. Pred. 0.32 0.62 0.26 0.58 0.23 0.55 0.29 0.58 0.25 0.55 Chaotic AR-NN Mackey-Glass .. o. 0.' 01 ~O8 Dual EKF Dual UFIEKF 0 Dual UF 0 JointEKF " ...... UF W en :::;; :::;; o. ]05 "~" . '5E035 " E go" C 03 03 0 25 02 0.1 5 0 ? 10 11 02 0 10 15 ~eration iteration 20 2S '" The final table below compares smoother performance used for the E-step in the EM algorithm. In this case, the network models are trained on the clean time-series, and then tested on the noisy data using either the standard Kalman smoother with linearized backward model (EKS 1), a Kalman smoother with a second nonlinear backward model (EKS2), and the unscented smoother (US). The forward (F), backward (B), and smoothed (S) estimation errors are reported. Again the performance benefits of the unscented approach is clear. Mackey-Glass Algorithm EKSI EKS2 US Norm. MSE B S 0.20 0.70 0.27 0.19 0.20 0.3] 0.]0 0.24 0.08 F Chaotic AR-NN Algorithm EKSI EKS2 US Norm. MSE B S 0.35 0.32 0.28 0.35 0.22 0.23 0.23 0.2] 0.16 F 5 UF Neural Network Training As part of the dual UF algorithm, we introduced the use of the UF for weight estimation. The approach can also be seen as a new method for the general problem of training neural networks (i.e., for regression or classification problems where the input x is observed and 672 E. A. Wan. R. v. d. Merwe and A. T. Nelson no state-estimation is required). The advantage of the UF over the EKF in this case is not as obvious, as the state-transition function is linear (See Equation 11). However, as pointed out earlier, the observation is nonlinear. Effectively, the EKF builds up an approximation to the expected Hessian by taking outer products of the gradient. The UF, however, may provide a more accurate estimate through direct approximation of the expectation of the Hessian. We have performed a number of preliminary experiments on standard benchmark data. The figure below shows the mean and std. oflearning curves (computed over 100 experiments with different initial weights) for the Mackay Robot Arm Mapping dataset. Note the faster convergence, lower variance, and lower final MSE performance of the UF weight training. While these results are encouraging, further study is still necessary to fully contrast differences between UF and EKF weight training. O.06n---.-----,"L"le;O;anmW ln;:;r,g1"it.;'"w lu rv:v:ec.s,..---r= I::;;U~F(;:=m.== an I r 7=>l) 0 .05 - UF(. ld) : : ~~~ ~~~n) W ~ 0 .004 c: ~ 0,03 E 0 .02 0.01 '\ ~_. -..~ __ ~ "--- __ -.. _ _ _ _ _ _ _ _ _ _ _ _ _ _ ? _ _ ...... - - .. .. 6 Conclusions The EKF has been widely accepted as a standard tool in the machine learning community. In this paper we have presented an alternative to the EKF using the unscented filter. The UF consistently achieves a better level of accuracy than the EKF at a comparable level of complexity. We demonstrated this performance gain on a number of dual estimation methods as well as standard regression modeling. Acknowledgements This work was sponsored in part by the NSF under grant IRI-9712346. References [1] J. F. G. de Freitas, M. Niranjan, A. H. Gee, and A. Doucet. Sequential Monte Carlo methods for optimisation of neural network models. Technical Report TR-328, Cambridge University Engineering Department, Cambridge, England, November 1998. [2] Z. Ghahramani and S. T. Roweis. Learning nonlinear dynamical systems using an EM algorithm. In M. J. Keams, S. A. Solla, and D. A. Cohn, editors, Advances in Neural Information Processing Systems II: Proceedings of the 1998 Conference. MIT Press, 1999. [3] S. J. Julier and J. K. Uhlmann. A New Extension of the Kalman Filter to Nonlinear Systems. In Proc. of AeroSense: The 11th International Symposium on Aerospace/Defence Sensing. Simulation and Controls. Orlando. Florida., 1997. [4] R. E. Kopp and R. J. Orford. Linear regression applied to system identification for adaptive control systems. AlAA 1., I :2300-06, October 1963. [5] M. B. Matthews and G. S. Moschytz. Neural-network nonlinear adaptive filtering using the extended Kalman filter algorithm. In INNC, pages 115-8, 1990. [6] A. T. Nelson. Nonlinear Estimation and Modeling of Noisy Time-Series by Dual Kalman Filtering Methods. PhD thesis, Oregon Graduate Institute, 1999. In preparation. [7] S. Singhal and L. Wu. Training multilayer perceptrons with the extended Kalman filter. In Advances in Neural Information Processing Systems 1, pages 133-140, San Mateo, CA, 1989. Morgan Kauffman. [8] E. A. Wan and A. T. Nelson. Dual Kalman filtering methods for nonlinear prediction, estimation, and smoothing. In Advances in Neural Information Processing Systems 9, 1997.
1681 |@word version:1 norm:2 simulation:1 propagate:1 linearized:2 covariance:11 wgn:2 minus:1 tr:1 klk:12 recursively:2 ld:1 initial:1 series:8 mmse:1 past:2 freitas:1 current:7 activation:1 additive:3 enables:1 plot:3 sponsored:1 update:2 mackey:4 stationary:1 kyk:1 xk:20 provides:2 direct:1 symposium:1 consists:1 combine:1 expected:1 eke:1 actual:2 encouraging:1 pf:1 considering:1 provided:1 estimating:2 linearity:1 underlying:1 maximizes:1 substantially:1 developed:1 transformation:8 thorough:1 every:1 exactly:1 uk:2 control:3 grant:1 producing:1 engineering:2 plus:1 mateo:1 factorization:2 graduate:2 statistically:1 testing:1 recursive:3 implement:1 differs:1 chaotic:5 procedure:1 refers:1 context:2 writing:1 ilf:1 map:1 demonstrated:1 center:1 maximizing:1 straightforward:1 go:1 iri:1 updated:1 pt:1 exact:3 us:2 approximated:4 std:1 observed:3 electrical:1 solved:1 calculate:1 solla:1 yk:16 substantial:1 complexity:3 dynamic:10 trained:2 basis:1 joint:15 additivity:1 train:2 monte:3 widely:1 statistic:3 g1:1 rudolph:1 noisy:7 final:2 advantage:1 propose:1 product:1 roweis:1 convergence:1 produce:1 linearize:1 propagating:1 expectationmaximization:1 predicted:1 come:1 filter:16 require:2 orlando:1 preliminary:1 extension:2 unscented:15 around:1 mapping:1 matthew:1 achieves:1 estimation:37 proc:1 tanh:1 uhlmann:1 tool:1 weighted:1 kopp:1 mit:1 gaussian:6 defence:1 ekf:39 rather:2 eks:1 improvement:3 vk:8 consistently:1 contrast:1 sense:1 glass:4 flaw:3 nn:3 typically:2 hidden:2 keams:1 dual:36 classification:1 proposes:1 development:1 smoothing:2 mackay:1 sampling:3 represents:1 future:1 report:1 simultaneously:4 comprehensive:1 phase:2 attempt:1 mlp:2 accurate:5 necessary:2 taylor:1 theoretical:1 increased:1 column:1 modeling:4 earlier:1 ar:3 maximization:1 oflearning:1 singhal:1 snr:2 reported:1 corrupted:1 combined:1 density:1 international:1 yl:2 nongaussian:1 again:2 thesis:1 wan:5 possibly:1 nonlinearities:1 de:1 wk:3 includes:1 oregon:3 performed:1 root:3 closed:2 square:4 il:3 accuracy:2 formed:1 merwe:4 variance:5 identification:1 lu:1 carlo:3 drive:1 simultaneous:1 pp:1 obvious:1 propagated:1 gain:3 dataset:1 treatment:1 wixi:1 ut:10 wherein:1 done:3 replacing:1 axk:1 nonlinear:21 cohn:1 propagation:3 undergoes:1 dnk:1 normalized:1 true:1 analytically:1 illustrated:1 white:1 ogi:3 during:1 prohibit:1 linearizing:1 generalized:1 recently:2 superior:2 functional:1 julier:1 discussed:2 interpretation:1 approximates:1 measurement:4 significant:1 cambridge:2 rd:1 pointed:1 particle:1 nonlinearity:1 stable:1 robot:1 longer:1 posterior:2 driven:3 der:1 yi:5 seen:1 morgan:1 additional:1 somewhat:1 signal:15 ii:1 smoother:10 rv:1 technical:1 xf:2 england:1 faster:1 calculation:1 offer:1 niranjan:1 calculates:1 prediction:7 regression:3 multilayer:1 optimisation:1 expectation:6 iteration:2 achieved:1 walker:1 source:1 rest:1 pass:2 tend:1 db:2 expression:1 wo:1 hessian:3 clear:4 exist:1 nsf:1 estimated:4 discrete:1 clean:5 backward:10 run:1 aerosense:1 wu:1 scaling:1 comparable:1 fl:2 ct:2 layer:1 alex:1 performing:1 px:1 uf:35 department:2 em:8 wi:3 inote:1 restricted:1 xo:4 taken:1 ln:1 equation:8 hh:1 know:1 autoregression:1 rewritten:1 alternative:2 florida:1 substitute:1 running:1 include:1 beaverton:1 calculating:1 concatenated:2 ghahramani:1 especially:1 build:1 quantity:2 gradient:2 separate:3 outer:1 nelson:7 assuming:2 kalman:17 illustration:1 providing:1 october:1 relate:1 gk:1 sigma:4 rise:1 implementation:1 unknown:1 observation:5 benchmark:1 november:1 ericwan:1 extended:8 smoothed:2 community:1 pred:4 introduced:1 required:2 aerospace:1 usually:1 below:3 dynamical:1 kauffman:1 wiyi:1 eration:1 recursion:2 arm:1 representing:1 technology:1 kj:1 prior:3 l2:1 acknowledgement:1 kf:3 determining:1 fully:1 interesting:1 filtering:4 consistent:1 o8:1 editor:1 gee:1 institute:2 taking:2 van:1 benefit:1 curve:3 dimension:1 transition:3 autoregressive:1 forward:5 made:1 adaptive:2 san:1 ec:1 erica:1 approximate:1 doucet:1 assumed:2 gem:1 xi:4 quoted:1 iterative:1 table:1 zk:2 ca:1 improving:1 expansion:1 mse:3 domain:1 noise:8 referred:1 en:1 explicit:1 xl:3 third:1 sensing:1 explored:1 sequential:1 effectively:1 phd:1 magnitude:1 ekfs:2 linearization:2 nk:6 bvk:1 simply:2 expressed:3 corresponds:1 conditional:2 goal:1 viewed:1 identity:1 rbf:1 determined:2 called:1 total:1 pas:1 ece:3 accepted:1 lv2:1 est:4 perceptrons:1 alaa:1 cholesky:2 preparation:1 tested:1
746
1,682
Independent Factor Analysis with Temporally Structured Sources Hagai Attias hagai@gatsby.ucl.ac.uk Gatsby Unit, University College London 17 Queen Square London WCIN 3AR, U.K. Abstract We present a new technique for time series analysis based on dynamic probabilistic networks. In this approach, the observed data are modeled in terms of unobserved, mutually independent factors, as in the recently introduced technique of Independent Factor Analysis (IFA). However, unlike in IFA, the factors are not Li.d.; each factor has its own temporal statistical characteristics. We derive a family of EM algorithms that learn the structure of the underlying factors and their relation to the data. These algorithms perform source separation and noise reduction in an integrated manner, and demonstrate superior performance compared to IFA. 1 Introduction The technique of independent factor analysis (IFA) introduced in [1] provides a tool for modeling L'-dim data in terms of L unobserved factors. These factors are mutually independent and combine linearly with added noise to produce the observed data. Mathematically, the model is defined by Yt = HXt + Ut, where Xt is the vector of factor activities at time t, L' x L mixing matrix, and Ut is the noise. (1) Yt is the data vector, H is the The origins of IFA lie in applied statistics on the one hand and in signal processing on the other hand. Its statistics ancestor is ordinary factor analysis (FA), which assumes Gaussian factors. In contrast, IFA allows each factor to have its own arbitrary distribution, modeled semi-parametrically by a I-dim mixture of Gaussians (MOG). The MOG parameters, as well as the mixing matrix and noise covariance matrix, are learned from the observed data by an expectation-maximization (EM) algorithm derived in [1]. The signal processing ancestor of IFA is the independent component analysis (ICA) method for blind source separation [2]-[6]. In ICA, the factors are termed sources, and the task of blind source separation is to recover them from the observed data with no knowledge of the mixing process. The sources in ICA have non-Gaussian distributions, but unlike in IFA these distributions are usually fixed by prior knowledge or have quite limited adaptability. More significant restrictions Dynamic Independent Factor Analysis 387 are that their number is set to the data dimensionality, i.e. L = L' ('square mixing'), the mixing matrix is assumed invertible, and the data are assumed noise-free (Ut = 0). In contrast, IFA allows any L, L' (including more sources than sensors, L > L'), as well as non-zero noise with unknown covariance. In addition, its use of the flexible MOG model often proves crucial for achieving successful separation [1]. Therefore, IFA generalizes and unifies FA and ICA. Once the model has been learned, it can be used for classification (fitting an IFA model for each class), completing missing data, and so on. In the context of blind separation, an optimal reconstruction of the sources Xt from data is obtained [1] using a MAP estimator. However, IFA and its ancestors suffer from the following shortcoming: They are oblivious to temporal information since they do not attempt to model the temporal statistics of the data (but see [4] for square, noise-free mixing). In other words, the model learned would not be affected by permuting the time indices of {yt}. This is unfortunate since modeling the data as a time series would facilitate filtering and forecasting, as well as more accurate classification. Moreover, for source separation applications, learning temporal statistics would provide additional information on the sources, leading to cleaner source reconstructions. To see this, one may think of the problem of blind separation of noisy data in terms of two components: source separation and noise reduction. A possible approach might be the following two-stage procedure. First, perform noise reduction using, e.g., Wiener filtering. Second, perform source separation on the cleaned data using, e.g., an ICA algorithm. Notice that this procedure directly exploits temporal (second-order) statistics of the data in the first stage to achieve stronger noise reduction. An alternative approach would be to exploit the temporal structure of the data indirectly, by using a temporal source model. In the resulting single-stage algorithm, the opemtions of source sepamtion and noise reduction are coupled. This is the approach taken in the present paper. In the following, we present a new approach to the independent factor problem based on dynamic probabilistic networks. In order to capture temporal statistical properties of the observed data, we describe each source by a hidden Markov model (HMM). The resulting dynamic model describes a multivariate time series in terms of several independent sources, each having its own temporal characteristics. Section 2 presents an EM learning algorithm for the zero-noise case, and section 3 presents an algorithm for the case of isotropic noise. The case of non-isotropic noise turns out to be computationally intractable; section 4 provides an approximate EM algorithm based on a variational approach. Notation: The multivariable Gaussian density is denoted by g(z, E) =1 27rE 1- 1 / 2 exp( -z T E- l z/2). We work with T-point time blocks denoted Xl:T = {Xt}[=I' The ith coordinate of Xt is x~. For a function f, (f(Xl:T)) denotes averaging over an ensemble of Xl:T blocks. 2 Zero Noise The MOG source model employed in IFA [1] has the advantages that (i) it is capable of approximating arbitrary densities, and (ii) it can be learned efficiently from data by EM. The Gaussians correspond to the hidden states of the sources, labeled by s. Assume that at time t, source i is in state s~ = s. Its signal x~ is then generated by sampling from a Gaussian distribution with mean JL! and variance v!. In order to capture temporal statistics of the data, we endow the sources with temporal structure by introducing a transition matrix a!,s between the states. Focusing on H. Attias 388 = 1, ... , T, the resulting probabilistic model is defined by i P( Sti=i Si St-l = S') = as's' P( Soi = S) = 7rsi , a time block t p(X~ IS~ = S) = g(x~ - J.L!,v!), P(Yl:T) =1 detG IT P(Xl:T), (2) xL where P(Xl:T) is the joint density of all sources i = 1, ... , L at all time points, and the last equation follows from Xt = GYt with G = H- 1 being the unmixing matrix. As usual in the noise-free scenario (see [2]; section 7 of [1]), we are assuming that the mixing matrix is square and invertible. The graphical model for the observed density P(Yl:T I W) defined by (2) is parametrized by W = {G ij , J.L!, v!, 7r!, a!, s}' This model describes each source as a first-order HMM; it reduces to a time-independent model if a!,s = 7r!. Whereas temporal structure can be described by other means, e.g. a moving-average [4] or autoregressive [6] model, the HMM is advantageous since it models high-order temporal statistics and facilitates EM learning. Omitting the derivation, maximization with respect to G ij results in the incremental update rule bG = ?G - 1 T ?T </>(Xt)x[G , L (3) t=l where </>(xn = Es 'Y:(s)(x~ - J.L!)/v!, and the natural gradient [3] was used; ? is an appropriately chosen learning rate. For the source parameters we obtain the update rules Et 'Yt(s)x~ Et 'Y1(s) i , as' s _ - t ~t( s' , s) i (')' u t 'Yt-l S E ~ (4) with the initial probabilities updated via 7r! = 'YA(s). We used the standard HMM notation 'Y:(s) = p(s~ = S I xLT)' ~t(s',s) = P(SLI = s',s~ = s I xLT)' These posterior densities are computed in the E-step for each source, which is given in terms of the data via x~ = E j Gijyl, using the forward-backward procedure [7]. The algorithm (3-4) may be used in several possible generalized EM schemes. An efficient one is given by the following two-phase procedure: (i) freeze the source parameters and learn the separating matrix G using (3); (ii) freeze G and learn the source parameters using (4), then go back to (i) and repeat. Notice that the rule (3) is similar to a natural gradient version of Bell and Sejnowski's leA rule [2]; in fact, the two coincide for time-independent sources where </>(Xi) = -alogp(xi)/axi. We also recognize (4) as the Baum-Welch method. Hence, in phase (i) our algorithm separates the sources using a generalized leA rule, whereas in phase (ii) it learns an HMM for each source. Remark. Often one would like to model a given L'-variable time series in terms of a smaller number L ~ L' of factors. In the framework of our noise-free model Yt = HXt, this can be achieved by applying the above algorithm to the L largest principal components of the data; notice that if the data were indeed generated by L factors, the remaining L' - L principal components would vanish. Equivalently, one may apply the algorithm to the data directly, using a non-square L x L' unmixing matrix G. Results. Figure 1 demonstrates the performance of the above method on a 4 x 4 mixture of speech signals, which were passed through a non-linear function to modify their distributions. This mixture is inseparable to leA because the source model used by the latter does not fit the actual source densities (see discussion in [1]). We also applied our dynamic network to a mixture of speech signals whose distributions Dynamic Independent Factor Analysis 389 HMM-ICA leA 3 3 0 .8 2 0 .7 0 .8 ):i"0.5 '>:! 'zs:O.4 0 0.3 -1 0 .2 -2 0.1 0 -4 -3 -3 -2 4 -2 o 2 -2 x1 0 2 x1 Figure 1: Left: Two of the four source distributions. Middle: Outputs of the EM algorithm (3-4) are nearly independent. Right: the outputs of leA (2) are correlated. were made Gaussian by an appropriate non-linear transformation. Since temporal information is crucial for separation in this case (see [4],[6]), this mixture is inseparable to leA and IFA; however, the algorithm (3-4) accomplished separation successfully. 3 Isotropic Noise We now turn to the case of non-zero noise Ut ::j:. O. We assume that the noise is white and has a zero-mean Gaussian distribution with covariance matrix A. In general, this case is computationally intractable (see section 4). The reason is that the Estep requires computing the posterior distribution P(SO:T, Xl:T I Yl:T) not only over the source states (as in the zero-noise case) but also over the source signals, and this posterior has a quite complicated structure. We now show that if we assume isotropic noise, i.e. Aij = )..6ij , as well as square invertible mixing as above, this posterior simplifies considerably, making learning and inference tractable. This is done by adapting an idea suggested in [8] to our dynamic probabilistic network. We start by pre-processing the data using a linear transformation that makes their covariance matrix unity, i.e., (YtyT) = I ('sphering'). Here (-) denotes averaging over T-point time blocks. From (1) it follows that HSHT = )..'1, where S = (XtxT) is the diagonal covariance matrix of the sources, and )..' = 1 -)... This, for a square invertible H, implies that HTH is diagonal. In fact, since the unobserved sources can be determined only to within a scaling factor, we can set the variance of each source to unity and obtain the orthogonality property HTH = )..'1. It can be shown that the source posterior now factorizes into a product over the individual sources, P(SO:T, Xl :T I Yl:T) = TIiP(sb:T, XLT I Yl:T), where P(Sb:T,xLT I Yl:T) ()( [rrg(X; -T):'aD ? v;p(s: I SLl)] vbp(sb)? (5) t=l The means and variances at time t in (5), as well as the quantities vL depend on both the data Yt and the states s~; in particular, T); = (~j Hjiyl + )..j1!)/(>..'vs +)..) and a-; = )..v!/(>..'vs + )..), using s = s1; the expression for the v; are omitted. The transition probabilities are the same as in (2). Hence, the posterior distribution (5) effectively defines a new HMM for each source, with yrdependent emission and transition probabilities. To derive the learning rule for H, we should first compute the conditional mean Xt of the source signals at time t given the data. This can be done recursively using (5) as in the forward-backward procedure. We then obtain 1 T c= T~YtXr. t=l (6) H. Attias 390 This fractional form results from imposing the orthogonality constraint HTH = >..'1 using Lagrange multipliers and can be computed via a diagonalization procedure. The source parameters are computed using a learning rule (omitted) similar to the noise-free rule (4). It is easy to derive a learning rule for the noise level ,\ as well; in fact, the ordinary FA rule would suffice. We point out that, while this algorithm has been derived for the case L = L', it is perfectly well defined (though sub-optimal: see below) for L :::; L'. Non-Isotropic Noise 4 The general case of non-isotropic noise and non-square mixing is computationally intractable. This is because the exact E-step requires summing over all possible source configurations (st, ... , SfL) at all times tl, ... , tL = 1, ... , T. The intractability problem stems from the fact that, while the sources are independent, the sources conditioned on a data vector Yl:T are correlated, resulting in a large number of hidden configurations. This problem does not arise in the noise-free case, and can be avoided in the case of isotropic noise and square mixing using the orthogonality property; in both cases, the exact posterior over the sources factorizes. The EM algorithm derived below is based on a variational approach. This approach was introduced in [9J in the context of sigmoid belief networks, but constitutes a general framework for ML learning in intractable probabilistic networks; it was used in a HMM context in [IOJ. The idea is to use an approximate but tractable posterior to place a lower bound on the likelihood, and optimize the parameters by maximizing this bound. A starting point for deriving a bound on the likelihood L is Neal and Hinton's [l1J formulation of the EM algorithm: T ~ L L Eq logp(Yt I Xt) + L Eq logp(sb:T' xi:T) - Eq logq, (7) t=l i=l where Eq denotes averaging with respect to an arbitrary posterior density over the hidden variables given the observed data, q = q(SO:T,Xl:T I Yl:T). Exact EM, as shown in [11], is obtained by maximizing the bound (7) with respect to both the posterior q (corresponding to the E-step) and the model parameters W (Mstep). However, the resulting q is the true but intractable posterior. In contrast, in variational EM we choose a q that differs from the true posterior, but facilitates a tractable E-step. L = lOgp(Yl:T) E-Step. We use q(sO:T,Xl:T I Yl :T) = parametrized as q(s~ = s I SLI = S',Yl:T) ex: q(Xt IYl :T) = TIiq(sb:T I Yl:T)TItq(Xt I Yl:T), q(sb = s I Yl :T) ex: '\!,ta!,s, Q(Xt - Pt, ~t) . ,\! ,t 7r! , (8) Thus, the variational transition probabilities in (8) are described by multiplying the original ones a!, s by the parameters '\~,t' subject to the normalization constraints. The source signals Xt at time t are jointly Gaussian with mean Pt and covariance ~t. The means, covariances and transition probabilities are all time- and datadependent, i.e., Pt = f(Yl:T, t) etc. This parametrization scheme is motivated by the form of the posterior in (5); notice that the quantities v~ ,t there become the variational parameters ~;j,,\~ t of (8). A related scheme was used in [IOJ in a different context. Since these parameters will be adapted independently of the model parameters, the non-isotropic algorithm is expected to give superior results compared to the isotropic one. pL 1]:, a-t, 391 Dynamic Independent Factor Analysis Mixing Reco nstruc tion O ~--------~------~ 5 -10 0 ~-20 ~ .L3Cfl>___~O ~ L.U -40 0 -5 0 -10 0 - 15 -5~S;----:0:------::5'-------:-::'0:------:-'? 15 -20 - 5 0 SNA (dB) 5 10 15 SNR (dB) Figure 2: Left: quality of the model parameter estimates. Right: quality of the source reconstructions. (See text). Of course, in the true posterior the Xt are correlated, both temporally among themselves and with St, and the latter do not factorize. To best approximate it, the variational parameters V = {p~, ~~j , >..! t} are optimized to maximize the bound on .c, or equivalently to minimize the KL' distance between q and the true posterior. This requirement leads to the fixed point equations Pt (HT A -lH + Bt)-l(HT A -lYt 1 [1 . --:- exp - - log V Z zZt 2 _ (pi _ t s + b t ), J-Li)2 + ~ii] s . t 2vZs ~t = (HT A-1H + Bt)-l , , (9) where Bij = Ls[rl(S)/v!]6ij , b~ = Ls ,l(s)J-L!/v!, and the factors zf ensure normalization. The HMM quantities ,f(s) are computed by the forward-backward procedure using the variational transition probabilities (8). The variational parameters are determined by solving eqs. (9) iteratively for each block Yl :T; in practice, we found that less then 20 iterations are usually required for convergence. M-Step. The update rules for W are given for the mixing parameters by 1~ T T T A = T L,)YtYt - YtPt H ), (10) t and for the source parameters by Lt ,f(s)p~ Lt ,I(s) , Lt ~f(s', s) Lt,Ll(S') , Vi s = Lt ,f(s)((p~ - J-L~)2 + ~~i) Lt ,f(s) (11) where the ~Hs' , s) are computed using the variational transition probabilities (8). Notice that the learning rules for the source parameters have the Baum-Welch form, in spite of the correlations between the conditioned sources. In our variational approach, these correlations are hidden in V, as manifested by the fact that the fixed point equations (9) couple the parameters V across time points (since ,:(s) depends on >"!,t=l:T) and sources. Source Reconstruction. From q(Xt I Yl :T) (8), we observe that the MAP source estimate is given by Xt = Pt(Yl:T), and depends on both Wand V. Results. The above algorithm is demonstrated on a source separation task in Figure 2. We used 6 speech signals, transformed by non-linearities to have arbitrary one-point densities, and mixed by a random 8 x 6 matrix Ho. Different signalto-noise (SNR) levels were used. The error in the estimated H (left, solid line) is quantified by the size ofthe non-diagonal elements of (HTH)-l HTHo relative to the H Attias 392 diagonal; the results obtained by IFA [1], which does not use temporal information, are plotted for reference (dotted line). The mean squared error of the reconstructed sources (right, solid line) and the corresponding IFA result (right, dashed line) are also shown. The estimate and reconstruction errors of this algorithm are consistently smaller than those of IFA, reflecting the advantage of exploiting the temporal structure of the data. Additional experiments with different numbers of sources and sensors gave similar results. Notice that this algorithm, unlike the previous two, allows both L ::; L' and L > L'. We also considered situations where the number of sensors was smaller than the number of sources; the separation quality was good, although, as expected, less so than in the opposite case. 5 Conclusion An important issue that has not been addressed here is model selection. When applying our algorithms to an arbitrary dataset, the number of factors and of HMM states for each factor should be determined. Whereas this could be done, in principle, using cross-validation, the required computational effort would be fairly large. However, in a recent paper [12] we develop a new framework for Bayesian model selection, as well as model averaging, in probabilistic networks. This framework, termed Variational Bayes, proposes an EM-like algorithm which approximates full posterior distributions over not only hidden variables but also parameters and model structure, as well as predictive quantities, in an analytical manner. It is currently being applied to the algorithms presented here with good preliminary results. One field in which our approach may find important applications is speech technology, where it suggests building more economical signal models based on combining independent low-dimensional HMMs, rather than fitting a single complex HMM. It may also contribute toward improving recognition performance in noisy, multispeaker, reverberant conditions which characterize real-world auditory scenes. References [1] Attias, H. (1999). Independent factor analysis. Neur. Camp. 11, 803-85l. [2] Bell, A.J. & Sejnowski, T .J. (1995). An information-maximization approach to blind separation and blind deconvolution. Neur. Camp. 7, 1129-1159. [3] Amari, S., Cichocki, A. & Yang, H.H. (1996). A new learning algorithm for blind signal separation. Adv. Neur. Info. Pmc. Sys. 8,757-763 (Ed. by Touretzky, D.S. et al). MIT Press, Cambridge, MA. [4] Pearlmutter , B.A. & Parra, L.C. (1997). Maximum likelihood blind source separation: A context-sensitive generalization of ICA. Adv. Neur. Info. Pmc. Sys. 9, 613-619 (Ed. by Mozer, M.C. et al). MIT Press, Cambridge, MA. [5] Hyviirinen, A. & Oja, E. (1997). A fast fixed-point algorithm for independent component analysis. Neur. Camp. 9, 1483-1492. [6] Attias, H. & Schreiner, C.E. (1998). Blind source separation and deconvolution: the dynamic component analysis algorithm. Neur. Camp. 10, 1373-1424. [7] Rabiner, L. & Juang, B.-H. (1993). Fundamentals of Speech Recognition. Prentice Hall, Englewood Cliffs, NJ. [8] Lee, D.D. & Sompolinsky, H. (1999) , unpublished; D.D. Lee, personal communication. [9] Saul, L.K., Jaakkola, T., and Jordan, M.L (1996). Mean field theory of sigmoid belief networks. J. Art. Int. Res. 4, 61-76. [10] Ghahramani, Z. & Jordan, M.L (1997). Factorial hidden Markov models. Mach. Learn. 29, 245-273. [11] Neal, R.M. & Hinton, G.E. (1998). A view of the EM algorithm that justifies incremental, sparse, and other variants. Learning in Graphical Models, 355-368 (Ed. by Jordan, M.L). Kluwer Academic Press. [12] Attias, H. (2000). A variational Bayesian framework for graphical models. Adv. Neur. Info. Pmc. Sys. 12 (Ed. by Leen, T. et al). MIT Press, Cambridge, MA.
1682 |@word h:1 version:1 middle:1 stronger:1 advantageous:1 covariance:7 solid:2 recursively:1 reduction:5 initial:1 configuration:2 series:4 si:1 j1:1 mstep:1 update:3 v:2 isotropic:9 parametrization:1 ith:1 sys:3 provides:2 contribute:1 become:1 combine:1 fitting:2 manner:2 indeed:1 expected:2 ica:7 themselves:1 actual:1 underlying:1 moreover:1 notation:2 suffice:1 linearity:1 xtxt:1 z:1 unobserved:3 transformation:2 nj:1 temporal:16 demonstrates:1 uk:1 unit:1 modify:1 mach:1 cliff:1 might:1 quantified:1 suggests:1 hmms:1 limited:1 practice:1 block:5 differs:1 procedure:7 bell:2 adapting:1 word:1 pre:1 spite:1 selection:2 prentice:1 context:5 applying:2 restriction:1 optimize:1 map:2 demonstrated:1 yt:8 missing:1 baum:2 go:1 maximizing:2 starting:1 independently:1 l:2 welch:2 schreiner:1 estimator:1 rule:12 deriving:1 coordinate:1 updated:1 pt:5 zzt:1 exact:3 origin:1 element:1 recognition:2 labeled:1 logq:1 observed:7 capture:2 adv:3 sompolinsky:1 mozer:1 dynamic:9 personal:1 depend:1 solving:1 iyl:1 predictive:1 joint:1 derivation:1 fast:1 shortcoming:1 london:2 describe:1 sejnowski:2 quite:2 whose:1 amari:1 statistic:7 think:1 jointly:1 noisy:2 xlt:4 advantage:2 analytical:1 ucl:1 reconstruction:5 product:1 sll:1 combining:1 mixing:12 achieve:1 exploiting:1 convergence:1 wcin:1 requirement:1 juang:1 unmixing:2 produce:1 incremental:2 sna:1 derive:3 develop:1 ac:1 ij:4 eq:5 reco:1 implies:1 generalization:1 preliminary:1 parra:1 hagai:2 mathematically:1 pl:1 considered:1 hall:1 exp:2 inseparable:2 omitted:2 currently:1 vbp:1 sensitive:1 largest:1 successfully:1 tool:1 mit:3 sensor:3 gaussian:7 rather:1 factorizes:2 jaakkola:1 endow:1 derived:3 emission:1 consistently:1 likelihood:3 contrast:3 camp:4 dim:2 inference:1 sb:6 integrated:1 vl:1 bt:2 hidden:7 relation:1 ancestor:3 transformed:1 issue:1 classification:2 flexible:1 among:1 denoted:2 proposes:1 art:1 fairly:1 field:2 once:1 having:1 sampling:1 nearly:1 constitutes:1 oblivious:1 oja:1 recognize:1 individual:1 phase:3 attempt:1 englewood:1 mixture:5 permuting:1 accurate:1 capable:1 lh:1 re:2 plotted:1 modeling:2 ar:1 logp:3 queen:1 maximization:3 ordinary:2 introducing:1 parametrically:1 snr:2 successful:1 characterize:1 considerably:1 st:3 density:8 fundamental:1 probabilistic:6 yl:18 lee:2 invertible:4 squared:1 choose:1 leading:1 li:2 ioj:2 int:1 blind:9 bg:1 ad:1 tion:1 vi:1 depends:2 view:1 start:1 recover:1 bayes:1 complicated:1 minimize:1 square:9 l1j:1 wiener:1 variance:3 characteristic:2 efficiently:1 ensemble:1 correspond:1 ofthe:1 rabiner:1 bayesian:2 unifies:1 multiplying:1 economical:1 touretzky:1 ed:4 couple:1 auditory:1 dataset:1 knowledge:2 ut:4 dimensionality:1 fractional:1 adaptability:1 back:1 reflecting:1 focusing:1 ta:1 formulation:1 done:3 though:1 leen:1 stage:3 correlation:2 hand:2 defines:1 quality:3 facilitate:1 omitting:1 building:1 multiplier:1 true:4 hence:2 iteratively:1 neal:2 white:1 ll:1 multivariable:1 generalized:2 demonstrate:1 pearlmutter:1 multispeaker:1 variational:12 recently:1 superior:2 sigmoid:2 rl:1 jl:1 approximates:1 kluwer:1 significant:1 freeze:2 cambridge:3 imposing:1 moving:1 etc:1 multivariate:1 own:3 posterior:16 recent:1 scenario:1 termed:2 manifested:1 accomplished:1 additional:2 employed:1 maximize:1 dashed:1 semi:1 signal:11 ii:4 full:1 reduces:1 stem:1 academic:1 cross:1 variant:1 mog:4 expectation:1 iteration:1 normalization:2 achieved:1 lea:6 addition:1 whereas:3 addressed:1 source:61 crucial:2 appropriately:1 unlike:3 subject:1 facilitates:2 db:2 jordan:3 yang:1 reverberant:1 easy:1 fit:1 gave:1 perfectly:1 opposite:1 simplifies:1 idea:2 attias:7 expression:1 motivated:1 passed:1 forecasting:1 effort:1 suffer:1 rsi:1 speech:5 remark:1 cleaner:1 factorial:1 notice:6 dotted:1 estimated:1 affected:1 four:1 achieving:1 ht:3 backward:3 wand:1 sti:1 soi:1 place:1 family:1 separation:17 scaling:1 bound:5 completing:1 alogp:1 activity:1 adapted:1 orthogonality:3 constraint:2 scene:1 sphering:1 estep:1 structured:1 neur:7 describes:2 smaller:3 em:14 across:1 unity:2 making:1 s1:1 taken:1 computationally:3 equation:3 mutually:2 turn:2 tractable:3 lyt:1 generalizes:1 gaussians:2 apply:1 observe:1 indirectly:1 appropriate:1 alternative:1 ho:1 original:1 denotes:3 assumes:1 remaining:1 ensure:1 unfortunate:1 graphical:3 ifa:17 exploit:2 ghahramani:1 prof:1 approximating:1 added:1 quantity:4 fa:3 usual:1 diagonal:4 gradient:2 distance:1 separate:1 separating:1 hmm:11 parametrized:2 reason:1 toward:1 assuming:1 modeled:2 index:1 equivalently:2 pmc:3 info:3 unknown:1 perform:3 zf:1 markov:2 situation:1 hinton:2 communication:1 y1:1 arbitrary:5 introduced:3 unpublished:1 cleaned:1 kl:1 required:2 optimized:1 learned:4 suggested:1 usually:2 below:2 including:1 belief:2 natural:2 scheme:3 technology:1 temporally:2 coupled:1 cichocki:1 text:1 prior:1 relative:1 mixed:1 filtering:2 validation:1 principle:1 intractability:1 pi:1 course:1 repeat:1 last:1 free:6 aij:1 saul:1 sparse:1 axi:1 xn:1 transition:7 world:1 autoregressive:1 forward:3 made:1 coincide:1 sli:2 avoided:1 reconstructed:1 approximate:3 ml:1 summing:1 assumed:2 xi:3 factorize:1 hyviirinen:1 learn:4 correlated:3 hth:4 improving:1 complex:1 linearly:1 noise:29 arise:1 x1:2 tl:2 gatsby:2 sub:1 xl:10 lie:1 vanish:1 learns:1 bij:1 tiip:1 xt:15 deconvolution:2 intractable:5 hxt:2 effectively:1 diagonalization:1 conditioned:2 justifies:1 signalto:1 lt:6 lagrange:1 datadependent:1 ma:3 conditional:1 determined:3 averaging:4 principal:2 e:1 ya:1 college:1 latter:2 ex:2
747
1,683
Recognizing Evoked Potentials in a Virtual Environment * Jessica D. Bayliss and Dana H. Ballard Department of Computer Science University of Rochester Rochester, NY 14627 {bayliss,dana}@cs.rochester.edu Abstract Virtual reality (VR) provides immersive and controllable experimental environments. It expands the bounds of possible evoked potential (EP) experiments by providing complex, dynamic environments in order to study cognition without sacrificing environmental control. VR also serves as a safe dynamic testbed for brain-computer .interface (BCl) research. However, there has been some concern about detecting EP signals in a complex VR environment. This paper shows that EPs exist at red, green, and yellow stop lights in a virtual driving environment. Experimental results show the existence of the P3 EP at "go" and "stop" lights and the contingent negative variation (CNY) EP at "slow down" lights. In order to test the feasibility of on-line recognition in VR, we looked at recognizing the P3 EP at red stop tights and the absence of this signal at yellow slow down lights. Recognition results show that the P3 may successfully be used to control the brakes of a VR car at stop lights. 1 Introduction The controllability of VR makes it an excellent candidate for use in studying cognition. It expands the bounds of possible evoked potential (EP) experiments by providing complex, dynamic environments in order to study decision making in cognition without sacrificing environmental control. We have created a flexible system for real-time EEG collection and analysis from within virtual environments. The ability of our system to give quick feedback enables it to be used in brain-computer interface (BCl) research, which is aimed at helping individuals with severe motor deficits to become more independent. Recent BCl work has shown the feasibility of on-line averaging and biofeedback methods in order to choose characters or move a cursor on a computer screen with up to 95% accuracy while sitting still and concentrating on the screen [McFarland et aI., 1993; Pfurtscheller et al., 1996; Vaughn et al., 1996; Farwell and Donchin, 1988]. Our focus is to dramatically extend the BCl by allowing evoked potentials to propel the user through alternate virtual environments. For example, a *This research was supported by NIHIPHS grantl-P41-RR09283. It was also facilitated in part by a National Physical Science Consortium Fellowship and by stipend support from NASA Goddard Space Flight Center. 4 J. D. Bayliss and D. H. Ballard Figure 1: (Left) An individual demonstrates driving in the modified go cart. (Right) A typical stoplight scene in the virtual environment. user could choose a virtual living room from a menu of rooms, navigate to the living room automatically in the head-mounted display, and then choose to turn on the stereo. As shown in [Farwell and Donchin, 1988], the P3 EP may be used for a brain-computer interface that picks characters on a computer monitor. Discovered by [Chapman and Bragdon, 1964; Sutton et aI., 1965] and extensively studied (see [Polich, 1998] for a literature review), the P3 is a positive waveform occurring approximately 300-500 ms after an infrequent task-relevant stimulus. We show that requiring subjects to stop or go at virtual traffic lights elicits this EP. The contingent negative variation (CNV), an EP that happens preceding an expected stimulus, occurs at slow down lights. In order to test the feasibility of on-line recognition in the noisy VR environment, we recognized the P3 EP at red stop lights and the lack of this signal at yellow slow down lights. Results using a robust Kalman filter for off-line recognition indicate that the car may be stopped reliably with an average accuracy of 84.5% while the on-line average for car halting is 83%. 2 The Stoplight Experiments The first experiment we performed in the virtual driving environment shows that a P3 EP is obtained when subjects stop or go at a virtual light and that a CNV occurs when subjects see a slow down light. Since all subjects received the same light colors for the slow down, go, and stop conditions we then performed a second experiment with different light colors in order to disambiguate light color from the occurrence of the P3 and CNV. Previous P3 research has concentrated primarily on static environments such as the continuous performance task [Rosvold et aI., 1956]. In the visual continuous performance task (VCPT), static images are flashed on a screen and the subject is told to press a button when a rare stimulus occurs or to count the number of occurrences of a rare stimulus. This makes the stimulus both rare and task relevant in order to evoke a P3. As an example, given red and yellow stoplight pictures, a P3 should occur if the red picture is less frequent than the yellow and subjects are told to press a mouse button only during the red light. We assumed a similar response would occur in a VR driving world if certain lights were infrequent and subjects were told to stop or go at them. This differs from the VCPT in two important ways: 1. In the VCPT subjects sit passively and respond to stimuli. In the driving task, Recognizing Evoked Potentials in a Virtual Environment 5 subjects control when the stimuli appear by where they drive in the virtual world. 2. Since subjects are actively involved and fully immersed in the virtual world, they make more eye and head movements. The movement amount can be reduced by a particular experimental paradigm, but it can not be eliminated. The first difference makes the VR environment a more natural experimental environment. The second difference means that subjects create more data artifacts with extra movement. We handled these artifacts by first manipulating the experimental environment to reduce movements where important stimulus events occurred. This meant that all stoplights were placed at the end of straight stretches of road in order to avoid the artifacts caused by turning a corner. For our on-line recognition, we then used the eye movement reduction technique described in [Semlitsch et al., 1986] in order to subtract a combination of the remaining eye and head movement artifact. 2.1 Experimental Setup All subjects used a modified go cart in order to control the virtual car (see Figure 1). The virtual reality interface is rendered on a Silicon Graphics Onyx machine with 4 processors and an Infinite Reality Graphics Engine. The environment is presented to the subject through a head-mounted display (HMD). Since scalp EEG recordings are measured in microvolts, electrical signals may easily interfere during an experiment. We tested the effects of wearing a VR4 HMD containing an ISCAN eye tracker and discovered that the noise levels inside of the VR helmet were comparable to noise levels while watching a laptop screen [Bayliss and Ballard, 1998]. A trigger pulse containing information about the color of the light was sent to the EEG acquisition system whenever a light changed. While an epoch size from -100 ms to 1 sec was specified, the data was recorded continuously. Information about head position as well as gas, braking, and steering position were saved to an external file. Eight electrodes sites (FZ, CZ, CPZ, PZ, P3, P4, as well as 2 vertical EOG channels) were arranged on the heads of seven subjects with a linked mastoid reference. Electrode impedances were between 2 and 5 kohms for all subjects. Subjects ranged in age from 19 to 52 and most had no previous experiences in a virtual environment. The EEG signal was amplified using Grass amplifiers with an analog bandwidth from 0.1 to 100 Hz. Signals were then digitized at a rate of 500 Hz and stored to a computer. 2.2 Ordinary Traffic Light Color Experiment Five subjects were instructed to slow down on yellow lights, stop for red lights, and go for green lights. These are normal traffic light colors. Subjects were allowed to drive in the environment before the experiment to get used to driving in VR. In order to make slow down lights more frequent, all stoplights turned to the slow down color when subjects were further than 30 meters aways from them. When the subject drove closer than 30 meters the light then turned to either the go or stop color with equal probability. The rest of the light sequence followed normal stoplights with the stop light turning to the go light after 3 seconds and the go light not changing. We calculated the grand averages over red, green, and yellow light trials (see Figure 2a). Epochs affected by artifact were ignored in the averages in order to make sure that any existing movements were not causing a P3-like signal. Results show that a P3 EP occurs for both red and green lights. Back averaging from the green/red lights to the yellow light shows the existence of a CNV starting at approximately 2 seconds before the light changes to red or green. 6 J. D. Bayliss and D. H Ballard Stop Light ~ Slow Down Light Go Light -5 uv 1\ .~'"\ i " \ I I ! \ I \ I i: \\ .v'l,,/AI \, I I) +lOuv -8 uv .E bO ;J u ~ b) ~ t) > \"~'~ ~ ; ,: , '::1 j '-lOOms :"' ,~t r/'" " .~ '\ f I".j" "\?.....\ ".; t<iooms' '-lOOms A, f...j Vv f I lOOOms I h'-3~000ii1S;;:::::::::==::;;2;;OO~ms~1 +12 uv Figure 2: a) Grand averages for the red stop, green go, and yellow slow down lights. b) Grand averages for the yellow stop, red go, and green slow down lights. All slow down lights have been back-averaged from the occurrence of the go/stop light in order to show the existence of a CNY. 2.3 Alternative Traffic Light Colors The P3 is related to task relevance and should not be related to color, but color needed to be disambiguated as the source of the P3 in the experiment. We had two subjects slow down at green lights, stop at yellow lights, and go at red lights. In order to get used to this combination of colors, subjects were allowed to drive in the town before the experiment. The grand averages for each light color were calculated in the same manner as the averages above and are shown in Figure 2b. As expected, a P3 signal existed for the stop condition and a CNV for the slow down condition. The go condition P3 was much noisier for these two subjects, although a slight P3-like signal is still visible. 3 Single Trial Recognition Results While averages show the existence of the P3 EP at red stop lights and the absence of such at yellow slow down lights, we needed to discover if the signal was clean enough for single trial recognition as the quick feedback needed by a BCI depends on quick recognition. While there were three light conditions to recognize, there were only two distinct kinds of evoked potentials. We chose to recognize the difference between the P3 and the CNV since their averages are very different. Recognizing the difference between two kinds of EPs gives us the ability to use a BCI in any task that can be performed using a series of binary decisions. We tried three methods for classification of the P3 EP: correlation, independent component analysis (ICA), and a robust Kalman filter. Approximately, 90 slow down yellow light and 45 stop red light trials from each subject were classified. The reason we allowed a yellow light bias to enter recognition is because the yellow light currently represents an unimportant event in the environment. In a real BCI unimportant events are likely to occur more than user-directed actions, making this bias justifiable. Recognizing Evoked Potentials in a Virtual Environment 7 Table 1: Recognition Results (p Subjects S1 S2 S3 S4 S5 Correlation %Correct Red Yel Total 81 51 64 95 73 63 89 56 66 81 60 67 63 66 65 ICA Red 76 86 72 73 65 < 0.01) %Correct Yel Total 77 77 87 88 82 87 71 69 74 79 Robust Kalman Filter %Correct Total Red Yel 55 86 77 82 94 90 74 81 85 91 82 65 78 92 87 Table 2: Recognition Results for Return Subjects Subjects S4 S5 Robust K-Filter % Correct Total Red Yel 73 90 85 67 87 80 As expected, the data obtained while driving contained artifacts, but in an on-line BCI these artifacts must be reduced in order to make sure that what the recognition algorithm is recognizing is not an artifact such as eye movement. In order to reduce these artifacts, we performed the on-line linear regression technique described in [Semlitsch et aI. , 1986] in order to subtract a combination of eye and head movement artifact. In order to create a baseline from which to compare the performance of other algorithms, we calculated the correlation of all sample trials with the red and yellow light averages from each subject's maximal P3 electrode site using the following formula: correlation = (sample * aveT)/(11 sample II * II ave II) (1) where sample and ave are both 1 x 500 vectors representing the trial epochs and light averages (respectively). We used the whole trial epoch for recognition because it yielded better recognition than just the time area around the P3. If the highest correlation of a trial epoch with the red and yellow averages was greater than 0.0, then the signal was classified as that type of signal. If both averages correlated negatively with the single trial, then the trial was counted as a yellow light signal. As can be seen in Table 1, the correct signal identification of red lights was extremely high while the yellow light identification pulled the results down. This may be explained by the greater variance of the yellow light epochs. Correlations in general were poor with typical correlations around 0.25. ICA has successfully been used in order to minimize artifacts in EEG data [Jung et at. , 1997; Vigario, 1997] and has also proven useful in separating P3 component data from an averaged waveform [Makeig et aI., 1997]. The next experiment used ICA in order to try to separate the background EEG signal from the P3 signal. Independent component analysis (lCA) assumes that n EEG data channels x are a linear combination of n statistically independent signals s : x= As (2) where x and s are n x 1 vectors. We used the matlab package mentioned in [Makeig et aI. , 1997] with default learning values, which finds a matrix W by stochastic gradient descent. 8 J D. Bayliss and D. H. Ballard This matrix W performs component separation. All data was sphered in order to speed convergence time. After training the W matrix, the source channel showing the closest P3-like signal (using correlation with the average) for the red light average data was chosen as the signal with which to correlate individual epochs. The trained W matrix was also used to find the sources of the yellow light average. The red and yellow light responses were then correlated with individual epoch sources in the manner of the first experiment. The third experiment used the robust Kalman filter framework formulated by Rao [Rao, 1998]. The Kalman filter assumes a linear model similar to the one ofICA in equation 2, but assumes the EEG output x is the observable output of a generative or measurement matrix A and an internal state vector s of Gaussian sources. The output may also have an additional noise component n, a Gaussian stochastic noise process with mean zero and a covariance matrix given by ~ = E[nnTj, leading to the model expression: x = As + n. In order to find the most optimal value of s, a weighted least-squares criterion is formulated: (3) where s follows a Gaussian distribution with mean s and covariance M. Minimizing this criterion by setting ~; = 0 and using the substitution N = (AT~-lU + M-1)-1 yields the Kalman filter equation, which is basically equal to the old estimate plus the Kalman gain times the residual error. (4) In an analogous manner, the measurement matrix A may be estimated (learned) if one assumes the physical relationships encoded by the measurement matrix are relatively stable. The learning rule for the measurement matrix may be derived in a manner similar to the rule for the internal state vector. In addition, a decay term is often needed in order to avoid overfitting the data set. See [Rao, 1998] for details. In our experiments both the internal state matrix s and the measurement matrix A were learned by training them on the average red light signal and the average yellow light signal. The signal is measured from the start of the trial which is known since it is triggered by the light change. We used a Kalman gain of 0.6 and a decay of 0.3. After training, the signal estimate for each epoch is correlated with the red and yellow light signal estimates in the manner of experiment 1. We made the Kalman filter statistically robust by ignoring parts of the EEG signal that fell outside a standard deviation of 1.0 from the training signals. The overall recognition results in Table 1 suggest that both the robust Kalman filter and ICA have a statistically significant advantage over correlation (p < 0.01). The robust Kalman filter has a very small advantage over ICA (not statistically significant). In order to look at the reliability of the best algorithm and its ability to be used on-line two of the Subjects (S4 and SS) returned for another VR driving session. In these sessions the brakes of the driving simulator were controlled by the robust Kalman filter recognition algorithm for red stop and yellow slow down lights. Green lights were ignored. The results of this session using the Robust Kalman Filter trained on the first session are shown in Table 2. The recognition numbers for red and yellow lights between the two sessions were compared using correlation. Red light scores between the sessions correlated fairly highly - 0.82 for S4 and 0.69 for SS. The yellow light scores between sessions correlated poorly with both S4 and SS at approximately -0.1. This indicates that the yellow light epochs tend to correlate poorly with each other due to the lack of a large component such as the P3 to tie them together. Recognizing Evoked Potentials in a Virtual Environment 9 4 Future Work This paper showed the viability of recognizing the P3 EP in a VR environment. We plan to allow the P3 EP to propel the user through alternate virtual rooms through the use of various binary decisions. In order to improve recognition for the BCI we need to experiment with a wider and more complex variety of recognition algorithms. Our most recent work has shown a dependence between the human computer interface used in the BCI and recognition. We would like to explore this dependence in order to improve recognition as much as possible. References [Bayliss and Ballard, 1998) lD. Bayliss and D.H. Ballard, ''The Effects of Eye Tracking in a VR Helmet on EEG Recording," TR 685, University of Rochester National Resource Laboratory for the Study of Brain and Behavior, May 1998. [Chapman and Bragdon, 1964) R.M. Chapman and H.R. Bragdon, "Evoked responses to numerical and non-numerical visual stimuli while problem solving.," Nature, 203: 1155-1157, 1964. [Farwell and Donchin, 1988) L. A. Farwell and E. Donchin, "Talking off the top of your head: toward a mental prosthesis utilizing event-related brain potentials," Electroenceph. Clin. Neurophysiol., pages 510-523, 1988. [Jung et al., 1997) 1'.P. Jung, C. Humphries,1'. Lee, S. Makeig, M.J. McKeown, Y. lragui, and 1'.l Sejnowski, "Extended ICA Removes Artifacts from Electroencephalographic Recordings," to Appear in Advances in Neural Information Processing Systems, 10, 1997. [Makeig et al., 1997) S. Makeig, 1'. Jung, A.J. Bell, D. Ghahremani, and 1'.J. Sejnowski, "Blind Separation of Auditory Event-related Brain Responses into Independent Components," Proc. Nat'l Acad. Sci. USA , 94:10979-10984, 1997. [McFarland et al., 1993) D.l McFarland, G.w. Neat, R.F. Read, and J.R. Wolpaw, "An EEG-based method for graded cursor control," Psychobiology, 21(1):77-81, 1993. [Pfurtscheller et al. , 1996) G. Pfurtscheller, D. Flotzinger, M. Pregenzer, J. Wolpaw, and D. McFarland, "EEG-based Brain Computer Interface (BCI)," Medical Progress through Technology, 21:111-121 , 1996. [Polich, 1998] J. Polich, "P300 Clinical Utility and Control of Variability," J. of Clinical NeurophYSiology, 15(1): 14-33, 1998. [Rao, 1998] R. P.N. Rao, "Visual Attention during Recognition," Advances in Neural Information Processing Systems, 10, 1998. [Rosvold et al., 1956] H.E. Rosvold, A.F. Mirsky, I. Sarason, E.D. Bransome Jr., and L.H. Beck, "A Continuous Performance Test of Brain Damage," 1. Consult. Psychol., 20, 1956. [SemIitschetal., 1986) H.Y. SemIitsch, P. Anderer, P Schuster, and O. Presslich, "A solution for reliable and valid reduction of ocular artifacts applied to the P300 ERP;' Psychophys., 23 :695703,1986. [Sutton et al., 1965) S. Sutton, M. Braren, J. Zublin, and E. John, "Evoked potential correlates of stimulus uncertainty," Science, 150: 1187-1188, 1965. [Vaughn et al., 1996) 1'.M. Vaughn, J.R. Wolpaw, and E. Donchin, "EEG-Based Communication: Prospects and Problems," IEEE Trans. on Rehabilitation Engineering, 4(4):425-430, 1996. [Vigario, 1997) R. Vigario, "Extraction of ocular artifacts from eeg using independent component analysis," Electroenceph. Clin. Neurophysiol., 103:395-404, 1997.
1683 |@word neurophysiology:1 trial:11 pulse:1 tried:1 covariance:2 pick:1 tr:1 ld:1 reduction:2 substitution:1 series:1 score:2 existing:1 must:1 john:1 numerical:2 visible:1 enables:1 motor:1 remove:1 iscan:1 grass:1 generative:1 mental:1 provides:1 detecting:1 five:1 become:1 inside:1 manner:5 ica:7 expected:3 behavior:1 simulator:1 brain:8 automatically:1 ofica:1 discover:1 pregenzer:1 laptop:1 what:1 kind:2 expands:2 tie:1 makeig:5 demonstrates:1 control:7 medical:1 appear:2 positive:1 before:3 engineering:1 acad:1 sutton:3 approximately:4 chose:1 plus:1 studied:1 mirsky:1 evoked:10 statistically:4 averaged:2 directed:1 differs:1 wolpaw:3 area:1 bell:1 road:1 suggest:1 consortium:1 get:2 flotzinger:1 humphries:1 quick:3 center:1 brake:2 go:17 attention:1 starting:1 rule:2 utilizing:1 menu:1 variation:2 analogous:1 drove:1 trigger:1 infrequent:2 user:4 recognition:22 ep:16 electrical:1 movement:9 highest:1 prospect:1 mentioned:1 environment:23 dynamic:3 trained:2 solving:1 negatively:1 neurophysiol:2 easily:1 various:1 distinct:1 sejnowski:2 outside:1 encoded:1 s:3 bci:7 ability:3 noisy:1 sequence:1 triggered:1 advantage:2 maximal:1 p4:1 frequent:2 causing:1 relevant:2 turned:2 p300:2 poorly:2 amplified:1 convergence:1 electrode:3 mckeown:1 vaughn:3 oo:1 wider:1 measured:2 received:1 progress:1 c:1 polich:3 indicate:1 safe:1 waveform:2 saved:1 correct:5 filter:12 stochastic:2 human:1 virtual:19 stipend:1 cnv:6 helping:1 stretch:1 tracker:1 around:2 normal:2 cognition:3 driving:9 proc:1 currently:1 helmet:2 create:2 successfully:2 weighted:1 gaussian:3 modified:2 avoid:2 derived:1 focus:1 indicates:1 electroencephalographic:1 ave:2 baseline:1 manipulating:1 overall:1 classification:1 flexible:1 plan:1 fairly:1 equal:2 extraction:1 eliminated:1 chapman:3 represents:1 look:1 future:1 stimulus:10 primarily:1 national:2 recognize:2 individual:4 beck:1 jessica:1 amplifier:1 highly:1 propel:2 severe:1 light:72 closer:1 experience:1 old:1 farwell:4 biofeedback:1 prosthesis:1 sacrificing:2 stopped:1 rao:5 ordinary:1 deviation:1 rare:3 recognizing:8 graphic:2 stored:1 grand:4 told:3 off:2 microvolt:1 lee:1 together:1 mouse:1 continuously:1 recorded:1 town:1 containing:2 choose:3 watching:1 corner:1 external:1 leading:1 return:1 actively:1 potential:10 halting:1 sec:1 caused:1 depends:1 blind:1 performed:4 try:1 linked:1 traffic:4 red:30 mastoid:1 start:1 rochester:4 minimize:1 square:1 accuracy:2 variance:1 sitting:1 yield:1 yellow:28 identification:2 basically:1 lu:1 psychobiology:1 drive:3 justifiable:1 straight:1 processor:1 classified:2 whenever:1 acquisition:1 involved:1 ocular:2 static:2 stop:21 gain:2 auditory:1 concentrating:1 color:13 car:4 back:2 nasa:1 response:4 arranged:1 just:1 correlation:10 flight:1 lack:2 interfere:1 artifact:14 usa:1 effect:2 requiring:1 ranged:1 read:1 laboratory:1 flashed:1 during:3 m:3 criterion:2 performs:1 interface:6 image:1 physical:2 cpz:1 extend:1 occurred:1 analog:1 braking:1 slight:1 eps:2 silicon:1 s5:2 measurement:5 significant:2 ai:7 enter:1 uv:3 session:7 had:2 reliability:1 stable:1 closest:1 recent:2 showed:1 certain:1 binary:2 seen:1 contingent:2 greater:2 preceding:1 steering:1 additional:1 recognized:1 paradigm:1 signal:26 living:2 ii:3 clinical:2 feasibility:3 controlled:1 regression:1 vigario:3 cz:1 background:1 fellowship:1 addition:1 source:5 extra:1 rest:1 file:1 sure:2 cart:2 subject:29 recording:3 sent:1 hz:2 fell:1 tend:1 consult:1 enough:1 viability:1 variety:1 bandwidth:1 reduce:2 expression:1 handled:1 utility:1 stereo:1 returned:1 action:1 matlab:1 dramatically:1 ignored:2 useful:1 aimed:1 unimportant:2 amount:1 s4:5 extensively:1 concentrated:1 reduced:2 fz:1 exist:1 s3:1 estimated:1 affected:1 monitor:1 changing:1 erp:1 clean:1 button:2 immersed:1 facilitated:1 yel:4 package:1 respond:1 uncertainty:1 p3:30 separation:2 decision:3 comparable:1 bound:2 followed:1 display:2 existed:1 yielded:1 scalp:1 occur:3 your:1 scene:1 speed:1 disambiguated:1 extremely:1 passively:1 rendered:1 relatively:1 department:1 lca:1 alternate:2 combination:4 poor:1 jr:1 aways:1 character:2 making:2 happens:1 s1:1 rehabilitation:1 explained:1 equation:2 resource:1 turn:1 count:1 needed:4 serf:1 end:1 studying:1 eight:1 rr09283:1 occurrence:3 alternative:1 existence:4 assumes:4 remaining:1 top:1 clin:2 goddard:1 graded:1 move:1 looked:1 occurs:4 damage:1 dependence:2 gradient:1 deficit:1 elicits:1 separating:1 separate:1 sci:1 seven:1 reason:1 toward:1 kalman:13 relationship:1 providing:2 minimizing:1 ghahremani:1 setup:1 negative:2 reliably:1 allowing:1 vertical:1 descent:1 controllability:1 gas:1 extended:1 variability:1 head:8 digitized:1 communication:1 discovered:2 specified:1 engine:1 learned:2 testbed:1 trans:1 mcfarland:4 psychophys:1 green:10 reliable:1 event:5 natural:1 turning:2 residual:1 representing:1 loom:2 improve:2 technology:1 eye:7 picture:2 created:1 psychol:1 eog:1 review:1 literature:1 epoch:10 meter:2 bcl:4 fully:1 mounted:2 proven:1 dana:2 age:1 changed:1 jung:4 supported:1 placed:1 neat:1 bias:2 allow:1 vv:1 pulled:1 feedback:2 calculated:3 default:1 world:3 valid:1 electroenceph:2 instructed:1 collection:1 made:1 counted:1 correlate:3 observable:1 evoke:1 overfitting:1 assumed:1 continuous:3 table:5 reality:3 disambiguate:1 nature:1 impedance:1 ballard:7 channel:3 robust:10 controllable:1 ignoring:1 eeg:14 excellent:1 complex:4 s2:1 noise:4 whole:1 allowed:3 site:2 bayliss:8 screen:4 ny:1 vr:14 slow:18 pfurtscheller:3 position:2 candidate:1 third:1 down:19 formula:1 navigate:1 showing:1 pz:1 decay:2 concern:1 sit:1 hmd:2 donchin:5 p41:1 nat:1 occurring:1 cursor:2 subtract:2 likely:1 explore:1 visual:3 contained:1 tracking:1 bo:1 talking:1 immersive:1 environmental:2 formulated:2 room:4 absence:2 change:2 typical:2 infinite:1 averaging:2 total:4 experimental:6 internal:3 support:1 sphered:1 meant:1 noisier:1 relevance:1 wearing:1 tested:1 schuster:1 correlated:5
748
1,685
Bayesian Network Induction via Local Neighborhoods Dimitris Margaritis Department of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 D.Margaritis@cs.cmu.edu Sebastian Thrun Department of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 S. Thrun@cs.cmu.edu Abstract In recent years, Bayesian networks have become highly successful tool for diagnosis, analysis, and decision making in real-world domains. We present an efficient algorithm for learning Bayes networks from data. Our approach constructs Bayesian networks by first identifying each node's Markov blankets, then connecting nodes in a maximally consistent way. In contrast to the majority of work, which typically uses hill-climbing approaches that may produce dense and causally incorrect nets, our approach yields much more compact causal networks by heeding independencies in the data. Compact causal networks facilitate fast inference and are also easier to understand. We prove that under mild assumptions, our approach requires time polynomial in the size of the data and the number of nodes. A randomized variant, also presented here, yields comparable results at much higher speeds. 1 Introduction A great number of scientific fields today benefit from being able to automatically estimate the probability of certain quantities of interest that may be difficult or expensive to observe directly. For example, a doctor may be interested in estimating the probability of heart disease from indications of high blood pressure and other directly measurable quantities. A computer vision system may benefit from a probability distribution of buildings based on indicators of horizontal and vertical straight lines. Probability densities proliferate the sciences today and advances in its estimation are likely to have a wide impact on many different fields . Bayesian networks are a succinct and efficient way to represent a joint probability distribution among a set of variables. As such, they have been applied to fields such as those mentioned [Herskovits90][Agosta88]. Besides their ability for density estimation, their semantics lend them to what is sometimes loosely referred to as causal discovery, namely directional relationships among quantities involved. It has been widely accepted that the most parsimonious representation for a Bayesian net is one that closely represents the causal independence relationships that may exist. For these reasons, there has been great interest in automatically inducing the structure of Bayesian nets automatically from data, preferably also preserving the independence relationships in the process. Two research approaches have emerged. The first employs independence properties of the underlying network that produced the data in order to discover parts of its structure. This approach is mainly exemplified by the SGS and PC algorithms in [Spirtes93], as well 506 D. Margaritis and S. Thrun Figure 1: On the left, an example of a Markov blanket of variable X is shown. The members of the blanket are shown shaded. On the right, an example reconstruction of a 5 x 5 rectangular net of branching factor 3 by the algorithm presented in this paper using 20000 samples. Indicated by dotted lines are 3 directionality errors. as for restricted classes such as trees [Chow68] and poly trees [Rebane87]. The second approach is concerned more with data prediction, disregarding independencies in the data. It is typically identified with a greedy hill-climbing or best-first beam search in the space of legal structures, employing as a scoring function a form of data likelihood, sometimes penalized for network complexity. The result is a local maximum score network structure for representing the data, and is one of the more popular techniques used today. This paper presents an approach that belongs in the first category. It addresses the two main shortcomings of the prior work which, we believe, are preventing its use from becoming more widespread. These two disadvantages are: exponential execution times, and proneness to errors in dependence tests used. The former problem is addressed in this paper in two ways. One is by identifying the local neighborhood of each variable in the Bayesian net as a preprocessing step, in order to facilitate the recovery of the local structure around each variable in polynomial time under the assumption of bounded neighborhood size. The second, randomized version goes one step further, employing a user-specified number of randomized tests (constant or logarithmic) in order to ascertain the same result with high probability. The second disadvantage of this research approach, namely proneness to errors, is also addressed by the randomized version, by using multiple data sets (if available) and Bayesian accumulation of evidence. 2 The Grow-Shrink Markov Blanket Algorithm The concept of the Markov blanket of a variable or a set of variables is central to this paper. The concept itself is not new. For example, see [PearI88]. It is surprising, however, how little attention it has attracted for all its being a fundamental property of a Bayesian net. What is new in this paper is the introduction of the explicit use of this idea to effectively limit unnecessary computation, as well as a simple algorithm to compute it. The definition of a Markov blanket is as follows: denoting V as the set of variables and X HS Y as the conditional dependence of X and Y given the set S, the Markov blanket BL(X) ~ V of X E V is any set of variables such that for any Y E V - BL(X) - {X}, X ft-BL(x) Y. In other words, BL(X) completely shields variable X from any other variable in V . The notion of a minimal Markov blanket, called a Markov boundary, is also introduced in [PearI88] and its uniqueness shown under certain conditions. The Markov boundary is not unique in certain pathological situations, such as the equality of two variables. In our following discussion we will assume that the conditions necessary for its existence and uniqueness are satisfied and we will identify the Markov blanket with the Markov boundary, using the notation B (X) for the blanket of variable X from now on. It is also illuminating to mention that, in the Bayesian net framework, the Markov blanket of a node X is easily identifiable from the graph: it consists of all parents, children and parents of children of X. An example Markov blanket is shown in Fig. 1. Note that any of these nodes, say Y, is dependent with X given B (X) - {Y}. 507 Bayesian Network Induction via Local Neighborhoods 1. S t- 0. 2. While:3 Y E V - {X} such that Y HS X, do S t- S U {Y}. 3. While:3 YES such that Y ft-S-{Y} X, do S t- S - {Y}. [Growing phase] [Shrinking phase] 4. B(X) t- S. Figure 2: The basic Markov blanket algorithm. The algorithm for the recovery of the Markov blanket of X is shown in Fig. 2. The idea behind step 2 is simple: as long as the Markov blanket property of X is violated (ie. there exists a variable in V that is dependent on X), we add it to the current set S until there are no more such variables. In this process however, there may be some variables that were added to S that were really outside the blanket. Such variables would have been rendered independent from X at a later point when "intervening" nodes of the underlying Bayesian net were added to S. This observation necessitates step 3, which identifies and removes those variables. The algorithm is efficient, requiring only O( n) conditional tests, making its running time O(n IDI), where n = IVI and D is the set of examples. For a detailed derivation of this bound as well as a formal proof of correctness, see [Margaritis99]. In practice one may try to minimize the number of tests in step 3 by heuristically ordering the variables in the loop of step 2, for example by ascending mutual information or probability of dependence between X and Y (as computed using the X2 test, see section 5). 3 Grow-Shrink (GS) Algorithm for Bayesian Net Induction The recovery of the local structure around each node is greatly facilitated by the knowledge of the nodes' Markov blankets. What would normally be a daunting task of employing dependence tests conditioned on an exponential number of subsets of large sets of variables-even though most of their members may be irrelevant-can now be focused on the Markov blankets of the nodes involved, making structure discovery much faster and more reliable. We present below the plain version of the GS algorithm that utilizes blanket information for inducing the structure of a Bayesian net. At a later point of this paper, we will present a robust, randomized version that has the potential of being faster and more reliable, as well as being able to operate in an "anytime" manner. In the following N (X) represents the direct neighbors of X. [ Compute Markov Blankets ] For all X E V, compute the Markov blanket B (X) . [ Compute Graph Structure] For all X E V and Y E B(X), determine Y to be a direct neighbor of X if X and Y are dependent given S for all S ~ T, where T is the smaller of B (X) - {Y} and B(Y) - {X}. [Orient Edges] For all X E V and YEN (X), orient Y -+ X if there exists a variable Z E N (X) - N (Y) - {Y} such that Y and Z are dependent given S U {X} for all S ~ U, where U is the smaller of B (Y) - {Z} and B (Z) - {Y}. [ Remove Cycles] Do the following while there exist cycles in the graph: 1. Compute the set of edges C = {X -+ Y such that X -+ Y is part of a cycle}. 2. Remove the edge in C that is part of the greatest number of cycles, and put it in R. 508 D. Margaritis and S. Thrun [ Reverse Edges] Insert each edge from R in the graph, reversed. [ Propagate Directions] For all X E V and Y E N(X) such that neither Y ~ X nor X ~ Y, execute the following rule until it no longer applies: If there exists a directed path from X to Y, orient X ~ Y . In the algorithm description above, step 2 determines which of the members of the blanket of each node are actually direct neighbors (parents and children). Assuming, without loss of generality, that B (X) - {Y} is the smaller set, if any of the tests are successful in separating (making independent) X from Y, the algorithm determines that there is no direct connection between them. That would happen when the conditioning set S includes all parents of X and no common children of X and Y. It is interesting to note that the motivation behind selecting the smaller set to condition on stems not only from computational efficiency but from reliability as well: a conditioning set S causes the data set to be split into 21 S 1 partitions; smaller conditioning sets cause the data set to be split into larger partitions and make dependence tests more reliable. Step 3 exploits the fact that two variables that have a common descendant become dependent when conditioning on a set that includes any such descendant. Since the direct neighbors of X and Y are known from step 2, we can determine whether a direct neighbor Y is a parent of X if there exists another node Z (which, coincidentally, is also a parent) such that any attempt to separate Y and Z by conditioning on a subset of the blanket of Y that includes X, fails (assuming that B(Y) is smaller than B(Z)). If the directionality is indeed Y ~ X ~ Z, there should be no such subset since, by conditioning on X, a permanent dependency path between Y and Z is created. This would not be the case if Y were a child of X. It is straightforward to show that the algorithm requires 0 (n 2 + nb2 2 b ) conditional independence tests, where b maxx(IB(X)I). Under the assumption that b is bounded by a constant, this algorithm is O( n 2 ) in the number of conditional independence tests. It is worthwhile to note that the time to compute a conditional independence test by a pass over the data set Dis O( n IDt) and not O(2IVI). An analysis and a formal proof of correctness of the algorithm is presented in [Margaritis99]. = Discussion The main advantage of the algorithm comes through the use of Markov blankets to restrict the size of the conditioning sets. The Markov blankets may be usually wrong in the side of including too many nodes because they are represented by a disjunction of tests for all values of the conditioning set, on the same data. This emphasizes the importance of the "direct neighbors" step which removes nodes that were incorrectly added during the Markov blanket computation step by admitting variables whose dependence was shown high confidence in a large number of different tests. It is also possible that an edge direction is wrongly determined during step 3 due to nonrepresentative or noisy data. This may lead to directed cycles in the resulting graph. It is therefore necessary to remove those cycles by identifying the minimum set of edges than need to be reversed for all cycles to disappear. This problem is closely related [Margaritis99] to the Minimum Feedback Arc Set problem, which is concerned with identifying a minimum set of edges that need to be removed from a graph that possibly contains directed cycles, in order for all such cycles to disappear. Unfortunately, this problem is NP-complete in its generality [Junger85]. We introduce here a reasonable heuristic for its solution that is based on the number of cycles that an edge that is part of a cycle is involved in. Not all edge directions can be determined during the last two steps. For example, nodes with a single parent or multi-parent nodes (called colliders) whose parents are directly connected do not apply to step 3, and steps 4 and 5 are only concerned with already directed edges. Step 6 attempts to ameliorate that, through orienting edges in a way that does not introduce 509 Bayesian Network Induction via Local Neighborhoods a cycle, if the reverse direction necessarily does. It is not obvious that, for example, if the direction X -t Y produces a cycle in an otherwise acyclic graph, the opposite direction Y -t X will not also. However, this is the case. For the proof of this, see [Margaritis99]. The algorithm is similar to the SGS algorithm presented in [Spirtes93], but differs in a number of ways. Its main difference lies in the use of Markov blankets to dramatically improve performance (in many cases where the bounded blanket size assumptions hold) . Its structure is similar to SGS, and the stability (frequently referred to as robustness in the following discussion) arguments presented in [Spirtes93] apply. Increased reliability stems from the use of smaller conditioning sets, leading to greater number of examples per test. The PC algorithm, also in [Spirtes93], differs from the GS algorithm in that it involves linear probing for a separator set, which makes it unnecessarily inefficient. 4 Randomized Version of the GS Algorithm The GS algorithm, as presented above, is appropriate for situations where the maximum Markov blanket of each of a set of variables is small. While it is reasonable to assume that in many real-life problems where high-level variables are involved this may be the case, other problems such as Bayesian image retrieval in computer vision, may employ finer representations. In these cases the variables used may depend in a direct manner on many others. For example, we may choose to use variables to characterize local texture in different parts of an image. If the resolution of the mapping from textures to variables is increasingly fine, direct dependencies among those variables may be plentiful and therefore the maximum Markov blanket size may be significant. Another problem that has plagued independence-test based algorithms for Bayesian net structure induction in general is that their decisions are based on a single or a few tests ("hard" decisions), making them prone to errors due to noise in the data. This also applies to the the GS algorithm. It would therefore be advantageous to employ multiple tests before deciding on a direct neighbor or the direction of an edge. The randomized version of the GS algorithm addresses these two problems. Both of them are tackled through randomized testing and Bayesian evidence accumulation. The problem of exponential running times in the maximum blanket size of steps 2 and 3 of the plain algorithm is overcome by replacing them by a series of tests, whose number may be specified by the user, with the members of the conditioning set chosen randomly from the smallest blanket of the two variables. Each such test provides evidence for or against the direct connection between the two variables, appropriately weighted by the probability that circumstances causing that event occur or not, and due to the fact that connectedness is the conjunction of more elementary events. This version of the algorithm is not shown here in detail due to space restrictions. Its operation follows closely the one of the plain GS version. The main difference lies in the usage of Bayesian updating of the posterior probability of a direct link (or a dependence through a collider) between a pair of variables X and Y using conditional dependence tests that take into account independent evidence. The posterior probability Pi of a link between X and Y after executing i dependence tests dj, j = 1, .. . , i is Pi= Pi-ldi Pi-d(G ------------~---------- Pi-ldi + (1 - +1- dd where G == G(X, Y) = 1 - (4)ITI is a factor that takes values in the interval [0,1) and can be interpreted as the "(un)importance" of the truth of each test di , while T is the smaller of B(X) - {Y} and B(Y) - {X}. We can use this accumulated evidence to guide our decisions to the hypothesis that we feel most confident about. Besides being able to do that in a timely manner due to the user-specified number of tests, we also note how this approach also addresses the robustness problem mentioned above through the use of mUltiple weighted tests, and leaving for the end the "hard" decisions that involve a threshold (ie. comparing the posterior probability with a threshold, which in our case is ~) . D. Margaritis and S. Thrun 510 Kl-divergance verSUS number of samples 0.00015 r--~--C...-~--~;;::Pla-:-in-::G::::SCN:-_ -_--, Randomized GS8N ....- .. Hill-Clil"l"tling. score' data likelihood .. ? ,. Hill-Glirnblng, soore: BIC - .Q--- 00001 5e-05 BOOO 12000 Nurrber of sarrples 4000 __- __ l00r-----------::p~lai~nG~Sr-BN~-~ -~ PlainGSBN Randomized GSBN .... ~ ... Hill-Clirrtling, score data likelihood 125 I 100 ~ 75 .s Hill-Climbing, soore: BIC - !w ' . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ... . . . . . . . . . . . . . . . . . . III- ...... . . ...... . . . . SO Randomized GSBN .... ~... Hill-Clirrbi~h~~i~~~~h~ .~_ . ,.. B - ~.~ GO 20000 Direction errors versus number of sarrplss Edge errors versus number of sarrples 1~r---_--_- 16000 1? 75 ~ 50 i 25 ~ S i5 25 o OL---~--~--~---~-~ o 4000 8000 12000 Nurrber of sarrples 18000 20000 0 4000 8000 12000 18000 20000 Number of 5a!1l)les Figure 3: Results for a 5 x 5 rectangular net with branching factor 2 (in both directions, blanket size 8) as a function of the number of samples. On the top, KL-divergence is depicted for the plain GS , randomized GS, and hill-climbing algorithms. On the bottom, the percentage of edge and direction errors are shown. Note that certain edge error rates for the hill-climbing algorithm exceed 100%. 5 Results Throughout the algorithms presented in this paper we employ standard chi-square (X 2 ) conditional dependence tests (as is done also in [Spirtes93]) in order to compare the histograms P(X) and P(X I Y). The X2 test gives us a probability of the error of assuming that the two variables are dependent when in fact they are not (type II error of a dependence test), from which we can easily derive the probability that X and Y are dependent. There is an implicit confidence threshold T involved in each dependence test, indicating how certain we wish to be about the correctness of the test without unduly rejecting dependent pairs, something that is always possible in reality due to the presence of noise. In all experiments we used T = 0.95, which corresponds to a 95% confidence test. We test the effectiveness of the algorithms through the following procedure: we generate a random rectangular net of specified dimensions and up/down branching factor. A number of examples are drawn from that net using logic sampling and they are used as input to the algorithm under test. The resulting nets can be compared with the original ones along dimensions of KL-divergence and difference in edges and edge directionality. The KLdivergence was estimated using a Monte Carlo procedure. An example reconstruction was shown in the beginning of the paper, Fig. 1. Fig. 3 shows how the KL-divergence between the original and the reconstructed net as well as edge omissions/false additions/reversals as a function of number of samples used. It demonstrates two facts. First, that typical KL-divergence for both GS and hill-climbing algorithms is low (with hill-climbing slightly lower), which shows good performance for applications where prediction is of prime concern. Second, the number of incorrect edges and the errors in the directionality of the edges present is much higher for the hill-climbing algorithm, making it unsuitable for accurate Bayesian net reconstruction. Fig. 4 shows the effects of increasing the Markov blanket through an increasing branching factor. As expected, we see a dramatic (exponential) increase in execution time of the plain Bayesian Network Induction via Local Neighborhoods 511 Edge I Direction Errors versus Branching Factor Execution Time versus Branching Factor 100r-------~~~----~~~----, Edge errors, plain GSBN ~ Edge errors, randomized GSBN ---~--Direction errors, plain GSBN Direction errors, randomized GSBN - &--- 90 80 70 ~ 60 ~ 50 ~ 40 i= 30 ___ __ ___ _ __ _...___ ___ ___ .... ___ __ _ 2?L==~==~~?----=----=-----=----~----------- ----l 10 -.,,""'-.. -----...... .. _. ? _ _" _______ .olII ___ ____ __ ? ___ O~------~--------~--------~ 2 3 4 Branching Factor 5 22000 20000 18000 16000 14000 12000 10000 8000 6000 4000 2000 Plain GSBN - Randomized GSBN ----K---- .----- O~------~--------~------~ 2 3 4 Branching Factor 5 Figure 4: Results for a 5 x 5 rectangular net from which 10000 samples were generated and used for reconstruction, versus increasing branching factor. On the left, errors are slowly increasing as expected, but comparable for the plain and randomized versions of the GS algorithm. On the right, corresponding execution times are shown. GS algorithm, though only a mild increase of the randomized version. The latter uses 200 (constant) conditional tests per decision, and its execution time increase can be attributed to the (quadratic) increase in the number of decisions. Note that the error percentages between the plain and the randomized version remain relatively close. The number of direction errors for the GS algorithm actually decreases due to the larger number of parents for each node (more "V" structures), which allows a greater number of opportunities to recover the directionality of an edge (using an increased number of tests). 6 Discussion In this paper we presented an efficient algorithm for computing the Markov blanket of a node and then used it in the two versions of the GS algorithm (plain and randomized) by exploiting the properties of the Markov blanket to facilitate fast reconstruction of the local neighborhood around each node, under assumptions of bounded neighborhood size. We also presented a randomized variant that has the advantages of faster execution speeds and added reconstruction robustness due to multiple tests and Bayesian accumulation of evidence. Simulation results demonstrate the reconstruction accuracy advantages of the algorithms presented here over hill-climbing methods. Additional results also show that the randomized version has a dramatical execution speed benefit over the plain one in cases where the assumption of bounded neighborhood does not hold, without significantly affecting the reconstruction error rate. References [Chow68] [Herskovits90] [Spirtes93] [PearI88] [Rebane87] [Verma90] [Agosta88] [Cheng97] [Margaritis99] [Jtinger85] C.K. Chow and C.N. Liu. Approximating discrete probability distributions with dependence trees. IEEE Transactions on Information Theory, 14, 1968. E.H. Herskovits and G.F. Cooper. Kutat6: An entropy-driven system for construction of probabilistic expert systems from databases. VAI-90. P. Spirtes, C. Glymour, and R. Scheines. Causation, Prediction, and Search, Springer, 1993. 1. Pearl. Probabilistic Reasoning in Intelligent Systems, Morgan Kaufmann, 1988. G. Rebane and J. Pearl. The recovery of causal poly-trees from statistical data. VAI-87. T.S. Verma, and J. Pearl. Equivalence and Synthesis of Causal Models. VAI-90. J.M. Agosta. The structure of Bayes networks for visual recognition. VAI-88. 1. Cheng, D.A. Bell, W. Liu, An algorithm for Bayesian network construction from data. AI and Statistics, 1997. D. Marg aritis , S. Thrun, Bayesian Network Induction via Local Neighborhoods. TR CMV-CS-99-134, forthcoming. M. Junger, Polyhedral combinatorics and the acyclic subdigraph problem, Heldermann, 1985.
1685 |@word mild:2 h:2 version:13 polynomial:2 advantageous:1 heuristically:1 simulation:1 propagate:1 bn:1 pressure:1 dramatic:1 mention:1 tr:1 plentiful:1 contains:1 score:3 selecting:1 series:1 liu:2 denoting:1 current:1 comparing:1 surprising:1 attracted:1 nb2:1 happen:1 partition:2 remove:5 greedy:1 beginning:1 provides:1 node:18 idi:1 along:1 direct:12 become:2 incorrect:2 prove:1 consists:1 descendant:2 polyhedral:1 introduce:2 manner:3 expected:2 indeed:1 nor:1 growing:1 multi:1 frequently:1 ol:1 chi:1 automatically:3 little:1 increasing:4 estimating:1 underlying:2 discover:1 bounded:5 notation:1 what:3 interpreted:1 kldivergence:1 booo:1 preferably:1 wrong:1 demonstrates:1 normally:1 causally:1 before:1 local:11 limit:1 path:2 becoming:1 connectedness:1 equivalence:1 shaded:1 directed:4 unique:1 pla:1 testing:1 practice:1 differs:2 procedure:2 maxx:1 significantly:1 bell:1 word:1 confidence:3 agosta:1 close:1 wrongly:1 put:1 marg:1 accumulation:3 measurable:1 restriction:1 go:2 attention:1 straightforward:1 rectangular:4 focused:1 resolution:1 identifying:4 recovery:4 rule:1 stability:1 notion:1 feel:1 construction:2 today:3 user:3 us:2 hypothesis:1 l00r:1 pa:2 expensive:1 recognition:1 updating:1 database:1 bottom:1 ft:2 cycle:13 connected:1 ordering:1 decrease:1 removed:1 disease:1 mentioned:2 complexity:1 depend:1 efficiency:1 completely:1 necessitates:1 easily:2 joint:1 represented:1 derivation:1 fast:2 shortcoming:1 monte:1 neighborhood:10 outside:1 disjunction:1 whose:3 emerged:1 widely:1 larger:2 heuristic:1 say:1 otherwise:1 ability:1 statistic:1 itself:1 noisy:1 advantage:3 indication:1 net:18 reconstruction:8 causing:1 loop:1 intervening:1 inducing:2 description:1 exploiting:1 parent:10 produce:2 executing:1 derive:1 ldi:2 c:3 involves:1 blanket:36 come:1 direction:14 collider:2 closely:3 dramatical:1 really:1 elementary:1 insert:1 hold:2 around:3 plagued:1 great:2 deciding:1 mapping:1 nurrber:2 smallest:1 uniqueness:2 estimation:2 cmv:1 correctness:3 tool:1 weighted:2 always:1 conjunction:1 likelihood:3 mainly:1 greatly:1 contrast:1 inference:1 dependent:8 accumulated:1 typically:2 chow:1 interested:1 semantics:1 among:3 mutual:1 field:3 construct:1 ng:1 sampling:1 represents:2 unnecessarily:1 np:1 idt:1 others:1 intelligent:1 employ:4 few:1 causation:1 pathological:1 randomly:1 divergence:4 phase:2 attempt:2 interest:2 highly:1 admitting:1 pc:2 behind:2 accurate:1 edge:25 necessary:2 tree:4 loosely:1 causal:6 minimal:1 increased:2 disadvantage:2 subset:3 successful:2 too:1 characterize:1 dependency:2 confident:1 density:2 fundamental:1 randomized:21 ie:2 probabilistic:2 connecting:1 synthesis:1 central:1 satisfied:1 choose:1 possibly:1 slowly:1 expert:1 inefficient:1 leading:1 account:1 potential:1 includes:3 permanent:1 combinatorics:1 later:2 try:1 doctor:1 bayes:2 recover:1 timely:1 yen:1 minimize:1 square:1 accuracy:1 kaufmann:1 yield:2 identify:1 climbing:9 directional:1 yes:1 bayesian:24 rejecting:1 produced:1 emphasizes:1 carlo:1 finer:1 straight:1 sebastian:1 definition:1 against:1 involved:5 obvious:1 proof:3 di:1 attributed:1 popular:1 knowledge:1 anytime:1 actually:2 higher:2 maximally:1 daunting:1 execute:1 shrink:2 though:2 generality:2 done:1 implicit:1 until:2 horizontal:1 replacing:1 widespread:1 indicated:1 scientific:1 believe:1 orienting:1 usage:1 effect:1 facilitate:3 concept:2 requiring:1 building:1 former:1 equality:1 spirtes:1 during:3 branching:9 hill:13 complete:1 demonstrate:1 reasoning:1 image:2 common:2 conditioning:10 mellon:2 significant:1 ai:1 dj:1 reliability:2 longer:1 add:1 something:1 posterior:3 recent:1 belongs:1 irrelevant:1 reverse:2 prime:1 driven:1 certain:5 life:1 scoring:1 preserving:1 minimum:3 greater:2 additional:1 morgan:1 determine:2 ii:1 multiple:4 stem:2 faster:3 long:1 retrieval:1 lai:1 impact:1 prediction:3 variant:2 basic:1 vision:2 cmu:2 circumstance:1 histogram:1 represent:1 sometimes:2 beam:1 addition:1 affecting:1 fine:1 addressed:2 interval:1 grow:2 leaving:1 appropriately:1 operate:1 sr:1 ivi:2 member:4 effectiveness:1 presence:1 exceed:1 split:2 iii:1 concerned:3 independence:7 bic:2 forthcoming:1 identified:1 restrict:1 opposite:1 idea:2 proneness:2 whether:1 cause:2 dramatically:1 detailed:1 involve:1 coincidentally:1 category:1 generate:1 exist:2 percentage:2 herskovits:1 dotted:1 estimated:1 per:2 diagnosis:1 carnegie:2 discrete:1 independency:2 threshold:3 blood:1 drawn:1 neither:1 graph:7 year:1 orient:3 facilitated:1 i5:1 ameliorate:1 throughout:1 reasonable:2 utilizes:1 parsimonious:1 decision:7 comparable:2 bound:1 tackled:1 cheng:1 quadratic:1 identifiable:1 g:15 occur:1 x2:2 speed:3 argument:1 rendered:1 relatively:1 glymour:1 department:2 smaller:8 ascertain:1 increasingly:1 slightly:1 remain:1 making:6 restricted:1 heart:1 legal:1 scheines:1 ascending:1 end:1 reversal:1 available:1 operation:1 apply:2 observe:1 worthwhile:1 appropriate:1 robustness:3 scn:1 existence:1 original:2 top:1 running:2 opportunity:1 unsuitable:1 exploit:1 disappear:2 approximating:1 bl:4 added:4 quantity:3 already:1 dependence:13 reversed:2 separate:1 link:2 thrun:6 separating:1 majority:1 reason:1 induction:7 assuming:3 besides:2 relationship:3 difficult:1 unfortunately:1 margaritis:5 vai:4 vertical:1 observation:1 markov:29 arc:1 iti:1 incorrectly:1 situation:2 omission:1 introduced:1 namely:2 pair:2 specified:4 kl:5 connection:2 unduly:1 pearl:3 address:3 able:3 below:1 exemplified:1 dimitris:1 usually:1 reliable:3 including:1 lend:1 greatest:1 event:2 indicator:1 representing:1 improve:1 identifies:1 created:1 prior:1 sg:3 discovery:2 loss:1 interesting:1 acyclic:2 versus:6 illuminating:1 consistent:1 dd:1 verma:1 pi:5 prone:1 penalized:1 last:1 dis:1 formal:2 side:1 understand:1 guide:1 wide:1 neighbor:7 benefit:3 boundary:3 plain:12 feedback:1 world:1 overcome:1 dimension:2 preventing:1 preprocessing:1 employing:3 transaction:1 reconstructed:1 compact:2 logic:1 pittsburgh:2 unnecessary:1 search:2 un:1 reality:1 robust:1 poly:2 necessarily:1 separator:1 domain:1 dense:1 main:4 motivation:1 proliferate:1 noise:2 succinct:1 child:5 fig:5 referred:2 cooper:1 probing:1 shield:1 shrinking:1 fails:1 explicit:1 wish:1 exponential:4 lie:2 ib:1 down:1 disregarding:1 evidence:6 concern:1 exists:4 false:1 effectively:1 importance:2 texture:2 execution:7 conditioned:1 heldermann:1 easier:1 entropy:1 depicted:1 logarithmic:1 likely:1 visual:1 applies:2 springer:1 corresponds:1 truth:1 determines:2 conditional:8 hard:2 directionality:5 determined:2 typical:1 called:2 pas:1 accepted:1 indicating:1 latter:1 violated:1
749
1,686
Audio-Vision: Using Audio-Visual Synchrony to Locate Sounds John Hershey .. Javier Movellan jhershey~cogsci.ucsd.edu movellan~cogsci.ucsd.edu Department of Cognitive Science University of California, San Diego La Jolla, CA 92093-0515 Department of Cognitive Science University of California, San Diego La Jolla, CA 92093-0515 Abstract Psychophysical and physiological evidence shows that sound localization of acoustic signals is strongly influenced by their synchrony with visual signals. This effect, known as ventriloquism, is at work when sound coming from the side of a TV set feels as if it were coming from the mouth of the actors. The ventriloquism effect suggests that there is important information about sound location encoded in the synchrony between the audio and video signals. In spite of this evidence, audiovisual synchrony is rarely used as a source of information in computer vision tasks. In this paper we explore the use of audio visual synchrony to locate sound sources. We developed a system that searches for regions of the visual landscape that correlate highly with the acoustic signals and tags them as likely to contain an acoustic source. We discuss our experience implementing the system, present results on a speaker localization task and discuss potential applications of the approach. Introd uction We present a method for locating sound sources by sampling regions of an image that correlate in time with the auditory signal. Our approach is inspired by psychophysical and physiological evidence suggesting that audio-visual contingencies play an important role in the localization of sound sources: sounds seem to emanate from visual stimuli that are synchronized with the sound. This effect becomes particularly noticeable when the perceived source of the sound is known to be false, as in the case of a ventriloquist's dummy, or a television screen. This phenomenon is known in the psychophysical community as the ventriloquism effect, defined as a mislocation of sounds toward their apparent visual source. The effect is robust in a wide variety of conditions, and has been found to be strongly dependent on the degree of "synchrony" between the auditory and visual signals (Driver, 1996; Bertelson, Vroomen, Wiegeraad & de Gelder, 1994). "1'0 whom correspondence should be addressed. 814 J. Hershey and J. R. Movellan The ventriloquism effect is in fact less speech-specific than first thought. For example the effect is not disrupted by an upside-down lip signal (Bertelson, Vroomen, Wiegeraad & de Gelder, 1994) and is just as strong when the lip signals are replaced by light flashes that are synchronized with amplitude peaks in the audio signal (Radeau & Bertelson, 1977). The crucial aspect here is correlation between visual and auditory intensity over time. When the light flashes are not synchronized the effect disappears. The ventriloquism effect is strong enough to produce an enduring localization bias, known as the ventriloquism aftereffect. Over time, experience with spatially offset auditory-visual stimuli causes a persistent shift in subsequent auditory localization. Exposure to audio-visual stimuli offset from each other by only 8 degrees of azimuth for 20-30 minutes is sufficient to shift auditory localization by the same amount. A corresponding shift in neural processing has been detected in macaque monkeys as early as primary auditory cortex(Recanzone, 1998). In barn owls a misalignment of visual and auditory stimuli during development causes the realignment of the auditory and visual maps in the optic tectum(Zheng & Knudsen, 1999; Stryker, 1999; Feldman & Knudsen, 1997). The strength of the psychophysical and physiological evidence suggests that audiovisual contingency may be used as an important source of information that is currently underutilized in computer vision tasks. Visual and auditory sensor systems carry information about the same events in the world, and this information must be combined correctly in order for a useful interaction of the two modalities. Audiovisual contingency can be exploited to help determine which signals in different modalities share a common origin. The benefits are two-fold: the two signals can help localize each other, and once paired can help interpret each other. To this effect we developed a system to localize speakers using input from a camera and a single microphone. The approach is based on searching for regions of the image which are "synchronized" with the acoustic signal. Measuring Synchrony The concept of audio-visual synchrony is not well formalized in the psychophysical literature, so for a working definition we interpret synchrony as the degree of mutual information between audio and spatially localized video signals. Ultimately it is a causal relationship that we are often interested in, but causes 'can only be inferred from effects such as synchrony. Let a(t) E IRn be a vector describing the acoustic signal at time t. The components of a(t) could be cepstral coefficients, pitch measurements, or the outputs of a filter bank. Let v(x, y, t) E IRm be a vector describing the visual signal at time t, pixel (x,y). The components ofv(x,y,t) could represent Gabor energy coefficients, RGB color values, etc. Consider now a set of s audio and visual vectors S = (a(tl), v(x, y, tl?l=k-s-l,. .. ,k sampled at times tk-s-l,'" ,tk and at spatial coordinates (x, y). Given this set of vectors our goal is to provide a number that describes the temporal contingency between audio and video at time tk' The approach we take is to consider each vector in S as an independent sample from a joint multivariate Gaussian process (A(tk), V(x, y, tk? and define audio-visual synchrony at time tk as the estimate of the mutual information between the audio and visual components of the process. Let A(tk) ,..., Nn(ltA(tk), ~A(tk?' and V(x,y, tk) ,..., Nm(ltv(x, y, t), ~v(x,y, tk)), where It represents means and ~ covariance matrices. Let A(tk) and V(x, y, tk) be jointly Gaussian, i.e., (A(tk), V(x, y, tk? ,..., Nn+m(ltA,V (x, y, tk), ~A,V(X, y, tk)' 815 Audio Vision: Using Audio-Vzsual Synchrony to Locate Sounds The mutual information between A(x, y, tk) and V(tk) can be shown to be as follows [(A(tk); V(x, y, tk)) = H(A(tk)) + H(V(x, y, tk)) 1 "2log(27re)nIEA(tk)1 - H(A(tk), V(x, y, tk)) 1 + "2log(27re)mIEv(x, y, tk)1 (1) 1 -"2log(27re)n+mIEA ,v(x,y, tk)1 (2) IEA(tk)IIEv(x,y,tk)1 -11og "'-----:-::::---'--'-'--;----'-----'-::-:--':"':' 2 IEA,V(X,y,tk)I' In the special case that n (3) = m = 1, then (4) where p(x, y, tk) is the Pearson correlation coefficient between A(tk) and V(x, y, tk)' For each triple (x, y, tk) we estimate the mutual information between A( tk) and V(x, y, tk) by considering each element of S as an independent sample from the random vector (A(tk), V(x, y, tk))' This amounts to computing estimates of the joint covariance matrix EA,V (x, y, tk). For example the estimate of the covariance between the ith audio component and the jLh video component would be as follows 1 SAi,v; (x, y, tk) 8-1 = s _ 1 I)ai(tk-l) - ai(tk))(Vj(X,y, tk-l) - Vj (x, y, tk)), (5) 1=0 where (6) (7) (8) These simple covariance estimates can be computed recursively in constant time with respect to the number of timepoints. The independent treatment of pixels would lend well to a parallel implementation. To measure performance, a secondary system produces a single estimate of the auditory location, for use with a database of labeled solitary audiovisual sources. Unfortunately there are many ways of producing such estimates so it becomes difficult to separate performance of the measure from the underlying system. The model used here is a centroid computation on the mutual information estimates, with some enhancements to aid tracking and reduce background noise. Implementation Issues A real time system was prototyped using a QuickCam on the Linux operating system and then ported to NT as a DirectShow filter. l'his platform provides input from real-time audio and video capture hardware as well as from static movie files. The video output could also be rendered live or compressed and saved in a movie file. The implementation was challenging in that it turns out to be rather difficult 816 J. Hershey and J. R. Movellan -2 -'.'----f:.20--"~--:':"--""=:----:':''':--~'20 F,_ (a) M is talking. ~.~~'~.~20~-~"=:---"~~~-~"-=~-~"~"? F". . . (b) J is talking. Figure 1: Normalized audio and visual intensity across sequences of frames in which a sequence of four numbers is spoken. The top trace is the contour of the acoustic energy from one of two speakers, M or J, and the bottom trace is the contour of intensity values for a single pixel, (147,100), near the mouth of J. to process precisely time-synchronized audio and video on a serial machine in real time. Multiple threads are required to read from the peripheral audio and visual devices. By the time the audio and visual streams reach the AV filter module, they are quite separate and asynchronous. The separately threaded auditory and visual packet streams must be synchronized, buffered, and finally matched and aligned by time-stamps before they can finally be processed. It is interesting that successful biologial audiovisual systems employ a parallel architecture and thus avoid this problem. Results To obtain a performance baseline we first tried the simplest possible approach: A single audio and visual feature per location: n = m = 1, v(x, y, t) E IR is the intensity of pixel (x, y) at time t, and a(t) E IR is the average acoustic energy over the interval [t - 6.t, tJ, where 6.t = 1/30 msec , the sampling period for the NTSC video signal. Figure 1 illustrates the time course of these signals for a non-synchronous and a synchronous pair of acoustic energy and pixel intensity. Notice in particular that in the synchonous pair, 1(b), where the sound and pixel values come from the same speaker, the relationship between the signals changes over time. There are regions of positive and negative covariance strung together in succession. Clearly the relationship over the entire sequence is far from linear. However over shorter time periods a linear relationship looks like a better approximation. Our window size of 16 samples (Le., s = 16 in 5 coincides approximately with this time-scale. Perhaps by averaging over many small windows we can capture on a larger scale what would be lost to the same method applied with a larger window. Of course there is a trade-off in the time-scale between sensitivity to spurious transients, and the response time of the system. We applied this mutual information measure to all the pixels in a movie, in the spirit of the perceptual maps of the brain. The result is a changing topographic map of audiovisual mutual information. Figure 2 illustrates two snapshots in which 817 Audio Vision: Using Audio-Visual Synchrony to Locate Sounds (a) Frame 206: M (at left) is talking. (b) Frame 104: J (at right) is talking. Figure 2: Estimated mutual information between pixel intensity and audio intensity (bright areas indicate greater mutual information) overlaid on stills from the video where one person is in mid-utterance. different parts of the face are synchronous (possibly with different sign) with the sound they take part in producing. It is interesting that the synchrony is shared by some parts, such as the eyes, that do not directly contribute to the sound, but contribute to the communication nonetheless. ___ To estimate the position of the speaker we computed a centroid were each point was weighted by the estimated mutual information between the correpsonding pixel and the audio signal. At each time step the mutual information was estimated using 16 past frames (Le., s = 16) In order to reduce the intrusion of spurious correlations from competing targets, once a target has been found, we employ a Gaussian influence function. (Goodall, 1983) The influence function reduces the weight given to mutual information from locations far from the current centroid when computing the next centroid. To allow for the speedy disengagement from a dwindling source of mutual information we set a threshold on the mutual information. Measurements under the threshold are treated as zero. This threshold also reduces the effects of unwanted background noise, such as camera and microphone jitter. A Sx(t) = Lx L x 8(1og(1 - f} (x , y, t)))'I/;(X, Sx(t - 1)) Y A Lx L y 8(log(1- p2(X,y,t)))'I/;(x, Sx(t -1)) (9) where Sx(t) represents the estimate of the x coordinate for the position of the speaker at time t. 8(.) is the thresholding function, and 'I/;(x, Sx(t - 1)) is the influence function , which depends upon the 'position x of the pixel being sampled and the prior estimate Sx(t-1). p2(X, y, t) is the estimate of the correlation between the intensity in pixel (x , y) and the acoustic enery, when using the 16 past video frames. -~ log(l- p2(x, y, t)) is the corresponding estimate of mutual information (the factor, -~ cancels out in the quotient after adjusting the threshold function accordingly. ) We tried the approach on a movie of two people (M and J) taking turns while saying random digits. Figure 3 shows the estimates of the actual positions of the speaker J. Hershey and J. R. Movellan 818 as a function of time. The estimates clearly provide information that could be used to localize the speaker, especially in combination with other approaches (e.g., flesh detection) . 180 180 ~ ~ 140 ~ ~ 120 i ~ 100 ~ !_ 80 40 20~----~----~----~----~----~------~--~ o 100 200 300 400 500 600 700 Frame Number Figure 3: Estimated and actual position of speaker at each frame for six hundred frames. The sources, M and J, took turns uttering a series of four digits, for three turns each. The actual positions and alternation times were measured by hand from the video recording Conclusions We have presented exploratory work on a system for localizing sound sources on a video signal by tagging regions of the image that are correlated in time with the auditory signal. The approach was motivated by the wealth of evidence in the psychophysical and physiological literature showing that sound localization is strongly influenced by synchrony with the visual signal. We presented a measure of local synchrony based on modeling the audio-visual signal as a non-stationary Gaussian process. We developed a general software tool that accepts as inputs all major video and audio file formats as well as direct input from a video camera. We tested the tool on a speaker localization task with very encouraging results. The approach could have practical applications for localizing sound sources in situations where where acoustic stereo cues are inexistent or unreliable. For example the approach could be used to help localize the actor talking in a video scene and put closed-captioned text near the audio source. The approach could also be used to guide a camera in teleconferencing applications. While the results reported here are very encouraging, more work needs to be done before practical applications are developed. For example we need to investigate more sophisticated methods for processing the audio and video signals. At this point we use average energy to represent the video and thus changes in the fundamental frequency that do not affect the average energy would not be captured by our model. Similarly local video decompositions, like spatio-temporal Gabor filtering, or approaches designed to enhance the lip regions may be helpful. The Audio Vision: Using Audio-Visual Synchrony to Locate Sounds 819 changing symmetry observed between audio and video signals might be addressed rectifying or squaring the normalized signals and derivatives. Finally, relaxing the Gaussian constraints in our measure of audio-visual contingency may help improve performance. While the work shown here is exploratory at this point, the approach is very promising: It emphasizes the idea of machine perception as a multimodal process it is backed by psychophysical evidence, and when combined with other approaches it may help improve robustness in tasks such as localization and separation of sound sources. References Bertelson, P., Vroomen, J., Wiegeraad, G., & de Gelder, B. (1994). Exploring the relation between McGurk interference and ventriloquism. In Proceedings of the 1994 International Conference on Spoken Language Processing, volume 2, pages 559-562. Driver, J. (1996). Enhancement of selective listening by illusory mislocation of speech sounds due to lip-reading. Nature, 381, 66-68. Feldman, D. E . & Knudsen, E. I. (1997). An anatomical basis for visual calibration of the auditiory space map in the barn owl's midbrain. The Journal of Neuroscience, 17(17), 6820-6837. Goodall, C. (1983). M-Estimators of Location: an outline of the theory. Wiley series in probability and mathematical statistics. Applied probability and statistics. Radeau, M. & Bertelson, P. (1977). Adaptation to auditory-visual discordance and ventriloquism in semi-realistic situations. Perception and Psychophysics, 22, 137-146. Recanzone, G. H. (1998). Rapidly induced auditory plasticity: The ventriloquism aftereffect. Proceedings of the National Academy of Sciences, USA, 95, 869- 875. Stryker, M. P. (1999) . Sensory Maps on the Move. Science, 925-926. Zheng, W. & Knudsen, E. I. (1999). Functional Selection of Adaptive Auditory Space Map by GABAA-Mediated Inhibition, 962-965.
1686 |@word tried:2 rgb:1 covariance:5 decomposition:1 recursively:1 carry:1 series:2 past:2 current:1 nt:1 must:2 john:1 realistic:1 subsequent:1 plasticity:1 designed:1 stationary:1 cue:1 device:1 accordingly:1 ith:1 provides:1 contribute:2 location:5 lx:2 mathematical:1 direct:1 driver:2 persistent:1 tagging:1 brain:1 audiovisual:6 inspired:1 actual:3 encouraging:2 window:3 considering:1 becomes:2 underlying:1 matched:1 what:1 gelder:3 monkey:1 developed:4 spoken:2 temporal:2 sai:1 unwanted:1 producing:2 before:2 positive:1 local:2 approximately:1 might:1 suggests:2 challenging:1 relaxing:1 practical:2 camera:4 lost:1 movellan:5 digit:2 emanate:1 area:1 thought:1 gabor:2 spite:1 selection:1 put:1 live:1 influence:3 map:6 backed:1 exposure:1 formalized:1 estimator:1 his:1 searching:1 exploratory:2 coordinate:2 feel:1 ntsc:1 tectum:1 diego:2 play:1 vzsual:1 target:2 origin:1 element:1 particularly:1 database:1 labeled:1 bottom:1 role:1 module:1 observed:1 capture:2 region:6 trade:1 ultimately:1 localization:9 upon:1 ofv:1 gabaa:1 misalignment:1 basis:1 teleconferencing:1 multimodal:1 joint:2 cogsci:2 detected:1 pearson:1 apparent:1 encoded:1 quite:1 larger:2 compressed:1 statistic:2 topographic:1 jointly:1 sequence:3 took:1 interaction:1 coming:2 adaptation:1 aligned:1 rapidly:1 academy:1 enhancement:2 produce:2 tk:45 help:6 measured:1 noticeable:1 p2:3 strong:2 quotient:1 flesh:1 come:1 indicate:1 synchronized:6 saved:1 filter:3 packet:1 transient:1 implementing:1 owl:2 exploring:1 barn:2 overlaid:1 major:1 early:1 perceived:1 inexistent:1 currently:1 tool:2 weighted:1 clearly:2 sensor:1 gaussian:5 rather:1 ltv:1 avoid:1 og:2 intrusion:1 centroid:4 baseline:1 helpful:1 dependent:1 squaring:1 nn:2 entire:1 spurious:2 irn:1 captioned:1 relation:1 selective:1 interested:1 pixel:11 issue:1 ported:1 development:1 spatial:1 special:1 platform:1 mutual:15 psychophysics:1 once:2 sampling:2 represents:2 look:1 cancel:1 bertelson:5 stimulus:4 employ:2 national:1 replaced:1 detection:1 highly:1 investigate:1 radeau:2 zheng:2 light:2 tj:1 experience:2 shorter:1 irm:1 re:3 causal:1 modeling:1 measuring:1 localizing:2 hundred:1 successful:1 azimuth:1 reported:1 combined:2 person:1 disrupted:1 peak:1 sensitivity:1 fundamental:1 international:1 off:1 enhance:1 together:1 realignment:1 linux:1 nm:1 mcgurk:1 possibly:1 cognitive:2 derivative:1 suggesting:1 potential:1 de:3 coefficient:3 depends:1 stream:2 jlh:1 closed:1 parallel:2 vroomen:3 synchrony:17 rectifying:1 bright:1 ir:2 succession:1 landscape:1 emphasizes:1 recanzone:2 influenced:2 reach:1 definition:1 energy:6 nonetheless:1 frequency:1 static:1 sampled:2 auditory:16 treatment:1 adjusting:1 illusory:1 color:1 javier:1 amplitude:1 sophisticated:1 ea:1 hershey:4 response:1 done:1 strongly:3 just:1 correlation:4 working:1 hand:1 perhaps:1 usa:1 effect:12 contain:1 concept:1 normalized:2 spatially:2 read:1 during:1 speaker:10 coincides:1 outline:1 image:3 common:1 functional:1 volume:1 interpret:2 measurement:2 buffered:1 feldman:2 ai:2 similarly:1 language:1 calibration:1 actor:2 cortex:1 operating:1 inhibition:1 etc:1 multivariate:1 jolla:2 alternation:1 exploited:1 captured:1 greater:1 ventriloquist:1 determine:1 period:2 signal:26 semi:1 upside:1 sound:22 multiple:1 reduces:2 serial:1 paired:1 pitch:1 vision:6 represent:2 background:2 separately:1 addressed:2 interval:1 wealth:1 source:15 crucial:1 modality:2 file:3 recording:1 induced:1 spirit:1 seem:1 near:2 enough:1 variety:1 affect:1 architecture:1 competing:1 reduce:2 idea:1 listening:1 shift:3 synchronous:3 thread:1 six:1 motivated:1 introd:1 stereo:1 locating:1 speech:2 cause:3 useful:1 amount:2 mid:1 hardware:1 processed:1 simplest:1 notice:1 sign:1 estimated:4 neuroscience:1 dummy:1 correctly:1 per:1 anatomical:1 four:2 threshold:4 localize:4 changing:2 jitter:1 solitary:1 saying:1 separation:1 correspondence:1 fold:1 iea:2 strength:1 optic:1 precisely:1 constraint:1 scene:1 software:1 tag:1 aspect:1 rendered:1 format:1 department:2 tv:1 peripheral:1 combination:1 underutilized:1 describes:1 across:1 midbrain:1 goodall:2 interference:1 aftereffect:2 discus:2 describing:2 lta:2 turn:4 robustness:1 top:1 especially:1 psychophysical:7 move:1 primary:1 stryker:2 separate:2 whom:1 threaded:1 toward:1 relationship:4 difficult:2 unfortunately:1 trace:2 negative:1 implementation:3 av:1 snapshot:1 knudsen:4 situation:2 communication:1 locate:5 frame:8 ucsd:2 community:1 intensity:8 inferred:1 pair:2 required:1 disengagement:1 acoustic:10 california:2 accepts:1 macaque:1 perception:2 reading:1 video:19 lend:1 mouth:2 event:1 treated:1 improve:2 movie:4 eye:1 disappears:1 mediated:1 jhershey:1 utterance:1 text:1 prior:1 literature:2 interesting:2 filtering:1 localized:1 triple:1 contingency:5 degree:3 sufficient:1 thresholding:1 bank:1 share:1 course:2 asynchronous:1 side:1 bias:1 allow:1 guide:1 wide:1 face:1 cepstral:1 taking:1 benefit:1 world:1 contour:2 sensory:1 uttering:1 adaptive:1 san:2 far:2 correlate:2 unreliable:1 spatio:1 search:1 lip:4 promising:1 nature:1 robust:1 ca:2 symmetry:1 vj:2 noise:2 tl:2 screen:1 aid:1 wiley:1 position:6 timepoints:1 msec:1 stamp:1 perceptual:1 down:1 minute:1 specific:1 showing:1 offset:2 physiological:4 evidence:6 uction:1 false:1 enduring:1 illustrates:2 television:1 sx:6 explore:1 likely:1 visual:31 tracking:1 talking:5 prototyped:1 goal:1 flash:2 shared:1 change:2 averaging:1 microphone:2 secondary:1 la:2 ventriloquism:9 rarely:1 speedy:1 people:1 audio:34 tested:1 phenomenon:1 correlated:1
750
1,687
A Geometric Interpretation of v-SVM Classifiers David J. Crisp Centre for Sensor Signal and Information Processing, Deptartment of Electrical Engineering, University of Adelaide, South Australia Christopher J.C. Burges Advanced Technologies, Bell Laboratories, Lucent Technologies Holmdel, New Jersey dcrisp@eleceng.adelaide.edu.au burges@lucent.com Abstract We show that the recently proposed variant of the Support Vector machine (SVM) algorithm, known as v-SVM, can be interpreted as a maximal separation between subsets of the convex hulls of the data, which we call soft convex hulls. The soft convex hulls are controlled by choice of the parameter v. If the intersection of the convex hulls is empty, the hyperplane is positioned halfway between them such that the distance between convex hulls, measured along the normal, is maximized; and if it is not, the hyperplane's normal is similarly determined by the soft convex hulls, but its position (perpendicular distance from the origin) is adjusted to minimize the error sum. The proposed geometric interpretation of v-SVM also leads to necessary and sufficient conditions for the existence of a choice of v for which the v-SVM solution is nontrivial. 1 Introduction Recently, SchOlkopf et al. [I) introduced a new class of SVM algorithms, called v-SVM, for both regression estimation and pattern recognition. The basic idea is to remove the user-chosen error penalty factor C that appears in SVM algorithms by introducing a new variable p which, in the pattern recognition case, adds another degree of freedom to the margin. For a given normal to the separating hyperplane, the size of the margin increases linearly with p. It turns out that by adding p to the primal objective function with coefficient -v, v 2: 0, the variable C can be absorbed, and the behaviour of the resulting SVM - the number of margin errors and number of support vectors - can to some extent be controlled by setting v. Moreover, the decision function produced by v-SVM can also be produced by the original SVM algorithm with a suitable choice of C. In this paper we show that v-SVM, for the pattern recognition case, has a clear geometric interpretation, which also leads to necessary and sufficient conditions for the existence of a nontrivial solution to the v-SVM problem. All our considerations apply to feature space, after the mapping of the data induced by some kernel. We adopt the usual notation: w is the normal to the separating hyperplane, the mapped 245 A Geometric Interpretation ofv-SVM Classifiers data is denoted by Xi E !RN , i = 1, ... ,1, with corresponding labels are scalars, and ~i' i = 1", ,,1 are positive scalar slack variables. 2 Yi E {?1}, b, p v-SVM Classifiers The v-SVM formulation, as given in [1], is as follows: minimize pI = 1 211w/112 - Vp' 1 + y l:~~ (1) i with respect to w', b' , p', ~i, subject to: Yi(W' . Xi + b/ ) ~ p' - ~~, ~i ~ 0, p' ~ o. (2) Here v is a user-chosen parameter between 0 and 1. The decision function (whose sign determines the label given to a test point x) is then: l' (x) = w' . x + b' . The Wolfe dual of this problem is: maximize to Ph (3) = -~ 2:ij OiOjYiYjXi . Xj subject (4) with w' given by w' = 2:i 0iYiXi . SchOlkopf et al. [1] show that v is an upper bound on the fraction of margin errors 1 , a lower bound on the fraction of support vectors, and that both of these quantities approach v asymptotically. Note that the point w' = b' = p = ~i = 0 is feasible, and that at this point, pI = O. Thus any solution of interest must have pI ::; O. Furthermore, if Vp' = 0, the optimal solution is at w' = b' = p = ~i = 02 ? Thus we can assume that v p' > 0 (and therefore v > 0) always. Given this, the constraint p' ~ 0 is in fact redundant: a negative value of p' cannot appear in a solution (to the problem with this constraint removed) since the above (feasible) solution (with p' = 0) gives a lower value for P'. Thus below we replace the constraints (2) by (5) 2.1 A Reparameterization of v-SVM We reparameterize the primal problem by dividing the objective function pI by v 2 /2, the constraints (5) by v, and by making the following substitutions: 2 w' b' p' I-' = -, w = - , b = -, p = -, ~i vl v v v ~i = -. v (6) 1 A margin error Xi is defined to be any point for which ?i > 0 (see [1]). 2In fact we can prove that, even if the optimal solution is not unique, the global solutions still all have w = 0: see Burges and Crisp, "Uniqueness of the SYM Solution" in this volume. 246 D. J. Crisp and C. J. C. Burges This gives the equivalent formulation: minimize (7) with respect to w, b, p, ~i' subject to: (8) = IT we use as decision function f(x) f'(x)/v, the formulation is exactly equivalent, although both primal and dual appear different. The dual problem is now: minimize (9) with respect to the ai, subject to: (10) with w given by w = 1 2: i aiYiXi. In the following, we will refer to the reparameterized version of v-StrM given above as J.'-SVM, although we emphasize that it describes the same problem. 3 A Geometric Interpretation of l/-SVM In the separable case, it is clear that the optimal separating hyperplane is just that hyperplane which bisects the shortest vector joining the convex hulls of the positive and negative polarity points 3 ? We now show that this geometric interpretation can be extended to the case of v-SVM for both separable and nonseparable cases. 3.1 The Separable Case We start by giving the analysis for the separable case. The convex hulls of the two classes are (11) and (12) Finding the two closest points can be written as the following optimization problem: min CIt (13) 3S ee, for example, K. Bennett, 1997, in http://www.rpi.edu/bennek/svmtalk.ps (also, to appear). 247 A Geometric Interpretation of v-SVM Classifiers subject to: L ai L = 1, ai = 1, a t' > _ 0 (14) i:y;=-l i:y;=+l Taking the decision boundary j(x) = w? x + b = 0 to be the perpendicular bisector of the line segment joining the two closest points means that at the solution, (15) and b= -w? p, where (16) Thus w lies along the line segment (and is half its size) and p is the midpoint of the line segment. By rescaling the objective function and using the class labels Yi = ?1 we can rewrite this as 4 : (17) subject to (18) The associated decision function is j( x) = w . x + b where w p = ~ L:i aiXi and b = -w.p = L:ij aiYiajXi . Xj. -t 3.2 = ~ L:i aiYiXi, The Connection with v-SVM Consider now the two sets of points defined by: H+ JJ = { '. ~ I.y;-+l aiXil .. ~ I.y.-+l ai = 1, 0 ~ ai ~ fL} (19) and (20) We have the following simple proposition: Proposition 1: H+ JJ C H+ and H-JJ C H_, and H+ JJ and H-JJ are both convex sets. Furthermore, the positions of the points H+ JJ and H-JJ with respect to the Xi do not depend on the choice of origin. Proof: Clearly, since the ai defined in H+ JJ is a subset of the ai defined in H+, H+ JJ C H+, similarly for H_. Now consider two points in H+ JJ defined by aI, a2. Then all points on the line joining these two points can be written as L:i:y;=+l ((1A)ali + Aa2i)Xi, 0 ~ A ~ 1. Since ali and a2i both satisfy 0 ~ ai ~ fL, so does (1- A)ali +Aa2i, and since also L:i:y;=+l (1- A)ali+Aa2i = 1, the set H+ JJ is convex. 4That one can rescale the objective function without changing the constraints follows from uniqueness of the solution. See also Burges and Crisp, "Uniqueness of the SVM Solution" in this volume. 248 D. J. Crisp and C. J. C. Burges The argument for H_~ is similar. Finally, suppose that every Xi is translated by Xo, i.e. Xi -+ Xi + Xo 'Vi. Then since L:i:Yi=+l ai = 1, every point in H+~ is also translated by the same amount, similarly for H-w 0 The problem of finding the optimal separating hyperplane between the convex sets H+~ and H_~ then becomes: (21) subject to (22) Since Eqs. (21) and (22) are identical to (9) and (10), we see that the v-SVM algorithm is in fact finding the optimal separating hyperplane between the convex sets H+~ and H-w We note that the convex sets H+~ and H_~ are not simply uniformly scaled versions of H + and H _. An example is shown in Figure 1. xl xl xl 1'=5112 1'=113 ...... '! 1/3 5::: xl xl 113 :"::.~ .' xl 112 116 5112 x2 xl -lo:rrrTTT17TTT17~ --t----"I---+-----. xl 112 Figure 1: The soft convex hull for the vertices of a right isosceles triangle, for various 1'. Note how the shape changes as the set grows and is constrained by the boundaries of the encapsulating convex hull. For I' < ~, the set is empty. Below, we will refer to the formulation given in this section as the soft convex hull formulation, and the sets of points defined in Eqs. (19) and (20) as soft convex hulls. 3.3 Comparing the Offsets and Margin Widths The natural value of the offset bin the soft convex hull approach, b= -w . p, arose by asking that the separating hyperplane lie halfway between the closest extremities of the two soft convex hulls. Different choices of b just amount to hyperplanes with the same normal but at different perpendicular distances from the origin. This value of b will not in general be the same as that for which the cost term in Eq. (7) is minimized. We can compare the two values as follows. The KKT conditions for the J.'-SVM formulation are (I' - ai)~i ai(Yi(w?Xi+b)-p+~i) - 0 0 Multiplying (24) by Yi, summing over i and using (23) gives (23) (24) A Geometric Interpretation ofv-SVM Classifiers 249 (25) Thus the separating hyperplane found in the J.'-SVM algorithm sits a perpendicular distance 12ifiorr l:i Yi~i I away from that found in the soft convex hull formulation. For the given w, this choice of b results in the lowest value of the cost, J.' l:i ~i. The soft convex hull approach suggests taking p = w . w, since this is the value Iii takes at the points l:Yi=+l (XiXi and l:Yi=-l (XiXi. Again, we can use the KKT conditions to compare this with p. Summing (24) over i and using (23) gives p= p+ ~ L~i. (26) i Since p = 3.4 W? w, this again shows that if p = 0 then w = ~i = 0, and, by (25), b = O. The Primal for the Soft Convex Hull Formulation By substituting (25) and (26) into the J.'-SVM primal formulation (7) and (8) we obtain the primal formulation for the soft convex hull problem: minimize (27) with respect to w, b, p, ~i, subject to: Yi (W ? Xi + b-) 2:: p_ - ~i + J.' "~ 1 + 2YiYj ~j, (28) j It is straightforward to check that the dual is exactly (9) and (10). Moreover, by summing the relevant KKT conditions, as above, we see that b = -w?p and p = w?w. Note that in this formulation the variables ~i retain their meaning according to (8). 4 Choosing v In this section we establish some results on the choices for v, using the J.'-SVM formulation. First, note that l:i (XiYi = 0 and l:i (Xi = 2 implies l:i:Yi=+l (Xi = l:i:Yi=-l (Xi = 1. Then (Xi 2:: 0 gives (Xi ~ 1, Vi. Thus choosing J.' > 1, which corresponds to choosing v < 2/1, results in the same solution of the dual (and hence the same normal w) as choosing J.' = 1. (Note that different values of J.' > 1 can still result in different values of the other primal variables, e.g. b). The equalities l:i:Yi=+l (Xi = l:i:y;=-l (Xi = 1 also show that if J.' < 2/1 then the feasible region for the dual is empty and hence the problem is insoluble. This corresponds to the requirement v < 1. However, we can improve upon this. Let 1+ (L) be the number of positive (negative) polarity points, so that 1+ + L = I. Let lmin == min{I+,L}. Then the minimal value of J.' which still results in a nonempty feasible region is J.'min = 1/lmin. This gives the condition v ~ 2Imin /l. We define a "nontrivial" solution of the problem to be any solution with w =I o. The following proposition gives conditions for the existence of nontrivial solutions. D. J. Crisp and C. J. C. Burges 250 Proposition 2: A value of v exists which will result in a nontrivial solution to the v-SVM classification problem if and only if {H+I-' : I-' = I-'min} n {H_I-' : I-' = I-'min} = 0. Proof: Suppose that {H+I-' : I-' = I-'min} n {H_I-' : I-' = I-'min} =1= 0. Then for all allowable values of I-' (and hence v), the two convex hulls will intersect, since {H+I-' : I-' = I-'min} C {H+I-' : I-' ~ I-'min} and {H_I-' : I-' = I-'min} C {H_I-' : I-' ~ I-'min}. IT the two convex hulls intersect, then the solution is trivial, since by definition there then exist feasible points z such that z = Li:Yi=+lOiXi and z = Li:Yi=_lOiXi, and hence 2w = Li 0iYiXi = Li:Yi=+lOiXi - Li:Yi=-l 0iXi = 0 (cf. (21), (22). Now suppose that {H+I-' : I-' = I-'min} n {H_I-' : I-' = I-'min} = 0. Then clearly a nontrivial solution exists, since the shortest distance between the two convex sets {H+1-' : I-' = I-'min} and {H -I-' : I-' = I-'min} is not zero, hence the corresponding w =1= o. 0 Note that when 1+ = L, the condition amounts to the requirement that the centroid of the positive examples does not coincide with that of the negative examples. Note also that this shows that, given a data set, one can find a lower bound on v, by finding the largest I-' that satisfies H_I-' n H+I-' = 0. 5 Discussion The soft convex hull interpretation suggests that an appropriate way to penalize positive polarity errors differently from negative is to replace the sum I-' Li ~i in (7) with 1-'+ Li:Yi=+l ~i + 1-'- Li:Yi=-l ~i? In fact one can go further and introduce a I-' for every train point. The I-'-SVM formulation makes this possibility explicit, which it is not in original v-SVM formulation. Note also that the fact that v-SVM leads to values of b which differ from that which would place the optimal hyperplane halfway between the soft convex hulls suggests that there may be principled methods for choosing the best b for a given problem, other than that dictated by minimizing the sum of the ~i 'so Indeed, originally, the sum of ~i 's term arose in an attempt to approximate the number of errors on the train set [21. The above reasoning in a sense separates the justification for w from that for b. For example, given w, a simple line search could be used to find that value of b which actually does minimize the number of errors on the train set. Other methods (for example, minimizing the estimated Bayes error [3]) may also prove useful. Acknowledgments C. Burges wishes to thank W. Keasler, V. Lawrence and C. Nohl of Lucent Technologies for their support. References [1] B. Scholkopf and A. Smola and R. Williamson and P. Bartlett. New support vector algorithms, neurocolt2 nc2-tr-1998-031. Technical report, GMD First and Australian National University, 1998. [2] C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20:273-297, 1995. [3] C. J. C. Burges and B. SchOlkopf. Improving the accuracy and speed of support vector learning machines. In M. Mozer, M. Jordan, and T. Petsche, editors, Advances in Neural Information Processing Systems 9, pages 375-381, Cambridge, MA, 1997. MIT Press.
1687 |@word version:2 tr:1 substitution:1 comparing:1 com:1 rpi:1 xiyi:1 must:1 written:2 shape:1 remove:1 half:1 sits:1 hyperplanes:1 along:2 scholkopf:4 prove:2 introduce:1 indeed:1 nonseparable:1 becomes:1 moreover:2 notation:1 lowest:1 interpreted:1 finding:4 every:3 exactly:2 classifier:5 scaled:1 appear:3 positive:5 engineering:1 joining:3 extremity:1 au:1 suggests:3 p_:1 perpendicular:4 unique:1 acknowledgment:1 intersect:2 bell:1 cannot:1 lmin:2 crisp:6 equivalent:2 www:1 straightforward:1 go:1 convex:28 reparameterization:1 justification:1 deptartment:1 suppose:3 user:2 aixi:1 ixi:1 origin:3 wolfe:1 recognition:3 electrical:1 region:2 removed:1 principled:1 mozer:1 depend:1 rewrite:1 segment:3 ali:4 upon:1 ofv:2 triangle:1 translated:2 differently:1 jersey:1 various:1 train:3 choosing:5 whose:1 maximal:1 relevant:1 bennek:1 empty:3 p:1 requirement:2 measured:1 rescale:1 ij:2 eq:3 dividing:1 implies:1 australian:1 differ:1 yiyj:1 hull:22 australia:1 bin:1 behaviour:1 proposition:4 adjusted:1 normal:6 lawrence:1 mapping:1 substituting:1 adopt:1 a2:1 uniqueness:3 estimation:1 label:3 largest:1 bisects:1 mit:1 clearly:2 sensor:1 always:1 arose:2 check:1 centroid:1 sense:1 vl:1 dual:6 classification:1 denoted:1 constrained:1 identical:1 minimized:1 report:1 h_:5 national:1 attempt:1 freedom:1 interest:1 possibility:1 primal:7 necessary:2 minimal:1 soft:14 asking:1 cost:2 introducing:1 vertex:1 subset:2 retain:1 again:2 rescaling:1 li:8 coefficient:1 satisfy:1 vi:2 start:1 bayes:1 nc2:1 minimize:6 accuracy:1 maximized:1 vp:2 produced:2 multiplying:1 definition:1 associated:1 proof:2 positioned:1 actually:1 appears:1 originally:1 formulation:14 furthermore:2 just:2 smola:1 christopher:1 grows:1 hence:5 equality:1 laboratory:1 width:1 allowable:1 reasoning:1 meaning:1 consideration:1 recently:2 volume:2 interpretation:9 refer:2 cambridge:1 ai:12 insoluble:1 similarly:3 centre:1 add:1 closest:3 dictated:1 yi:19 maximize:1 redundant:1 shortest:2 signal:1 technical:1 controlled:2 variant:1 regression:1 basic:1 kernel:1 penalize:1 south:1 induced:1 subject:8 jordan:1 call:1 ee:1 iii:1 xj:2 idea:1 bartlett:1 penalty:1 jj:11 useful:1 clear:2 amount:3 ph:1 cit:1 gmd:1 http:1 exist:1 sign:1 estimated:1 changing:1 asymptotically:1 halfway:3 sum:4 fraction:2 place:1 separation:1 decision:5 holmdel:1 bound:3 fl:2 nontrivial:6 constraint:5 x2:1 speed:1 argument:1 reparameterize:1 min:15 separable:4 imin:1 according:1 describes:1 making:1 xo:2 turn:1 slack:1 nonempty:1 encapsulating:1 apply:1 away:1 appropriate:1 petsche:1 a2i:1 existence:3 original:2 cf:1 giving:1 establish:1 objective:4 quantity:1 neurocolt2:1 usual:1 distance:5 separate:1 mapped:1 separating:7 thank:1 extent:1 trivial:1 polarity:3 xixi:2 minimizing:2 negative:5 upper:1 reparameterized:1 extended:1 rn:1 david:1 introduced:1 connection:1 below:2 pattern:3 h_i:6 suitable:1 natural:1 advanced:1 improve:1 technology:3 nohl:1 geometric:8 degree:1 sufficient:2 editor:1 pi:4 lo:1 sym:1 burges:9 taking:2 midpoint:1 boundary:2 coincide:1 approximate:1 emphasize:1 global:1 kkt:3 summing:3 xi:17 search:1 improving:1 williamson:1 linearly:1 position:2 explicit:1 wish:1 xl:8 lie:2 lucent:3 offset:2 svm:33 cortes:1 exists:2 vapnik:1 adding:1 margin:6 intersection:1 simply:1 absorbed:1 scalar:2 corresponds:2 determines:1 satisfies:1 ma:1 iyixi:2 replace:2 bennett:1 feasible:5 change:1 determined:1 uniformly:1 hyperplane:11 called:1 bisector:1 support:7 adelaide:2
751
1,688
Graded grammaticality in Prediction Fractal Machines Shan Parfitt, Peter Tiilo and Georg Dorffner Austrian Research Institute for Artificial Intelligence, Schottengasse 3, A-IOIO Vienna, Austria. { shan,petert,georg} @ai. univie. ac. at Abstract We introduce a novel method of constructing language models, which avoids some of the problems associated with recurrent neural networks. The method of creating a Prediction Fractal Machine (PFM) [1] is briefly described and some experiments are presented which demonstrate the suitability of PFMs for language modeling. PFMs distinguish reliably between minimal pairs, and their behavior is consistent with the hypothesis [4] that wellformedness is 'graded' not absolute. A discussion of their potential to offer fresh insights into language acquisition and processing follows. 1 Introduction Cognitive linguistics has seen the development in recent years of two important, related trends. Firstly, a widespread renewal of interest in the statistical, 'graded' nature of language (e.g. [2]-[4]) is showing that the traditional all-or-nothing notion of well-formedness may not present an accurate picture of how the congruity of utterances is represented internally. Secondly, the analysis of state space trajectories in artificial neural networks (ANNs) has provided new insights into the types of processes which may account for the ability of learning devices to acquire and represent language, without appealing to traditional linguistic concepts [5]-[7]. Despite the remarkable advances which have come out of connectionist research (e.g. [8]), and the now common use of recurrent networks, and Simple Recurrent Networks (SRNs) [9] especially, in the study of language (e.g. [10]), recurrent neural networks suffer from particular problems which make them imperfectly suited to language tasks. The vast majority of work in this field employs small networks and datasets (usually artificial), and although many interesting linguistic issues may be thus tackled, real progress in evaluating the potentials of state trajectories and graded 'grammaticality' to uncover the underlying processes responsible for overt linguistic phenomena must inevitably be limited whilst the experimental tasks remain so small. Nevertheless, there are certain obstacles to the scaling-up of networks trained by back-propagation (BP). Such networks tend towards ever Graded Grammaticality in Prediction Fractal Machines 53 longer training times as the sizes of the input set and of the network increase, and although Real-Time Recurrent Learning (RTRL) and Back-propagation Through Time are potentially better at modeling temporal dependencies, training times are longer still [11]. Scaling-up is also difficult due to the potential for catastrophic interference and lack of adaptivity and stability [12]-[14]. Other problems include the rapid loss of information about past events as the distance from the present increases [15] and the dependence of learned state trajectories not only on the training data, but also upon such vagaries as initial weight vectors, making their analysis difficult [16]. Other types of learning device also suffer problems. Standard Markov models require the allocation of memory for every n-gram, such that large values of n are impractical; variable-length Markov models are more memory-efficient, but become unmanageable when trained on large data sets [17]. Two important, related concerns in cognitive linguistics are thus (a) to find a method which allows language models to be scaled up, which is similar in spirit to recurrent neural networks, but which does not encounter the same problems of scale, and (b) to use such a method to evince new insights into graded grammaticality from the state trajectories which arise given genuinely large, naturally-occurring data sets. Accordingly, we present a new method of generating state trajectories which avoids most of these problems. Previously studied in a financial prediction task, the method creates a fractal map of the training data, from which state machines are built. The resulting models are known as Prediction Fractal Machines (PFMs) [18] and have some useful properties. The state trajectories in the fractal representation are fast and computationally efficient to generate, and are accurate and well-understood; it may be inferred that, even for very large vocabularies and training sets, catastrophic interference and lack of adaptivity and stability will not be a problem, given the way in which representations are built (demonstrating this is a topic for future work); training times are significantly less than for recurrent networks (in the experiments described below, the smallest models took a few minutes to build, while the largest ones took only around three hours; in comparison, all of the ANNs took longer - up to a day - to train); and there is little or no loss of information over the course of an input sequence (allowing for the finite precision of the computer). The scalability of the PFM was taken advantage of by training on a large corpus of naturally-occurring text. This enabled an assessment of what potential new insights might arise from the use of this method in truly large-scale language tasks. 2 Prediction Fractal Machines (PFMs) A brief description of the method of creating a PFM will now be given. Interested readers should consult [1], since space constraints preclude a detailed examination here. The key idea behind our predictive model is a transformation F of symbol sequences from an alphabet (here, tagset) {I, 2, ... , N} into points in a hypercube H = [0, I]D. The dimensionality D of the hypercube H should be large enough for each symbol 1, 2, ... , N to be identified with a unique vertex of H. The particular assignment of symbols to vertices is arbitrary. The transformation F has the crucial property that symbol sequences sharing the same suffix (context) are mapped close to each other. Specifically, the longer the common suffix shared by two sequences, the smaller the (Euclidean) distance between their point representations. The transformation F used in this study corresponds to an Iterative Function System [19] S. Parfitt, P TIno and G. Dorffner 54 consisting of N affine maps i : H -+ H, i = 1,2, ... , N, i(x) = ~(x + ti), tj E {a, l}D, ti =F tj for i =F j. (1) Given a sequence 5182 ... 5L of L symbols from the alphabet 1,2, .. . , N, we construct its point representation as where x* is the center {l}D of the hypercube H. (Note that as is common in the Iterative Function Systems literature, i refers either to a symbol or to a map, depending upon the context.) PFMs are constructed on point representations of subsequences appearing in the training sequence. First, we slide the window of length L > 1 over the training sequence. At each position we transform the sequence of length L appearing in the window into a point. The set of points obtained by sliding through the whole training sequence is then partitioned into several classes by k-means vector quantization (in the Euclidean space), each class represented by a particular codebook vector. The number of code book vectors required is chosen experimentally. Since quantization classes group points lying close together, sequences having point representations in the same class (potentially) share long suffixes. The quantization classes may then be treated as prediction contexts, and the corresponding predictive symbol probabilities computed by sliding the window over the training sequence again and counting, for each quantization class, how often a sequence mapped to that class was followed by a particular symbol. In test mode, upon seeing a new sequence of L symbols, the transformation F is again performed, the closest quantization center found, and the corresponding predictive probabilities used to predict the next symbol. 3 An experimental comparison of PFMs and recurrent networks The performance of the PFM was compared against that of a RTRL-trained recurrent network on a next-tag prediction task. Sixteen grammatical tags and a 'sentence start' character were used. The models were trained on a concatenated sequence (22781 tags) of the top three-quarters of each of the 14 sub-corpora of the University of Pennsylvania 'Brown' corpus 1 . The remainder was used to create test data, as follows. Because in a large training corpus of naturally-occurring data, contexts in most cases have more than one possible correct continuation, simply counting correctly predicted symbols is insufficient to assess performance, since this fails to count correct responses which are not targets. The extent to which the models distinguished between grammatical and ungrammatical utterances was therefore additionally measured by generating minimal pairs and comparing their negative log likelihoods (NLLs) per symbol with respect to the model. Likelihood is computed by sliding through the test sequence and for each window position, determining the probability of the symbol that appears immediately beyond it. As processing progresses, these probabilities are multiplied. The negative of the natural logarithm is then taken and divided by the number of symbols. Significant differences in NLLs Ihttp://www.ldc.upenn.edu/ Graded Grammaticality in Prediction Fractal Machines 55 are much harder to achieve between members of minimal pairs than between grammatical and random sequences, and are therefore a good measure of model validity. Minimal pairs generated by theoretically-motivated manipulations tend to be no longer ungrammatical given a small tagset, because the removal of grammatical sub-classes necessarily also removes a large amount of information. Manipulations were therefore performed by switching the positions of two symbols in each sentence in the test sets. Symbols switched could be any distance apart within the sentence, as long as the resulting sentence was ungrammatical under all surface instantiations. By changing as little as possible to make the sentence ungrammatical, the goal was retained that the task of distinguishing between grammatical and ungrammatical sequences be as difficult as possible. The test data then consisted of 28 paired grammatical/ungrammatical test sets (around 570 tags each), plus an ungrammatical, 'meaningless' test set containing all 17 codes listed several times over, used to measure baseline performance. Ten 1st-order randomly-initialised networks were trained for 100 epochs using RTRL. The networks consisted of 1 input and 1 output layer, each with 17 units corresponding to the 17 tags, 2 hidden layers, each with 10 units, and 1 context layer of 10 units connected to the first hidden layer. The second hidden layer was used to increase the flexibility of the maps between the hidden representations in the recurrent portion and the tag activations at the output layer. A logistic sigmoid activation function was used, the learning rate and momentum were set to 0.05, and the training sequence was presented at the rate of one tag per clock tick. The PFMs were derived by clustering the fractal representation of the training data ten times for various numbers of codebook vectors between 5 and 200. More experiments were performed using PFMs than neural networks because in the former case, experience in choosing appropriate numbers of codebook vectors was initially lacking for this type of data. The results which follow are given as averages, either over all neural networks, or else over all PFMs derived from a given number of codebook vectors. The networks correctly predicted 36.789% and 32.667% of next tags in the grammatical and ungrammatical test sets, respectively. The PFMs matched this performance at around 30 codebook vectors (37 .134% and 32.814% respectively), and exceeded it for higher numbers of vectors (39.515% and 34.388% respectively at 200 vectors). The networks generated mean NLLs per symbol of 1.966 and 2.182 for the grammatical and ungrammatical test sets, respectively (a difference of 0.216) and 4.157 for the 'meaningless' test set (the difference between NLLs for grammatical and 'meaningless' data = 2.191). The PFMs matched this difference in NLLs at 40 codebook vectors (NLL grammatical = 1.999, NLL ungrammatical = 2.217; difference = 0.218). The NLL for the 'meaningless' data at 40 codebook vectors was 6.075 (difference between NLLs for grammatical and 'meaningless' data = 4.076). The difference between NLLs for grammatical and ungrammatical, and for grammatical and 'meaningless' data sets, became even larger with increased numbers of codebook vectors. The difference in performance between grammatical and ungrammatical test sets was thus highly significant in all cases (p < .0005): all the models distinguished what was grammatical from what was not. This conclusion is supported by the fact that the mean, NLLs for the 'meaningless' test set were always noticeably higher than those for the minimal pair sets. S. Parfitt, P. Tina and G. Daiffner 56 4 Discussion The PFMs exceeded the performance of the networks for larger numbers of codebook vectors, but it is possible that networks with more hidden nodes would also do better. In terms of ease of use, however, as well as in their scaling-up potential, PFMs are certainly superior. Their other great advantage is that the representations created are dependable (see section 1), making hypothesis creation and testing not just more rapid, but also more straightforward: the speed with which PFMs may be trained made it possible to make statistically significant observations for a large number of clustering runs. In the introduction, 'graded' wellformedness was spoken of as being productive of new hypotheses about the nature of language. Our use of minimal pairs, designed to make a clear-cut distinction between grammatical and ungrammatical utterances, appears to leave this issue to one side. But in reality, our results were rather pertinent to it, as the use of the likelihood measure might indeed imply. The Brown corpus consists of subcorpora representative of 14 different discourse types, from fiction to government documents. Whereas traditional notions of grammaticality would lead us to treat all of the 'ungrammatical' sentences in the minimal pair test sets as equally ungrammatical, the NLLs in our experiments tell a different story. The grammatical versions consistently had a lower associated NLL (higher probability) than the ungrammatical versions, but the difference between these was much smaller than that between the 'meaningless' data and either the grammatical or the ungrammatical data. This supports the concept of 'graded grammaticality', and NLLs for 'meaningless' data such as ours might be seen as a sort of benchmark by which to measure all lesser degrees ofungrammaticality. (Note incidentally that the PFMs appear to associate with the 'meaningless' data a significantly higher NLL than did the networks, even though the difference between the NLLs of the grammatical and ungrammatical data was the same. This is suggestive of PFMs having greater powers of discrimination between grades of wellformedness than the recurrent networks used, but further research will be needed to ascertain the validity of this.) Moreover, the NLL varied not just between grammatical and ungrammatical test sets, but also from sentence to sentence, from word to word and from discourse style to discourse style. While it increased, often dramatically, when the manipulated portion of an ungrammatical sentence was encountered, some words in grammatical sentences exhibited a similar effect: thus, if a subsequence in a well-formed utterance occurs only rarely - or never - in a training set, it will have a high associated NLL in the same way as an ungrammatical one does. This is likely to happen even for very large corpora, since some grammatical structures are very rare. This is consistent with recent findings that, during human sentence processing, well-formedness is linked to conformity with expectation [20] as measured by CLOZE scores. Interesting also was the remarkable variation in NLL between discourse styles. Although the mean NLL across all discourse styles (test sets) is lower for the grammatical than for the ungrammatical versions, it cannot be guaranteed that the grammatical version of one test set will have a lower NLL than the ungrammatical version of another. Indeed, the grammatical and ungrammatical NLLs interleave, as may be observed in figure 1, which shows the NLLs for the three discourse styles which lie at the bottom, middle and top of the range. Even more interestingly, if the NLLs for the grammatical versions of all discourse styles are ordered according to where they lie within this range, it becomes clear that NLL is a predictor of discourse style. Styles which linguists class as 'formal', e.g. those of 57 Graded Grammaticality in Prediction Fractal Machines NU,s associaIed wi1h grammatical and ungrammalical versions of 3 discourse types 3r----------r~~------~--------~------~--~ Leamed text: grammatical - leamed text ungrammalical ....... Romantic: fiction: grammatical .?. ... Romantfc fiction: ungrammalical -+2.8 Science fiction: grammatical ? .? .? Science fiction: ungrammatical -D .? .. ....... 2.8 ae G. :::I z : 1:).. 2.4 \. _.-a ... ? ..?t ? .. ? q ?.. B:Ie...... _a?- ..... ... __ .. _ ..- - -...- __----'" .--+-'---- ._ ..----...- . ...... _....-.-.-.- -- t::::~~~::-::-:-?-=~?? ~ --"'- .......' 2.2 \'--_.. ___ --?? _.-? ? 2 ~~~~~~:::.:~~~:~~~~-::~~:?.:;:::-:::::.::::.:.:;:::.:;:::.:;;::::;;;:::;:::.::~::"::,. ~ 1.8 o 50 100 No. of codebook vectors 150 200 Figure 1: NLLs of minimal pair test sets containing different discourse styles suggest grades of wellformedness based upon prototypicality. the Learned and Government Document test sets, have the lowest NLLs, with the three Press test sets clustering just above, and the Fiction test sets, exemplifying creative language use, clustering at the high end. Similarly, that the Learned and Government test sets have the lowest NLLs conforms with the intuition that their usage lies closest to what is grammatically 'prototypical ' - even though in the training set, 6 out of the 14 test sets are fiction and thus might be expected to contribute more to the prototype. That they do not, suggests that usage varies significantly across fiction test sets. 5 Conclusion Work on the use of PFMs in language modeling is at an early stage, but as results to date show, they have a lot to offer. A much larger project is planned, which will examine further Allen and Seidenberg's hypothesis that 'graded grammaticality' (or wellformedness) applies not only to syntax, but also to other language subdomains such as semantics, an integral part of this being the use of larger corpora and tagsets, and the identification of vertices with semantic/syntactic features rather than atomic symbols. Identifying the possibilities of combining PFMs with ANNs, for example as a means of bypassing the normal method of creating state-space trajectories, is the subject of current study. Acknowledgments This work was supported by the Austrian Science Fund (FWF) within the research project "Adaptive Information Systems and Modeling in Economics and Management Science" (SFB 010). The Austrian Research Institute for Artificial Intelligence is supported by the Austrian Federal Ministry of Science and Transport. 58 S. Parfitt, P Tino and G. DorjJner References [1] P. Tino & G. Dorffner (1998). Constructing finite-context sources from fractal representations of symbolic sequences. Technical Report TR-98-18, Austrian Research Institute for AI, Vienna. [2] J. R. Taylor (1995). Linguistic categorisation: Prototypes in linguistic theory. Clarendon, Oxford. [3] J. R. Saffran, R. N. Aslin & E. L. Newport (1996) . Statistical cues in language acquisition: Word segmentation by infants. In Proc. of the Cognitive Science Society Conference, 376-380, La Jolla, CA. [4] J. Allen & M. S. Seidenberg (in press). The emergence of grammaticality in connectionist networks. In B. Macwhinney (ed.), Emergentist approaches to language: Proc . of the 28th Carnegie Symposium on cognition. Erlbaum. [5] S. Parfitt (1997). Aspects of anaphora resolution in artificial neural networks: Implications for nativism. PhD thesis, Imperial College, London. [6] D. Servan-Schreiber et al (1989). Graded state machines: The representation of temporal contingencies in Simple Recurrent Networks. In Advances in Neural Information Processing Systems, 643-652. [7] W. Tabor & M. Tanenhaus (to appear). Dynamical models of sentence processing. Cognitive Science. [8] J. L. Elman et al (1996). Rethinking innateness: A connectionist perspective on development. Bradford. [9] J . L. Elman (1990). Finding structure in time. In: Cognitive Science, 14: 179-211. [10] S. Lawrence, C. Lee Giles & S. Fong (in press). Natural language grammatical inference with recurrent neural networks. IEEE Trans. on knowledge and data engineering. [11] J. Hertz, A. Krogh & R. G. Palmer (1991) . Introduction to the theory of neural computation. Addison Wesley. [12] M. McCloskey & N. J. Cohen (1989). Catastrophic interference in connectionist networks: The sequential learning problem. In G. Bower (ed.), The psychology of learning and motivation, vol 24. Academic, NY. [13] J . K. Kruschke (1991) . ALCOVE: A connectionist model of human category learning. In R. P. Lippman et al (eds.), Advances in Neural Information Processing 9, 649-655. Kaufmann, San Mateo, CA. [14] S. Grossberg (ed.) (1988). Neural networks and natural intelligence. Bradford, MIT, Cambs, MA. [15] Y. Bengio, P. Simard & P. Frasconi (1994). Learning long-term dependencies with gradient descent is difficult. IEEE Trans. on neural networks, 5(2). [16] M. P. Casey (1996). The dynamics of discrete-time computation, with application to recurrent neural networks and finite-state machine extraction. Neural Computation, 8(6):1135-1178. [17] D. Ron, Y. Singer & N. Tishby (1996). The power of amnesia. Machine Learning, 25. [18] P. Tino, B. G. Horne, C. Lee Giles & P. C. Collingwood (1998). Finite state machines and recurrent neural networks - automata and dynamical systems approaches. In J. E. Dayhoff & O. Omidvar (eds.), Neural Networks and Pattern Recognition, 171- 220. Academic. [19] M. F. Barnsley (1988). Fractals everywhere. Academic, NY. [20] S. Coulson, J. W. King & M. Kutas (1998). Expect the unexpected: Responses to morphosyntactic violations. Language and Cognitive Processes, 13(1).
1688 |@word version:7 middle:1 briefly:1 interleave:1 tr:1 harder:1 omidvar:1 initial:1 score:1 document:2 ours:1 interestingly:1 past:1 current:1 comparing:1 activation:2 must:1 happen:1 pertinent:1 remove:1 designed:1 fund:1 discrimination:1 infant:1 intelligence:3 cue:1 device:2 accordingly:1 leamed:2 codebook:10 node:1 contribute:1 ron:1 firstly:1 constructed:1 become:1 symposium:1 amnesia:1 consists:1 dayhoff:1 introduce:1 theoretically:1 upenn:1 expected:1 indeed:2 rapid:2 elman:2 examine:1 behavior:1 grade:2 little:2 preclude:1 window:4 becomes:1 provided:1 project:2 underlying:1 matched:2 moreover:1 horne:1 lowest:2 what:4 whilst:1 spoken:1 finding:2 transformation:4 impractical:1 temporal:2 every:1 ti:2 scaled:1 unit:3 internally:1 appear:2 understood:1 engineering:1 treat:1 switching:1 despite:1 oxford:1 might:4 plus:1 studied:1 mateo:1 suggests:1 ease:1 limited:1 palmer:1 range:2 statistically:1 grossberg:1 unique:1 responsible:1 acknowledgment:1 testing:1 atomic:1 lippman:1 innateness:1 significantly:3 word:4 refers:1 seeing:1 suggest:1 symbolic:1 cannot:1 close:2 context:6 www:1 map:4 center:2 straightforward:1 economics:1 kruschke:1 automaton:1 resolution:1 identifying:1 immediately:1 insight:4 financial:1 enabled:1 stability:2 notion:2 variation:1 target:1 distinguishing:1 hypothesis:4 associate:1 trend:1 recognition:1 genuinely:1 cut:1 observed:1 bottom:1 connected:1 intuition:1 productive:1 dynamic:1 trained:6 predictive:3 creation:1 upon:4 creates:1 represented:2 various:1 alphabet:2 train:1 fast:1 london:1 artificial:5 tell:1 choosing:1 larger:4 ability:1 syntactic:1 transform:1 emergence:1 nll:11 sequence:19 advantage:2 took:3 remainder:1 combining:1 date:1 flexibility:1 achieve:1 description:1 scalability:1 generating:2 incidentally:1 leave:1 depending:1 recurrent:15 ac:1 measured:2 progress:2 krogh:1 predicted:2 come:1 correct:2 dependable:1 human:2 newport:1 noticeably:1 require:1 government:3 suitability:1 formedness:2 secondly:1 bypassing:1 lying:1 around:3 normal:1 great:1 lawrence:1 cognition:1 predict:1 early:1 smallest:1 proc:2 overt:1 largest:1 schreiber:1 create:1 alcove:1 federal:1 mit:1 always:1 rather:2 linguistic:5 derived:2 tabor:1 casey:1 consistently:1 likelihood:3 baseline:1 inference:1 suffix:3 initially:1 hidden:5 interested:1 semantics:1 issue:2 development:2 renewal:1 field:1 construct:1 never:1 having:2 frasconi:1 extraction:1 anaphora:1 future:1 connectionist:5 report:1 aslin:1 employ:1 few:1 randomly:1 manipulated:1 consisting:1 prototypicality:1 interest:1 highly:1 possibility:1 certainly:1 violation:1 truly:1 behind:1 tj:2 implication:1 accurate:2 integral:1 experience:1 conforms:1 euclidean:2 logarithm:1 taylor:1 minimal:8 increased:2 modeling:4 obstacle:1 giles:2 planned:1 servan:1 assignment:1 vertex:3 rare:1 imperfectly:1 predictor:1 erlbaum:1 tishby:1 dependency:2 varies:1 fong:1 st:1 ie:1 lee:2 together:1 again:2 thesis:1 management:1 containing:2 cognitive:6 creating:3 book:1 simard:1 style:9 account:1 potential:5 grammaticality:10 performed:3 lot:1 linked:1 portion:2 start:1 sort:1 ass:1 formed:1 became:1 kaufmann:1 identification:1 trajectory:7 saffran:1 anns:3 sharing:1 ed:5 against:1 acquisition:2 initialised:1 naturally:3 associated:3 austria:1 knowledge:1 dimensionality:1 segmentation:1 uncover:1 back:2 appears:2 exceeded:2 clarendon:1 higher:4 wesley:1 day:1 follow:1 response:2 though:2 just:3 stage:1 clock:1 transport:1 assessment:1 lack:2 propagation:2 widespread:1 mode:1 logistic:1 usage:2 effect:1 validity:2 concept:2 brown:2 consisted:2 former:1 semantic:1 vagary:1 tino:4 during:1 syntax:1 demonstrate:1 allen:2 novel:1 common:3 sigmoid:1 superior:1 quarter:1 cohen:1 significant:3 ai:2 tanenhaus:1 similarly:1 language:17 had:1 longer:5 surface:1 closest:2 recent:2 perspective:1 jolla:1 apart:1 manipulation:2 certain:1 seen:2 ministry:1 greater:1 pfm:4 sliding:3 technical:1 academic:3 offer:2 long:3 divided:1 equally:1 paired:1 prediction:10 austrian:5 ae:1 expectation:1 represent:1 whereas:1 else:1 source:1 crucial:1 meaningless:10 exhibited:1 subject:1 tend:2 member:1 spirit:1 grammatically:1 fwf:1 consult:1 counting:2 pfms:18 enough:1 bengio:1 psychology:1 pennsylvania:1 identified:1 idea:1 lesser:1 prototype:2 motivated:1 sfb:1 suffer:2 peter:1 dorffner:3 schottengasse:1 linguist:1 fractal:12 dramatically:1 useful:1 detailed:1 listed:1 clear:2 amount:1 slide:1 ten:2 category:1 generate:1 continuation:1 fiction:8 correctly:2 per:3 carnegie:1 discrete:1 vol:1 georg:2 group:1 key:1 nevertheless:1 demonstrating:1 imperial:1 changing:1 vast:1 year:1 run:1 everywhere:1 reader:1 scaling:3 layer:6 shan:2 followed:1 distinguish:1 tackled:1 guaranteed:1 encountered:1 constraint:1 categorisation:1 bp:1 exemplifying:1 tag:8 aspect:1 speed:1 macwhinney:1 according:1 creative:1 hertz:1 remain:1 smaller:2 ascertain:1 rtrl:3 character:1 partitioned:1 appealing:1 across:2 making:2 interference:3 taken:2 computationally:1 previously:1 count:1 needed:1 addison:1 singer:1 end:1 multiplied:1 appropriate:1 appearing:2 distinguished:2 encounter:1 subdomains:1 top:2 clustering:4 include:1 linguistics:2 tina:1 vienna:2 concatenated:1 especially:1 graded:12 build:1 hypercube:3 society:1 occurs:1 dependence:1 traditional:3 gradient:1 distance:3 mapped:2 rethinking:1 majority:1 conformity:1 topic:1 extent:1 fresh:1 length:3 code:2 retained:1 insufficient:1 acquire:1 difficult:4 potentially:2 negative:2 reliably:1 allowing:1 observation:1 datasets:1 markov:2 benchmark:1 finite:4 descent:1 inevitably:1 nlls:17 ever:1 varied:1 arbitrary:1 inferred:1 pair:8 required:1 sentence:12 learned:3 distinction:1 hour:1 nu:1 trans:2 beyond:1 usually:1 univie:1 below:1 dynamical:2 pattern:1 srns:1 built:2 memory:2 ldc:1 power:2 event:1 treated:1 examination:1 natural:3 brief:1 imply:1 picture:1 created:1 utterance:4 text:3 epoch:1 literature:1 removal:1 determining:1 lacking:1 loss:2 expect:1 adaptivity:2 interesting:2 prototypical:1 allocation:1 remarkable:2 sixteen:1 switched:1 contingency:1 degree:1 affine:1 consistent:2 story:1 share:1 romantic:1 course:1 supported:3 tick:1 side:1 formal:1 institute:3 unmanageable:1 absolute:1 grammatical:31 vocabulary:1 evaluating:1 avoids:2 gram:1 made:1 adaptive:1 san:1 suggestive:1 instantiation:1 corpus:7 subsequence:2 iterative:2 seidenberg:2 reality:1 additionally:1 barnsley:1 nature:2 ca:2 ungrammatical:25 necessarily:1 constructing:2 did:1 cloze:1 whole:1 motivation:1 arise:2 nothing:1 wellformedness:5 representative:1 ny:2 precision:1 sub:2 position:3 fails:1 momentum:1 lie:3 bower:1 evince:1 minute:1 showing:1 symbol:18 concern:1 quantization:5 sequential:1 phd:1 occurring:3 suited:1 simply:1 likely:1 tagsets:1 unexpected:1 ordered:1 mccloskey:1 applies:1 corresponds:1 discourse:10 ma:1 goal:1 king:1 towards:1 shared:1 experimentally:1 specifically:1 bradford:2 catastrophic:3 experimental:2 la:1 rarely:1 college:1 support:1 tagset:2 phenomenon:1
752
1,689
Some Theoretical Results Concerning the Convergence of Compositions of Regularized Linear Functions Tong Zhang Mathematical Sciences Department IBM T.1. Watson Research Center Yorktown Heights, NY 10598 tzhang@watson.ibm.com Abstract Recently, sample complexity bounds have been derived for problems involving linear functions such as neural networks and support vector machines. In this paper, we extend some theoretical results in this area by deriving dimensional independent covering number bounds for regularized linear functions under certain regularization conditions. We show that such bounds lead to a class of new methods for training linear classifiers with similar theoretical advantages of the support vector machine. Furthermore, we also present a theoretical analysis for these new methods from the asymptotic statistical point of view. This technique provides better description for large sample behaviors of these algorithms. 1 Introduction In this paper, we are interested in the generalization performance of linear classifiers obtained from certain algorithms. From computational learning theory point of view, such performance measurements, or sample complexity bounds, can be described by a quantity called covering number [11, 15, 17], which measures the size of a parametric function family. For two-class classification problem, the covering number can be bounded by a combinatorial quantity called VC-dimension [12, 17]. Following this work, researchers have found other combinatorial quantities (dimensions) useful for bounding the covering numbers. Consequently, the concept of VC-dimension has been generalized to deal with more general problems, for example in [15, 11]. Recently, Vapnik introduced the concept of support vector machine [16] which has been successful applied to many real problems. This method achieves good generalization by restricting the 2-norm of the weights of a separating hyperplane. A similar technique has been investigated by Bartlett [3], where the author studied the performance of neural networks when the I-norm of the weights is bounded. The same idea has also been applied in [13] to explain the effectiveness of the boosting algorithm. In this paper, we will extend their results and emphasize the importance of dimension independence. Specifically, we consider the following form of regularization method (with an emphasis on classification problems) which has been widely studied for regression problems both in statistics and in Convergence of Regularized Linear Functions 371 numerical mathematics: inf Ex yL(w, 2:, y) = inf Ex yl(wT 2:Y) w W I I + Ag(W), (1) where Ex ,y is the expectation over a distribution of (2:, y), and y E {-1, 1} is the binary label of data vector 2:. To apply this fonnulation for the purpose oftraining linear classifiers. we can choose I as a decreasing function, such that I (.) ~ 0, and choose 9 (w) ~ 0 as a function that penalizes large w (liIl1w~oo g( w) -4 00). A is an appropriately chosen positive parameter to balance the two tenns. The paper is organized as follows. In Section 2, we briefly review the concept of covering numbers as well as the main results related to analyzing the perfonnance of learning algorithms. In Section 3, we introduce the regularization idea. Our main goal is to construct regularization conditions so that dimension independent bounds on covering numbers can be obtained. Section 4 extends results from the previous section to nonlinear compositions of linear functions. In Section 5. we give an asymptotic fonnula for the generalization perfonnance of a learning algorithm, which will then be used to analyze an instance of SVM. Due to the space limitation, we will only present the main results and discuss their implications. The detailed derivations can be found in [18]. 2 Covering numbers We fonnulate the learning problem as to find a parameter from random observations to minimize risk: given a loss function L( a, x) and n observations Xl = {x 1, ... , x n } independently drawn from a fixed but unknown distribution D, we want to find a that minimizes the expected loss over 2: (risk): R(a) = ExL(a,x)= / L(a,x)dP(x). (2) The most natural method for solving (2) using a limited number of observations is by the empirical risk minimization (ERM) method (cf [15, 16]). We simply choose a parameter a that minimizes the observed risk: 1 n R(a,X l ) = - LL(a,xi). (3) n i=l We denote the parameter obtained in this way as a erm (Xl)' The convergence behavior of this method can be analyzed by using the VC theoretical point of view. which relies on the unifonn convergence of the empirical risk (the unifonn law of large numbers): SUPa IR(a, Xl) - R(a)l. Such a bound can be obtained from quantities that measure the size of a Glivenko-Cantelli class. For finite number of indices, the family size can be measured simply by its cardinality. For general function families, a well known quantity to measure the degree ofunifonn convergence is the covering number which can be be dated back to Kolmogrov [8, 9]. The idea is to discretize (which can depend on the data Xl) the parameter space into N values a1, . .. ,aN SO that each L(a, .) can be approximated by L( ai, .) for some i. We shall only describe a simplified version relevant for our purposes. Definition 2.1 Let B be a metric space with metric p. Given a norm p, observations Xl = [Xl, ... ,xn ]. and vectors I(a, Xl) = [/(a, Xl)"" ,/(a, x n )] E Bn parameterized by a, the covering number in p-norm, denoted as Np (I, ?, Xl)' is the minimum number of a collection o/vectors V1, ... ,Vm E B n such that Va. 3Vi: IIp(l(a,Xl),vi)lIp ::; n 1/P ?. We also denote Np(l, ?, n) = maxx~ Np(l, ?, Xl). Note that from the definition and the Jensen's inequality, we have N p ::; N q for p ::; q. We will always assume the metric on R to be IX1 - x21 if not explicitly specified otherwise. The following theorem is due to Pollard [11]: T. Zhang 372 Theorem 2.1 ([11]) \;/n, f > ? and distribution D. P(s~p IR(a, X~) - R(a)1 > ?j -nf 2 ~ 8E(Af1 (L , f/8, X~)] exp( 128M2)' where M = sUPa,:z: L(a, x) - infa,:z: L(a, x). and X~ = {Xl, . .. ,X' l } are independently drawn from D. The constants in the above theorem can be improved for certain problems; see [4. 6, 15, 16] for related results. However, they yield very similar bounds. The result most relevant for this paper is a lemma in [3] where the 1-nonn covering number is replaced by the oo-nonn covering number. The latter can be bounded by a scale-sensitive combinatorial dimension [1], which can be bounded from the I-norm covering number if this covering number does not depend on n. These results can replace Theorem 2.1 to yield better estimates under certain circumstances. Since Bartlett's lemma in [3] is only for binary loss functions, we shall give a generalization so that it is comparable to Theorem 2.1 : Theorem 2.2 Let It and 12 be two functions: R n -+ [0, 1] such that /Y1 - Y21 ~ I implies ~ h(Y2) ~ h(Y1) where h : R n -+ [0,1] is a reference separatingfunction, then It (Y1) P[s~p[E:z:It(L(a, -nf 2 x?) - Ex-;-h(L(a, x))] > f] ~ 4E[Afoo(L, I, X~)] exp( 32)' Note that in the extreme case that some choice of a achieves perfect generalization: E:z:h(L(a, x)) 0, and assume that our choices of a(X1) always satisfy the condition EXf h(L( a, x? = 0, then better bounds can be obtained by using a refined version of the Chernoffbound. = 3 Covering number bounds for linear systems In this section, we present a few new bounds on covering numbers for the following form of real valued loss functions: d L(w, x) = xT w= L XiWi ? (4) i=l As we shall see later, these bounds are relevant to the convergence properties of (1). Note that in order to apply Theorem 2.1, since Afl < Af2 , therefore it is sufficient to estimate Af2(L, ?, n) for ? > O. It is clear that Af2(L, f, ~ is not finite ifno restrictions on x and w are imposed. Therefore in the following, we will assume that each I/xil/p is bounded. and study conditions ofllw// q so that logAf(j, f, n) is independent or weakly dependent of d. Our first result generalizes a theorem of Bartlett [3]. The original results is with p = 00 and q 1, and the related technique has also appeared in [10, 13]. The proof uses a lemma that is attributed to Maurey (cf. [2, 7]). = Theorem 3.1 V/lxi/lp ~ band Ilw/lq ~ a, where lip + 1/q == 1 and 2 ~ p ~ log2 Af2(L, f, n) ~ 00, then r7a b 1Iog (2d + 1). 2 2 2 The above bound on the covering number depends logarithmically on d, which is already quite weak (as compared to linear dependency on d in the standard situation). However, the bound in Theorem 3.1 is nottightforp < 00. For example, the following theorem improves the above bound for p = 2. Our technique of proof relies on the SVD decomposition [5] for matrices, which improves a similar result in [14 J by a logarithmic factor. Convergence of Regularized Linear Functions 373 The next theorem shows that if lip + llq mdependent of dimension. = Theorem 3.3 Let L(w, x) xTw. J = lip + 1jq - 1 > 0, then > 1, then the 2-nonn covering number is also {f'llxillp :::; band Ilwllq :::; a, where 1 :::; q :::; 2 and One consequence of this theorem is a potentially refined explanation for the boosting algorithm. In [13], the boosting algorithm has been analyzed by using a technique related to results in [3] which essentially rely on Theorem 3.1 withp = 00. Unfortunately, the bound contains a logarithmic dependency on d (in the most general case) which does not seem to fully explain the fact that in many cases the perfonnance of the boosting algorithm keeps improving as d increases. However, this seemingly mysterious behavior might be better understood from Theorem 3.3 under the assumption that the data is more restricted than simply being oo-nonn bounded. For example, when the contribution of the wrong predictions is bounded by a constant (or grow very slowly as d increases), then we can regard its p-th nonn bounded for some p < 00 . In this case, Theorem 3.3 implies dimensional independent generalization. If we want to apply Theorem 2.2, then it is necessary to obtain bounds for infinity-nonn covering numbers. The following theorem gives such bounds by using a result from online learning. Theorem 3.4 lfllxillp :::; band Ilwllq :::; a, where 2 :::; p < 00 and lip + 11q = 1, then tiE> O. In the case of p = 00, an entropy condition can be used to obtain dimensional independent covering number bounds. = = Definition 3.1 Let f1. [f1.i] be a vector with positive entries such that 11f1.lll 1 (in this case, we call f1. a distribution vector). Let x = [Xi] "# 0 be a vector of the same length, then we define the weighted relative entropy of x with re5pect to f1. as: entro~(x) ~ IXil = ~ IXil ln J-Lillxlh' ? Theorem 3.5 Given a distribution vector f1., If llxi lloo :::; band Ilwlll :::; :::; c, where we assume that w has non-negative entries, then tiE> 0, entro ~ (w) log2 Noo(L, E, n) :::; a and 36b 2 ( a 2 + ac) E2 log2[2 r4ab/ E+ 21n + 1] . Theorems in this section can be combined with Theorem 4.1 to fonn more complex covering number bounds for nonlinear compositions oflinear functions. 374 4 T. Zhang Nonlinear extensions Consider the following system: + wTh(a, x)) , L([a, w], x) = I(g(a, x) (5) where x is the observation, and [a, w] is the parameter. We assume that function with bounded total variation. 1 is a nonlinear Definition 4.1 A/unction 1 : R -+ R is said to satisfy the Lipschitz condition with parameter"Y ifVx, y: I/( x) - I(y) I ~ )'Ix - yl? Definition 4.2 The total variation of a/unction 1 : R -+ R is defined as L TV(f, x) = sup :2:0<X1 L I/(xi) - I(xi-dl ? ' <Xl~X t=l We also denote TV(f, (0) as TV(f). Theorem 4.1 .if L([a, w], x) = I(g(a, x) + w T h(a, x)), where TV(f) < 00 and 1 is Lipschitz with parameter),. Assume also that w is a d-dimensional vector and Ilwllq :s; c, then VEl, E2 > 0, and n > 2(d + 1): Iog 2 Nr (L, E1 + E2, n) < (d + 1) log2[d en max(l TV(f) J, 1)] + log2 Nr([g , h], E2h, n) , - +1 where the metric o/[g, h) is defined as Ig1 - g21 2E1 + cllh1 - h211p (l/p + l/q = 1). Example 4.1 Consider classification by hyperplane: L( w, x) = J( w T x < 0) where J is the set indicator function. Let L' (w, x) = 10 (w T x) be another loss function where 1 z<0 lo(z) = { 1 - z z E [0 , 1] . o z>1 Instead of using ERM for estimating parameter that minimizes the risk of L , consider the scheme of minimize empirical risk associated with L', under the assumption that II x 112 :s; b and constraint that JJwl12 :s; a. Denote the estimated parameter by w n . It follows from the covering number bounds and Theorem 2.1 that with probability of at least 1 - 1]: n 1 / 2 ab In( nab + 2) + In 1.. _ _ _ _ _ _ _ _--'-'7 ). n If we apply a slight generalization of Theorem 2.2 and the covering number bound of Theorem 3.4, then with probability of at least 1 - T/: ExJ(w~ x ~ 0) :s; EXfJ(w~ x :s; 2)') + O( 1 a 2 b2 - ( - 2 In(abh + 2) n )' 1 + In n + In -)) T/ for all)' E (0,1]. 0 Bounds given in this paper can be applied to show that under appropriate regularization conditions and assumptions on the data, methods based on (1) lead to generalization performances of the form 0(1/ .jn), where 0 symbol (which is independent of d) is used to indicate that the hidden constant may include a polynomial dependency on Iog( n). It is also important to note that in certain cases, ,\ will not appear (or it has a small influence on the convergence) in the constant of 0, as being demonstrated by the example in the next section. 375 Convergence of Regularized Linear Functions 5 Asymptotic analysis The convergence results in the previous sections are in the form of VC style convergence in probability, which has a combinatorial flavor. However, for problems with differentiable function families involving vector parameters, it is often convenient to derive precise asymptotic results using the differential structure. Assume that the parameter a E Rm in (2) is a vector and L is a smooth function. Let a* denote the optimal parameter; "\1 ex denote the derivative with respect to a; and 'It( a, x) denote "\1 exL(a, x) . Assume that V = U= J J "\1 ex'lt(a* , x) dP(x) 'It ( a * , x) 'It ( a * , x f dP (x) . Then under certain regularity conditions, the asymptotic expected generalization error is given by 1 (6) E R(a erm ) R(a*) + 2n tr(V-1U). = More generally, for any evaluation function h( a) such that "\1 h( a*) = 0: 1 E h(a erm ) I=::j h(a*) + -tr(V- 1 "\12h? V-1U), (7) 2n where "\1 2 h is the Hessian matrix of hat a*. Note that this approach assumes that the optimal solution is unique. These results are exact asymptotically and provide better bounds than those from the standard PAC analysis. Example 5.1 We would like to study a form of the support vector machine: Consider L(a, x) = f(a T x) + ~Aa2 , z <1 z>1. Because of the discontinuity in the derivative of f , the asymptotic formula may not hold. However, if we make an assumption on the smoothness of the distribution x, then the expectation of the derivative over x can still be smooth. In this case, the smoothness of f itself is not crucial. Furthermore, in a separate report. we shall illustrate that similar small sample bounds without any assumption on the smoothness of the distribution can be obtained by using techniques related to asymptotic analysis. Consider the optimal parameter a* and letS = {x : a*Tx::; 1}. Note that Aa* = ExEsx, and U = EXES(X - ExEsx)(x - EXEsxf. Assume that 3')' > 0 S.t. P(a*T x ::; ')') = 0, then V = AI + B where B is a positive semi-definite matrix. It follows that E x2 tr(V-1U) ::; tr(U)jA ::; EXES *T Ila*I I ~::; sup Ilxll~ l la*ll~j')'. xESa X Now, consider an obtained from observations Xl risk associated with loss function L( a, x), then ExL(a emp , x) ::; inf ExL(a, x) ex = [Xl, '" ,x n ] by minimizing empirical 1 sup I lx l l~lla*ll~ + -2 ')'n asymptotically. Let A --+ 0, this scheme becomes the optimal separating hyperplane [16]. This asymptotic bound is better than typical PAC bounds with fixed A. 0 Note that although the bound obtained in the above example is very similar to the mistake bound for the perceptron online update algorithm, we may in practice obtain much better estimates from (6) by plugging in the empirical data. 376 T. Zhang References [I] N. Alon, S. Ben-David, N. Cesa-Bianchi, and D. Haussler. Scale-sensitive dimensions, uniform convergence, and learnability. Journal of the ACM, 44(4):615-631, 1997. [2] A.R. Barron. Universal approximation bounds for superpositions of a sigmoidal function. IEEE Transactions on Injormation Theory, 39(3):930-945, 1993. [3] P.L. Bartlett. The sample complexity of pattern classification with neural networks: the size of the weights is more important than the size of the network. IEEE Transactions on Information Theory, 44(2):525-536, 1998. [4] R.M. Dudley. A course on empirical processes, volume 1097 of Lecture Notes in Mathematics. 1984. [5] G.H. Golub and C.P. Van Loan. Matrix computations. Johns Hopkins University Press, Baltimore, MD, third edition, 1996. [6] D. Haussler. Generalizing the PAC model: sample size bounds from metric dimension-based uniform convergence results. In Proc. 30th IEEE Symposium on Foundations of Computer Science, pages 40-45, 1989. [7] Lee K. Jones. A simple lemma on greedy approximation in Hilbert space and convergence rates for projection pursuit regression and neural network training. Ann. Statist., 20(1) : 60~13, 1992. [8] A.N. Kolmogorov. Asymptotic characteristics of some completely bounded metric spaces. Dokl. Akad. Nauk. SSSR, 108:585-589, 1956. [9] A.N. Kolmogorov and Y.M. Tihomirov. f-entropyand f-capacity of sets in functional spaces. Amer. Math. Soc. Trans!., 17(2):277-364,1961. [10] Wee Sun Lee, P.L. Bartlett, and R.C. Williamson. Efficient agnostic learning of neural networks with bounded fan-in. IEEE Transactions on Information Theory, 42(6):2118-2132,1996. [II] D. Pollard. Convergence of stochastic processes. Springer-Verlag, New York, 1984. [12] N. Sauer. On the density of families of sets. Journal of Combinatorial Theory (Series A), 13: 145-147,1972. [13] Robert E. Schapire, Yoav Freund, Peter Bartlett, and Wee Sun Lee. Boosting the margin: a new explanation for the effectiveness of voting methods. Ann. Statist., 26(5): 1651-1686,1998. [14] 1. Shawe-Taylor, P.L. Bartlett, R.C. Williamson, and M. Anthony. Structural risk minimization over data-dependent hierarchies. IEEE Trans. In! Theory, 44(5): 19261940, 1998. [15] Y.N. Vapnik. Estimation of dependences based on empirical data. Springer-Verlag, New York, 1982. Translated from the Russian by Samuel Kotz. [16] Y.N. Vapnik. The nature of statistical learning theory. Springer-Verlag, New York, 1995. [17] Y.N. Vapnik and AJ. Chervonenkis. On the uniform convergence of relative frequencies of events to their probabilities. Theory of Probability and Applications, 16:264-280, 1971. [18] Tong Zhang. Analysis of regularized linear functions for classification problems. Technical Report RC-21572, IBM, 1999. PART IV ALGORITHMS AND ARCHITECTURE
1689 |@word version:2 briefly:1 polynomial:1 norm:5 bn:1 decomposition:1 fonn:1 tr:4 contains:1 series:1 chervonenkis:1 com:1 unction:2 afl:1 john:1 ixil:2 numerical:1 update:1 exl:4 greedy:1 wth:1 provides:1 boosting:5 math:1 lx:1 sigmoidal:1 zhang:5 height:1 mathematical:1 rc:1 differential:1 symposium:1 introduce:1 expected:2 behavior:3 decreasing:1 lll:1 cardinality:1 becomes:1 estimating:1 bounded:11 agnostic:1 xiwi:1 minimizes:3 ag:1 nf:2 voting:1 tie:2 classifier:3 wrong:1 rm:1 appear:1 positive:3 understood:1 ilxll:1 mistake:1 consequence:1 analyzing:1 might:1 emphasis:1 studied:2 limited:1 unique:1 practice:1 definite:1 xtw:1 area:1 lla:1 universal:1 empirical:7 maxx:1 convenient:1 projection:1 ila:1 risk:9 influence:1 fonnula:1 restriction:1 imposed:1 demonstrated:1 center:1 independently:2 m2:1 haussler:2 deriving:1 variation:2 hierarchy:1 exact:1 us:1 logarithmically:1 approximated:1 observed:1 sun:2 complexity:3 depend:2 solving:1 weakly:1 completely:1 translated:1 tx:1 kolmogorov:2 derivation:1 describe:1 glivenko:1 refined:2 quite:1 widely:1 valued:1 otherwise:1 statistic:1 nab:1 itself:1 injormation:1 seemingly:1 online:2 advantage:1 differentiable:1 relevant:3 nauk:1 description:1 convergence:16 regularity:1 xil:1 perfect:1 ben:1 oo:3 derive:1 ac:1 illustrate:1 alon:1 measured:1 soc:1 implies:2 indicate:1 sssr:1 stochastic:1 vc:4 ja:1 f1:6 generalization:9 exf:1 extension:1 hold:1 exp:2 achieves:2 purpose:2 estimation:1 proc:1 combinatorial:5 label:1 superposition:1 sensitive:2 weighted:1 minimization:2 always:2 derived:1 cantelli:1 dependent:2 hidden:1 jq:1 interested:1 classification:5 denoted:1 construct:1 jones:1 np:3 report:2 few:1 wee:2 replaced:1 entro:2 ab:1 llq:1 evaluation:1 withp:1 golub:1 analyzed:2 extreme:1 implication:1 necessary:1 sauer:1 perfonnance:3 iv:1 taylor:1 penalizes:1 theoretical:5 instance:1 yoav:1 entry:2 uniform:3 successful:1 ilwlll:1 learnability:1 dependency:3 combined:1 tihomirov:1 density:1 lee:3 yl:3 vm:1 hopkins:1 cesa:1 iip:1 choose:3 slowly:1 derivative:3 style:1 b2:1 satisfy:2 explicitly:1 fonnulate:1 vi:2 depends:1 later:1 view:3 analyze:1 sup:3 contribution:1 minimize:2 ir:2 characteristic:1 yield:2 weak:1 researcher:1 kolmogrov:1 explain:2 definition:5 mysterious:1 frequency:1 e2:3 proof:2 attributed:1 associated:2 improves:2 organized:1 hilbert:1 back:1 improved:1 amer:1 vel:1 furthermore:2 abh:1 nonlinear:4 aj:1 russian:1 concept:3 y2:1 regularization:5 deal:1 ll:3 covering:22 yorktown:1 samuel:1 generalized:1 recently:2 functional:1 volume:1 extend:2 slight:1 measurement:1 composition:3 nonn:6 ai:2 smoothness:3 mathematics:2 exj:1 shawe:1 inf:3 certain:6 verlag:3 inequality:1 binary:2 watson:2 tenns:1 minimum:1 ii:2 semi:1 smooth:2 technical:1 concerning:1 e1:2 plugging:1 a1:1 va:1 iog:3 prediction:1 involving:2 regression:2 y21:1 essentially:1 expectation:2 ilwllq:3 circumstance:1 metric:6 want:2 baltimore:1 grow:1 crucial:1 appropriately:1 fonnulation:1 effectiveness:2 seem:1 call:1 structural:1 independence:1 architecture:1 idea:3 bartlett:7 peter:1 pollard:2 hessian:1 york:3 useful:1 generally:1 detailed:1 clear:1 band:4 statist:2 schapire:1 unifonn:2 estimated:1 shall:4 drawn:2 v1:1 asymptotically:2 parameterized:1 tzhang:1 extends:1 family:5 kotz:1 comparable:1 bound:30 oftraining:1 fan:1 infinity:1 constraint:1 x2:1 department:1 tv:5 lp:1 restricted:1 erm:5 ln:1 discus:1 generalizes:1 pursuit:1 apply:4 barron:1 appropriate:1 dudley:1 lxi:1 hat:1 jn:1 original:1 assumes:1 cf:2 include:1 x21:1 log2:5 already:1 quantity:5 parametric:1 dependence:1 md:1 nr:2 said:1 dp:3 separate:1 separating:2 capacity:1 ix1:1 g21:1 length:1 index:1 balance:1 minimizing:1 akad:1 unfortunately:1 robert:1 potentially:1 ilw:1 negative:1 unknown:1 bianchi:1 discretize:1 observation:6 finite:2 situation:1 precise:1 y1:3 supa:2 introduced:1 david:1 specified:1 discontinuity:1 trans:2 dokl:1 pattern:1 appeared:1 max:1 explanation:2 event:1 natural:1 rely:1 regularized:6 indicator:1 scheme:2 dated:1 review:1 asymptotic:9 law:1 relative:2 loss:6 fully:1 lecture:1 freund:1 maurey:1 limitation:1 foundation:1 degree:1 sufficient:1 ibm:3 lo:1 course:1 perceptron:1 emp:1 van:1 regard:1 dimension:9 xn:1 author:1 collection:1 simplified:1 transaction:3 lloo:1 emphasize:1 keep:1 e2h:1 xi:4 lip:5 nature:1 improving:1 williamson:2 investigated:1 complex:1 anthony:1 main:3 bounding:1 edition:1 x1:2 en:1 ny:1 tong:2 oflinear:1 lq:1 xl:15 third:1 ix:1 theorem:27 formula:1 xt:1 pac:3 jensen:1 symbol:1 svm:1 dl:1 vapnik:4 restricting:1 importance:1 margin:1 flavor:1 entropy:2 generalizing:1 logarithmic:2 lt:1 simply:3 springer:3 aa:1 relies:2 acm:1 goal:1 llxi:1 consequently:1 ann:2 replace:1 lipschitz:2 loan:1 specifically:1 typical:1 hyperplane:3 wt:1 lemma:4 called:2 total:2 svd:1 la:1 support:4 noo:1 latter:1 ex:9
753
169
687 AN ANALOG VLSI CHIP FOR THIN-PLATE SURFACE INTERPOLATION John G. Harris California Institute of Technology Computation and Neural Systeins Option, 216-76 Pasadena, CA 91125 ABSTRACT Reconstructing a surface from sparse sensory data is a well-known problem iIi computer vision. This paper describes an experimental analog VLSI chip for smooth surface interpolation from sparse depth data. An eight-node ID network was designed in 3J.lm CMOS and successfully tested. The network minimizes a second-order or "thinplate" energy of the surface. The circuit directly implements the coupled depth/slope model of surface reconstruction (Harris, 1987). In addition, this chip can provide Gaussian-like smoothing of images. INTRODUCTION Reconstructing a surface from sparse sensory data is a well-known problem in computer vision. Early vision modules typically supply sparse depth, orientation, and discontinuity information. The surface reconstruction module incorporates these sparse and possibly conflicting measurements of a surface into a consistent, dense depth map. The coupled depth/slope model provides a novel solution to the surface reconstruction problem (Harris, 1987). A ID version of this model has been implemented; fortunately, its extension to 2D is straightforward. Figure 1 depicts a high-level schematic of the circuit. The di voltages represent noisy and possibly sparse input data, the ZiS are the smooth output values, and the PiS are the explicitly computed slopes. The vertical data resistors (with conductance g) control the confidence in the input data. In the absence of data these resistors are open circuits. The horizontal chain of smoothness resistors of conductance ..\ forces the derivative of the data to be smooth. This model is called the coupled depth/slope model because of the coupling between the depth and slope representations provided by the subtractor elements. The subtractors explicitly calculate a slope representation of the surface. Any depth or slope node can be made into a constraint by fixing a voltage source to the proper location in the network. Intuitively, any sudden change in slope is smoothed out with the resistor mesh. 688 Harris Figure 1. The coupled depth/slope model. The tri-directional subtractor device (shown in Figure 2) is responsible for the coupling between the depth and slope representations. If nodes A and B are set with ideal voltage sources, then node C will be forced to A - B by the device. This circuit element is unusual in that all of its terminals can act as inputs or outputs. If nodes Band C are held constant with voltage sources, then the A terminal is fixed to B + C. If A and C are input, then B becomes A - C. When further constraints are added, this device dissipates a power proportional to (A - B - C)2. In the limiting case of a continuous network, the total dissipated power is (1) The three terms arise from the power dissipated in the sub tractors and in the two different types of resistors. Energy minimization techniques and standard calculus of variations have been used to formally show that the reconstructed surfaces, z, satisfy the 1D biharmonic equation between input data points (Harris, 1987). In the tw~dimensional formulation, z is a solution of (2) This interpolant, therefore, provides the same results as minimizing the energy of a thin plate, which has been commonly used in surface reconstruction algorithms on digital computers (Grimson, 1981; Terzopoulos, 1983). IMPLEMENTATION The eight-node 1D network shown in Figure 1 was designed in 3J.lm CMOS (Mead, 1988) and fabricated through MOSIS. Three important components of the model must be mapped to analog VLSI: the two different types of resistors and the subtractors. The vertical confidence resistors are built with simple transconductance An Analog VLSI Chip for Thin-Plate Surface Interpolation A B c Figure 2. Tri-directional subtract constraint device. amplifiers (transamps) connected as followers. The bias voltage of the transamp follower determines its conductance (g) and therefore signifies the certainty of the data. If there are no data for a given location, the corresponding transamp follower is turned off. The horizontal smoothness resistors are implemented with Mead's saturating resistor (Mead, 1988). Since conventional CMOS processes lack adequate resistive elements, we are forced to build resistors out of transistor elements. The bias voltage for Mead's resistor allows the effective conductance of these circuit elements to vary over many orders of magnitude. The most difficult component to implement in analog VLSI is the subtract constraint device. Its construction led to a general theory of constraint boxes which can be used to implement all sorts of constraints which are useful in early vision (Harris, 1988). The implementation of the subtract constraint device is a straightforward application of constraint box theory. Figure 3 shows a generic n terminal constraint box enforcing a constraint F on its voltage terminals. The constraints are enforced by generating a feedback current lie for each constrained voltage terminal. Suppose F can be written as One possible feedback equation which implements this constraint is given by 1.: 8F = - F8Vl: - (4) When this particular choice of feedback current is used, the constraint box minimizes the least-squares error in the constraint equation (Harris, 1989). Notice that F can be scaled by any arbitrary scaling factor. This scaling factor and the capacitance at each node determine the speed of convergence of a single constraint box. 689 690 Harris V t-------i~-.. k Figure 3. Generic n terminal constraint box. The subtract constraint box given in Figure 2 requires a constraint of A - B which leads to the following error equation: F(A,B,C) =A- B - =C, (5) C Straightforward application of constraint box theory yields 8F -F 8A B - C) -F~ = (A- B - 1B Ie =-(A - - -F :~ = (A - C) (6) B - C) where lA, I B , and Ie represent feedback currents that must be generated by the device. These current feedback equations can be implemented with two modified widerange transamps (see Figure 4). In its linear range, a single transamp produces a current proportional to the difference of its two inputs. The negative input to each transamp is indicated by an inverting circle. The transamps have been modified to produce four outputs, two positive and two negative. The negative outputs are also represented by inverting circles. Because the difference terminal C can be positive or negative, it is measured with respect to a voltage reference VREF. VREF is a global signal which defines zero slope. As seen in Figure 4, the An Analog VLSI Chip for Thin-Plate Surface Interpolation A I----V REF c B IB Ie ~ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _- J Figure 4. Tri-directional subtract constraint box. proper combination of positive and negative outputs from the two transamps are fed back to the voltage terminals to implement the feedback equations given in eq. (6). Analog networks which solve most regularizable early vision problems can be designed with networks consisting solely of linear resistances and batteries (Poggio and Koch, 1985). Unfortunately, many times these networks contain negative resistances that are troublesome to implement in analog VLSI. For example, the circuit shown in Figure 5 computes the same solutions as the coupled depth/slope network described in this paper. Interestingly, a 2-D implementation of this idea was implemented in the 1960s using inductors and capacitors (Volynskii and Bukhman, 1965). Proper choice of the frequency of alternating current allowed the circuit elements to act as pure positive and negative impedances. Unfortunately, negative resistances are troublesome to implement, especially in analog VLSI. One of the big advantages of using constraint boxes to implement early vision algorithms is that the resulting networks do not require negative resistances. ANALYSIS Figure 6 shows a sample output of the circuit. Data (indicated by vertical dashed lines) were supplied at nodes 2, 5, and 8. As expected, the chip finds a smooth solution (solid line) which extrapolates beyond the known data points. It is wellknown that a single resistive grid minimizes the first-order or membrane energy of a surface. Luo, Koch, and Mead (1988) have implemented a 48x48 resistive grid to perform surface interpolation. Figure 6 also shows the simulated performance of a first-order energy or membrane energy minimization. Data points are again supplied at nodes 2, 5, and 8. In contrast to the second-order chip results, the solution (dashed line) is much more jagged and does not extrapolate outside of 691 692 Harris ? g -R -R -R -R -R -R Figure 5. A negative-resistor resistor solution to the ID biharmonic equation. I .? .... .... ~ 1.7 ~ .... ~ ~ ~ ~ ~ -. ~ I I / 1.6 ~ ~ / / ~ ~ 1.5 1.4 13+------+------+------+------~----~------~----~ I 2 3 4 5 6 7 ? Figure 6. Measured data from the second-order chip (solid line) and simulated first-order result (dashed line). An Analog VLSI Chip for Thin-Plate Surface Interpolation 1.0 0.' 0.6 ,, ,, , I 0.4 I I I I / G.2 / ,, " / ..-.-----.: . ' . .. . ... o.o+r-~.~ .... .... ... ,. --..... .... 42+---~~--~----~--~~-~---~--~~--+--~--~ -5 "" -3 -2 -I 0 2 3 4 5 Figure 7. Graphical comparison of ID analytic Green's functions for first-order (dashed line), second-order (dotted line) and Gaussian (solid line). the known data points (for example, see node 1). Interestingly, psychophysics experiments support the smoother interpolant used by the second-order coupled depth/slope chip (Grimson, 1981). Unlike the second-order network, the firstorder network is not rigid enough to incorporate either orientation constraints or orientation discontinuities (Terzopoulos, 1983). Image smoothing is a special case of surface interpolation where the data are given on a dense grid. The first-order network is a poor smoothing operator. A comparison of analytic Green's function of first and second-order networks is shown in Figure 7 (the first-order shown with a dashed line and the secondorder with a solid line). Note that the analytic Green's function of the secondorder network (solid line) and that of standard Gaussian convolution (dotted line) are nearly identical. This fact was pointed out by Poggio, Voorhees, and Yuille (1986), when they suggested the use of the second-order energy to regularize the edge detection problem. Gaussian convolution has been claimed by many authors to be the "optimal" smoothing operator and is commonly used as the first stage of edge detection. Though the second-order network can be used to smooth images, Gaussian convolution cannot be used to solve the more difficult problem of interpolating from sparse data points. 693 694 Harris CONCLUSION Biharmonic surface interpolation has been successfully demonstrated in analog VLSI. To test true performance, we plan to combine a larger version of this chip with an analog stereo network. Work has already started on building the necessary circuitry for discontinuity detection during surface reconstruction. The Gaussianlike smoothing effect of this network will be further explored through building a network with photoreceptors supplying dense data input. Acknowledgements Support for this research was provided by the Office of Naval Research and the System Development Foundation. The author is a Hughes Aircraft Fellow and thanks Christof Koch and Carver Mead for their ongoing support. Additional thanks to Berthold Horn for several helpful suggestions. References Grimson, W.E.L. From Images to Surfaces, MIT Press, Cambridge, (1981). Harris, J.G. A new approach to surface reconstruction: the coupled depth/slope model, Proc. IEEE First Inti. Con! Computer Vision, pp. 277-283, London, (1987). Harris, J.G. Solving early vision problems with VLSI constraint networks, Neural Architectures for Computer Vision Workshop, AAAI-88, Minneapolis, Minnesota, Aug. 20 (1988). Harris, J .G. Designing analog constraint boxes to solve energy minimization problems in vision, submitted to INNS Neural Networks Conference, Washington D.C., June (1989) Luo, J., Koch, C., and Mead, C. An experimental subthreshold, analog CMOS tw<r dimensional surface interpolation circuit, Neural Information and Processing Systems Conference, Denver, Nov. (1988). Mead, C.A. Analog VLSI and Neural Systems, Addison-Wesley, Reading, (1989). Poggio, T. and Koch, C. Ill-posed problems in early vision: from computational theory to analogue networks, Proc. R. Soc. Lond. B 226: 303-323 (1985). Poggio, T., Voorhees, H., and Yuille, A. A regularized solution to edge detection, Arti! Intell. Lab Memo No. 833, MIT, Cambridge, (1986). Terzopoulos, D. Multilevel computational processes for visual surface reconstruction, Compo Vision Graph. Image Proc. 24: 52-96 (1983). Volynskii, B. A. and Bukhman, V. Yeo Analogues for Solution of Boundary- Value Problems, Pergamon Press, New York, (1965).
169 |@word aircraft:1 version:2 open:1 calculus:1 arti:1 solid:5 interestingly:2 current:6 luo:2 follower:3 must:2 written:1 john:1 mesh:1 analytic:3 designed:3 device:7 supplying:1 compo:1 sudden:1 provides:2 node:10 location:2 supply:1 resistive:3 combine:1 expected:1 terminal:8 becomes:1 provided:2 circuit:9 vref:2 minimizes:3 fabricated:1 certainty:1 fellow:1 act:2 firstorder:1 scaled:1 control:1 christof:1 positive:4 troublesome:2 id:4 mead:8 solely:1 interpolation:9 range:1 minneapolis:1 horn:1 responsible:1 hughes:1 implement:8 zis:1 confidence:2 cannot:1 operator:2 conventional:1 map:1 demonstrated:1 straightforward:3 pure:1 regularize:1 variation:1 limiting:1 construction:1 suppose:1 designing:1 secondorder:2 element:6 module:2 calculate:1 connected:1 grimson:3 battery:1 interpolant:2 solving:1 yuille:2 chip:11 represented:1 forced:2 effective:1 london:1 transamp:4 outside:1 larger:1 solve:3 posed:1 noisy:1 advantage:1 transistor:1 inn:1 reconstruction:7 turned:1 convergence:1 produce:2 generating:1 cmos:4 coupling:2 fixing:1 measured:2 aug:1 eq:1 soc:1 implemented:5 require:1 multilevel:1 extension:1 koch:5 lm:2 circuitry:1 vary:1 early:6 proc:3 successfully:2 minimization:3 mit:2 gaussian:5 modified:2 voltage:10 office:1 june:1 naval:1 contrast:1 helpful:1 rigid:1 typically:1 pasadena:1 vlsi:12 orientation:3 ill:1 development:1 plan:1 smoothing:5 subtractor:2 constrained:1 psychophysics:1 special:1 washington:1 identical:1 nearly:1 thin:5 intell:1 consisting:1 amplifier:1 conductance:4 detection:4 held:1 chain:1 edge:3 necessary:1 poggio:4 x48:1 carver:1 circle:2 signifies:1 thanks:2 ie:3 off:1 again:1 aaai:1 possibly:2 derivative:1 yeo:1 satisfy:1 jagged:1 explicitly:2 biharmonic:3 lab:1 sort:1 option:1 slope:14 square:1 yield:1 subthreshold:1 directional:3 submitted:1 energy:8 frequency:1 pp:1 di:1 con:1 back:1 wesley:1 formulation:1 box:11 though:1 stage:1 horizontal:2 lack:1 defines:1 indicated:2 building:2 effect:1 contain:1 true:1 alternating:1 during:1 plate:5 image:5 novel:1 denver:1 analog:15 measurement:1 cambridge:2 smoothness:2 grid:3 pointed:1 minnesota:1 surface:24 wellknown:1 claimed:1 seen:1 fortunately:1 additional:1 determine:1 signal:1 dashed:5 smoother:1 smooth:5 schematic:1 vision:12 represent:2 addition:1 source:3 unlike:1 tri:3 incorporates:1 capacitor:1 ideal:1 iii:1 enough:1 architecture:1 idea:1 stereo:1 resistance:4 york:1 adequate:1 useful:1 band:1 supplied:2 notice:1 dotted:2 four:1 mosis:1 graph:1 enforced:1 scaling:2 inductor:1 extrapolates:1 constraint:24 speed:1 lond:1 transconductance:1 combination:1 poor:1 membrane:2 describes:1 reconstructing:2 tw:2 intuitively:1 inti:1 equation:7 addison:1 fed:1 unusual:1 eight:2 generic:2 graphical:1 build:1 especially:1 capacitance:1 added:1 already:1 pergamon:1 thinplate:1 mapped:1 simulated:2 enforcing:1 minimizing:1 difficult:2 unfortunately:2 negative:10 memo:1 implementation:3 proper:3 perform:1 vertical:3 convolution:3 smoothed:1 arbitrary:1 inverting:2 california:1 conflicting:1 discontinuity:3 beyond:1 suggested:1 reading:1 built:1 green:3 analogue:2 power:3 force:1 regularized:1 dissipates:1 technology:1 started:1 dissipated:2 coupled:7 acknowledgement:1 tractor:1 suggestion:1 proportional:2 digital:1 foundation:1 consistent:1 pi:1 bias:2 terzopoulos:3 institute:1 sparse:7 feedback:6 depth:13 berthold:1 boundary:1 computes:1 sensory:2 author:2 made:1 commonly:2 reconstructed:1 nov:1 global:1 photoreceptors:1 continuous:1 impedance:1 ca:1 interpolating:1 dense:3 big:1 arise:1 allowed:1 ref:1 depicts:1 sub:1 resistor:13 lie:1 ib:1 explored:1 voorhees:2 workshop:1 magnitude:1 subtract:5 led:1 visual:1 saturating:1 determines:1 harris:13 absence:1 change:1 called:1 total:1 experimental:2 la:1 formally:1 support:3 ongoing:1 incorporate:1 tested:1 extrapolate:1
754
1,690
Bifurcation Analysis of a Silicon Neuron Girish N. Patel] , Gennady s. Cymbalyuk2,3, Ronald L. Calabrese2 , and Stephen P. DeWeerth 1 lSchool of Electrical and Computer Engineering Georgia Institute of Technology Atlanta, Ga. 30332-0250 {girish.patel, steve.deweerth} @ece.gatech.edu 2Department of Biology Emory University 1510 Clifton Road, Atlanta, GA 30322 {gcym, rcalabre}@biology.emory.edu 3Institute of Mathematical Problems in Biology RAS Pushchino, Moscow Region, Russia 142292 (on leave) Abstract We have developed a VLSI silicon neuron and a corresponding mathematical model that is a two state-variable system. We describe the circuit implementation and compare the behaviors observed in the silicon neuron and the mathematical model. We also perform bifurcation analysis of the mathematical model by varying the externally applied current and show that the behaviors exhibited by the silicon neuron under corresponding conditions are in good agreement to those predicted by the bifurcation analysis. 1 Introduction The use of hardware models to understand dynamical behaviors in biological systems is an approach that has a long and fruitful history [1 ][2]. The implementation in silicon of oscillatory neural networks that model rhythmic motor-pattern generation in animals is one recent addition to these modeling efforts [3][4]. The oscillatory patterns generated by these systems result from intrinsic membrane properties of individual neurons and their synaptic interactions within the network [5]. As the complexity of these oscillatory silicon systems increases, effective mathematical analysis becomes increasingly more important to our understanding their behavior. However, the nonlinear dynamical behaviors of the model neurons and the large-scale interconnectivity among these neurons makes it very difficult to analyze theoretically the behavior of the resulting very large-scale integrated (VLSI) systems. Thus, it is important to first identify methods for modeling the model neurons that underlie these oscillatory systems. Several simplified neuronal models have been used in the mathematical simulations of pattern generating networks [6][7][8] . In this paper, we describe the implementation of a 732 G. N Patel, G. S. Cymbalyuk, R. L. Calabrese and S. P. DeWeerth two-state-variable silicon neuron that has been used effectively to develop oscillatory networks [9][10]. We then derive a mathematical model of this implementation and analyze the neuron and the model using nonlinear dynamical techniques including bifurcation analysis [11]. Finally, we compare the experimental data derived from the silicon neuron to that obtained from the mathematical model. 2 The silicon model neuron The schematic for our silicon model neuron is shown in Figure 1. This silicon neuron is inspired by the two-state, Morris-Lecar neuron model [12][ 13]. Transistor M I ' analogous to the voltage-gated calcium channel in the Morris-Lecar model, provides an instantaneous inward current that raises the membrane potential towards V High when the membrane is depolarized. Transistor M2 ' analogous to the voltage-gated potassium channel in the Morris-Lecar model, provides a delayed outward current that lowers the membrane potential toward V Low when the membrane is depolarized. V H and V L are analogous to the half-activation voltages for the inward and outward currents, respectively. The voltages across C I and C2 are the state variables representing the membrane potential, V, and the slow "activation" variable of the outward current, W, respectively. The W -nullcline represents its steady-state activation curve. Unlike the Morris-Lecar model, our silicon neuron model does not possess a leak current. Using current conservation at node V, the net current charging C I is given by (1) where iH and iL are the output currents of a differential pair circuit, and a p and aN describe the ohmic effects of transistors M J and M2 , respectively. The net current into C 2 is given by (2) where ix is the output current of the OTA, and ~p and ~N account for ohmic effects of the pull-up and the pull-down transistors inside the OTA. V High OTA w v '------<>--- V Low Figure 1: Circuit diagram of the silicon neuron. The circuit incorporates analog building blocks including two differential pair circuits composed of a bias current, IB H , and transistors M4-M s, and a bias current, IB L , and transistors M6-M7' and a single followerintegrator circuit composed of an operational transconductance amplifier (OTA), Xl in the configuration shown and a load capacitor, C 2 . The response of the follower-integrator circuit is similar to a first-order low-pass filter. 733 Bifurcation Analysis ofa Silicon Neuron The output currents of the differential-pair and an OTA circuits, derived by using subthreshold transistor equations [2], are a Fenni function and a hyperbolic-tangent function, respectively [2]. Substituting these functions for i H , i L , and ix in (1) and (2) yields . C 1V = Iexta p + IBH e K(V-YH) / U T l+e e K(V _ YH) / U T a p - K(W - YL) / U T IBL l+e K(W _ YL) / U T aN (3) where v - V High / UT a p = I -e aN ~N = 1- e = 1- e Y Low - V I UT - W / UT (4) U T is the thennal voltage, V dd is the supply voltage, and K is a fabrication dependent parameter. The tenns a p and aN limit the range of V to within V High and V Low' and the terms ~p and ~N limit the range of W to within the supply rails (Vdd and Gnd). In order to compare the model to the experimental results, we needed to determine values for all of the model parameters. V Hi!\h' V Low' V H' V L ' and V dd were directly measured in experiments. The parameters IBH and IBL were measured by voltage-clamp experiments performed on the silicon neuron. At room temperature, U T ::::: 0.025 volts. The value of K ::::: 0.65 was estimated by measuring the slope of the steady-state activation curve of inward current. Because W was implemented as an inaccessible node, IT could only be estimated. Based on the circuit design, we can assume that the bias currents IT and IBH are of the same order of magnitude. We choose IT::::: 2.2 nA to fit the bifurcation diagram (see Figure 3). Cl and C2, which are assumed to be identical according to the physical design, are time scaling parameters in the model. We choose their values (Cl =C2 =28 pF) to fit frequency dependence on lext (see Figure 4). 3 Bifurcation analysis The silicon neuron and the mathematical model! described by (3) demonstrate various dynamical behaviors under different parametric conditions. In particular, stable oscillations and steady-state equilibria are observed for different values of the externally applied current, I ext . We focused our analysis on the influence of I ext on the neuron behavior for two reasons: (i) it provides insight about effects of synaptic currents, and (ii) it allows comparison with neurophysiological experiments in which polarizing current is used as a primary control parameter. The main results of this work are presented as the comparison between the mathematical models and the experimental data represented as bifurcation diagrams and frequency dependencies. The null clines described by (3) and for lext = 32 nA are shown in Figure 2A. In the regime that we operate the circuit, the W -null cline is an almost-linear curve and the Vnullcline is an N-shaped curve. From (3), it can be seen that when IBH + lext > IBL the nullclines cross at (V, W)::::: (V High , V High ) and the system has high voltage (about 5 volts) steady-state equilibrium. Similarly, for I ext close to zero, the system has one stable equilibrium point close to (V, W) ::::: (V Low' V Low). !The parameters used throughout the analyses of the model are V Low = 0 V , V High = 5 V, V L = V H = 2.5 V, I BH = 6.5 nA , I BL = 42 nA, IT = 2.2 nA , V dd = 5 V, V t = 0.025 mV, and K = 0.65. G. N Patel, G. S. Cymba/yule, R. L. Calabrese and S. P. De Weerth 734 A 2.85 W-nullcline / / @ 2.8 / 2.75 ~ 2.7 (5 > .....,.. ~ 2.65 V-nullcline 2.6 2.55 2.5 I 2.45~______~______~______~______~______~1 o 3 2 5 4 V (volts) 3.2 B 3 -- 2.8 In l 2.6 > 2.4 2.2 2 0 5 10 15 20 25 30 35 time (msec) Figure 2: Nullclines and trajectories in the model of the silicon neuron for lex! = 32 nA. The system exhibits a stable limit-cycle (filled circles), an unstable limit-cycle (unfilled circles), and stable equilibrium point. Unstable limit-cycle separates the basins of attraction of the stable limit-cycle and stable equilibrium point. Thus, trajectories initiated within the area bounded by the unstable limit-cycle approach the stable equilibrium point (solid line in A's inset, and "x's" in B). Trajectories initiated outside the unstable limitcycle approach the stable limit-cycle . In A, the inset shows an expansion at the intersection of the V - and W -nullclines. 735 Bifurcation Analysis ofa Silicon Neuron Experimental data A ........................... 5 .. 003 .0 x ~ ~ ? x ..?? I 1 0 ",>0< x xxxx ? ? x >t#* xxxx >2 x ? ? ? ? ? 4 2:- ~ ?? x ? ? ? ??????????????????????????? 10 0 20 ? ?? 30 40 50 40 50 lext (nAmps) Modeling data B 5 ( 0 -> 2 > .? ? ? 4 003 .- ....................... -. ?? ,J rt---------; ? .'- ....................... 1 0 0 10 20 ? ? ~ 30 Iext (nAmps) Figure 3: Bifurcation diagrams of the hardware implementation (A) and of the mathematical model (B) under variation of the externally applied current. In A, the steadystate equilibrium potential of V is denoted by "x"s. The maximum and minimum values of V during stable oscillations are denoted by the filled circles. In B, the stable and unstable equilibrium points are denoted by the solid and dashed curve, respectively, and the minimum and maximum values of the stable and unstable oscillations are denoted by the filled and unfilled circles, respectively. In B, limit-cycle oscillations appear and disappear via sub-critical Andronov-Hopf bifurcations. The bifurcation diagram (B) was computed with the LOCBIF program [14]. G. N. Patel, G. S. Cymbalyuk, R. L. Calabrese and S. P. De Weerth 736 A B Experimental data 100 ? 80 ? N ? >. 60 ? . ? ? 100~--~--~----~~ -N 80 ?? >. 60 ? ~ g . .~~.. .., ... ~ LL ? ? ??? t:T 20 -::c ? ~ 40 Modeling data c ~. ? ??? ??? ~ Q) :l t:T ... 40 Q) LL o~--~----~------~ o 0 10 20 lext (nAmps) 30 20 0 0 10 20 30 lext (nAmps) Figure 4: Frequency dependence of the silicon neuron (A) and the mathematical model (B) on the externally applied current. For moderate values of lext ([1 nA,34 nA)), the stable and unstable equilibrium points are close to (V, W) ::::: (V H' V L) (Figure 3). In experiments in which lext was varied, we observed a hard loss of the stability of the steady-state equilibrium and a transition into oscillations at lext = 7.2 nA (I ext = 27.5 nA). In the mathematical model, at the critical value of lext = 7.7 nA (lext = 27.8 nA), an unstable limit cycle appears via a subcritical Andronov-Hopf bifurcation. This unstable limit cycle merges with the stable limit cycle at the fold bifurcation at lext = 3.4 nA (lext = 32.1 nA). Similarly, in the experiments, we observed hard loss of stability of oscillations at lext = 2.0 nA (I ext = 32.8 nA). Thus, the system demonstrates hysteresis. For example. when lext = 20 nA the silicon neuron has only one stable regime, namely, stable oscillations. Then if external current is slowly increased to lext = 32.8 nA. the form of oscillations changes. At this critical value of the current, the oscillations suddenly lose stability, and only steady-state equilibrium is stable. Now, when the external current is reduced, the steady-state equilibrium is observed at the values of the current where oscillations were previously exhibited. Thus, within the ranges of externally applied currents (2.0,7.2) and (27.5,32.8), oscillations and a steady-state equilibrium are stable regimes as shown in Figure 2. 4 Discussion We have developed a two-state silicon neuron and a mathematical model that describes the behavior of this neuron. We have shown experimentally and verified mathematically that this silicon neuron has three regions of operation under the variation of its external current (one of its parameters). We also perform bifurcation analysis of the mathematical model by varying the externally applied current and show that the behaviors exhibited by the silicon neuron under corresponding conditions are in good agreement to those predicted by the bifurcation analysis. This analysis and comparison to experiment is an important step toward our understanding of a variety of oscillatory hardware networks that we and others are developing. The Bifurcation Analysis ofa Silicon Neuron 737 model facilitates an understanding of the neurons that the hardware alone does not provide. In particular for this neuron, the model allows us to determine the location of the unstable fixed points and the types of bifurcations that are exhibited. In higher-order systems, we expect that the model will provide us insight about observed behaviors and complex bifurcations in the phase space. The good matching between the model and the experimental data described in this paper gives us some confidence that future analysis efforts will prove fruitful. Acknowledgments S. DeWeerth and G. Patel are funded by NSF grant IBN-95 II 721 , G.S. Cymbalyuk is supported by Russian Foundation of Fundamental Research grant 99-04-49112, R.L. Calabrese and G.S. Cymbalyuk are supported by NIH grants NS24072 and NS34975. References [1] Van Der Pol, B (1939) Biological rhythms considered as relaxation oscillations In H. Bremmer and c.J. Bouwkamp (eds) Selected Scientific Papers, Vol 2, North Holland Pub. Co., 1960. [2] Mead, C.A. Analog VLSI and Neural Systems. Addison-Wesley, Reading, MA, 1989. [3] Simoni, M.E, Patel, G.N., DeWeerth, S.P., & Calabrese, RL. Analog VLSI model of the leech heartbeat elemental oscillator. Sixth Annual Computational Neuroscience Meeting, 1997. in Big Sky, Montana. [4] DeWeerth, S., Patel, G., Schimmel, D., Simoni, M. and Calabrese, R (1997). In Proceedings of the Seventeenth Conference on Advanced Research in VLSI, RB. Brown and A.T. Ishii (eds), Los Alamitos, CA: IEEE Computer Society, 182-200. [5] Marder, E. & Calabrese, RL. (1996) Principles of rhythmic motor pattern generation. Physiological Reviews 76 (3): 687-717. [6] Kopell, N. & Ermentrout, B. (1988) Coupled oscillators and the design of central pattern generators. Mathematical biosciences 90: 87-109. [7] Skinner, EK., Turrigiano, G.G., & Marder, E. (1993) Frequency and burst duration in oscillating neurons and two-cell networks. Biological Cybernetics 69: 375-383. [8] Skinner, EK., Gramoll, S., Calabrese, R.L., Kopell, N. & Marder, E. (1994) Frequency control in biological half-center oscillators. In EH. Eeckman (ed.), Computation in neurons and neural systems, pp. 223-228, Boston: Kluwer Academic Publishers. [9] Patel, G. Holleman, J., DeWeerth, S. Analog VLSI model of intersegmental coordination with nearest-neighbor coupling. In , 1997. [10] Patel, G. A neuromorphic architecture for modelling intersegmental coordination. Ph.D. dissertation, Georgia Institute of Technology, 1999. [11] J. Guckenheimer and P. Holmes. Nonlinear Oscillations, Dynamical Systems, and Bifurcation of Vector Fields. Applied Mathematical Sciences, 42. Springer-Verlag, New York, New York, Heidelberg, Berlin, 1983. [12] Morris, C. and Lecar, H. (1981) Voltage oscillations in the barnacle giant muscle fiber. Biophys. J, 35: 193-213. [13] Rinzel, J. & Ermentrout, G.B. (1989) Analysis of Neural Excitability and Oscillations. In C. Koch and I. Segev (eds) Methods in Neuronal Modeling from Synapses to Networks. MIT press, Cambridge, MA. [14] Khibnik, A. I. , Kuznetsov, Yu.A., Levitin, v.v., Nikolaev, E.V. (1993) Continuation techniques and interactive software for bifurcation analysis of ODEs and iterated maps. Physica D 62 (1-4): 360-367.
1690 |@word simulation:1 gennady:1 solid:2 configuration:1 pub:1 current:28 emory:2 activation:4 follower:1 ronald:1 ota:5 motor:2 rinzel:1 alone:1 half:2 selected:1 dissertation:1 provides:3 node:2 location:1 mathematical:17 burst:1 c2:3 differential:3 m7:1 supply:2 hopf:2 prove:1 inside:1 theoretically:1 ra:1 behavior:11 integrator:1 inspired:1 nullcline:3 pf:1 becomes:1 bounded:1 circuit:10 inward:3 null:2 developed:2 giant:1 sky:1 ofa:3 interactive:1 demonstrates:1 control:2 underlie:1 grant:3 appear:1 engineering:1 limit:12 ext:5 initiated:2 mead:1 montana:1 co:1 range:3 seventeenth:1 acknowledgment:1 block:1 area:1 hyperbolic:1 matching:1 confidence:1 road:1 schimmel:1 ga:2 close:3 bh:1 influence:1 fruitful:2 map:1 center:1 duration:1 focused:1 m2:2 insight:2 attraction:1 holmes:1 pull:2 stability:3 variation:2 analogous:3 agreement:2 observed:6 electrical:1 region:2 cycle:10 leech:1 leak:1 complexity:1 inaccessible:1 pol:1 ermentrout:2 vdd:1 raise:1 heartbeat:1 various:1 represented:1 fiber:1 ohmic:2 describe:3 effective:1 ibh:4 outside:1 transistor:7 net:2 turrigiano:1 clamp:1 interaction:1 los:1 holleman:1 elemental:1 potassium:1 oscillating:1 generating:1 leave:1 derive:1 develop:1 coupling:1 measured:2 nearest:1 implemented:1 predicted:2 ibn:1 filter:1 cline:2 biological:4 mathematically:1 physica:1 koch:1 considered:1 equilibrium:13 substituting:1 intersegmental:2 lose:1 coordination:2 interconnectivity:1 guckenheimer:1 mit:1 nullclines:3 varying:2 voltage:9 gatech:1 derived:2 modelling:1 ibl:3 ishii:1 dependent:1 integrated:1 vlsi:6 among:1 denoted:4 animal:1 bifurcation:21 field:1 shaped:1 biology:3 represents:1 identical:1 yu:1 future:1 others:1 composed:2 individual:1 delayed:1 m4:1 phase:1 amplifier:1 atlanta:2 pushchino:1 bremmer:1 filled:3 circle:4 increased:1 modeling:5 measuring:1 neuromorphic:1 fabrication:1 dependency:1 fundamental:1 yl:2 na:18 central:1 choose:2 russia:1 slowly:1 external:3 ek:2 account:1 potential:4 de:2 north:1 hysteresis:1 mv:1 performed:1 analyze:2 slope:1 il:1 subthreshold:1 identify:1 yield:1 andronov:2 iterated:1 calabrese:8 trajectory:3 kopell:2 cybernetics:1 history:1 oscillatory:6 synapsis:1 synaptic:2 ed:4 sixth:1 frequency:5 polarizing:1 pp:1 bioscience:1 ut:3 appears:1 wesley:1 steve:1 higher:1 response:1 deweerth:7 xxxx:2 nonlinear:3 scientific:1 russian:1 building:1 effect:3 brown:1 skinner:2 excitability:1 volt:3 unfilled:2 ll:2 during:1 steady:8 rhythm:1 demonstrate:1 temperature:1 steadystate:1 instantaneous:1 limitcycle:1 nih:1 physical:1 rl:2 lext:16 analog:4 kluwer:1 silicon:24 cambridge:1 eeckman:1 similarly:2 funded:1 stable:17 recent:1 moderate:1 verlag:1 barnacle:1 tenns:1 meeting:1 der:1 muscle:1 seen:1 minimum:2 determine:2 dashed:1 stephen:1 ii:2 academic:1 cross:1 long:1 schematic:1 girish:2 cell:1 addition:1 ode:1 diagram:5 publisher:1 operate:1 depolarized:2 exhibited:4 unlike:1 posse:1 facilitates:1 incorporates:1 capacitor:1 m6:1 variety:1 fit:2 architecture:1 effort:2 york:2 outward:3 morris:5 hardware:4 gnd:1 ph:1 reduced:1 continuation:1 nsf:1 estimated:2 neuroscience:1 rb:1 levitin:1 vol:1 verified:1 subcritical:1 relaxation:1 almost:1 throughout:1 oscillation:15 scaling:1 hi:1 fold:1 annual:1 marder:3 segev:1 software:1 transconductance:1 department:1 developing:1 according:1 membrane:6 across:1 describes:1 increasingly:1 equation:1 previously:1 needed:1 addison:1 operation:1 lecar:5 moscow:1 disappear:1 society:1 suddenly:1 bl:1 alamitos:1 lex:1 parametric:1 primary:1 dependence:2 rt:1 exhibit:1 separate:1 berlin:1 unstable:10 toward:2 reason:1 simoni:2 difficult:1 implementation:5 design:3 calcium:1 perform:2 gated:2 neuron:34 varied:1 pair:3 namely:1 merges:1 dynamical:5 pattern:5 regime:3 reading:1 program:1 including:2 charging:1 critical:3 eh:1 advanced:1 representing:1 technology:2 coupled:1 review:1 understanding:3 tangent:1 loss:2 expect:1 generation:2 generator:1 foundation:1 basin:1 dd:3 principle:1 supported:2 bias:3 understand:1 institute:3 neighbor:1 rhythmic:2 van:1 curve:5 transition:1 simplified:1 iext:1 patel:10 conservation:1 assumed:1 channel:2 ca:1 operational:1 heidelberg:1 expansion:1 cl:2 complex:1 main:1 big:1 neuronal:2 georgia:2 slow:1 sub:1 msec:1 xl:1 rail:1 ib:2 yh:2 ix:2 externally:6 down:1 yule:1 load:1 inset:2 physiological:1 intrinsic:1 ih:1 effectively:1 magnitude:1 biophys:1 boston:1 intersection:1 neurophysiological:1 holland:1 kuznetsov:1 clifton:1 springer:1 ma:2 towards:1 oscillator:3 room:1 hard:2 change:1 experimentally:1 pas:1 ece:1 experimental:6
755
1,691
Greedy importance sampling Dale Schuurmans Department of Computer Science University of Waterloo dale@cs.uwaterloo.ca Abstract I present a simple variation of importance sampling that explicitly searches for important regions in the target distribution. I prove that the technique yields unbiased estimates, and show empirically it can reduce the variance of standard Monte Carlo estimators. This is achieved by concentrating samples in more significant regions of the sample space. 1 Introduction It is well known that general inference and learning with graphical models is computationally hard [1] and it is therefore necessary to consider restricted architectures [13], or approximate algorithms to perform these tasks [3, 7]. Among the most convenient and successful techniques are stochastic methods which are guaranteed to converge to a correct solution in the limit oflarge samples [10, 11, 12, 15]. These methods can be easily applied to complex inference problems that overwhelm deterministic approaches. The family of stochastic inference methods can be grouped into the independent Monte Carlo methods (importance sampling and rejection sampling [4, 10, 14]) and the dependent Markov Chain Monte Carlo (MCMC) methods (Gibbs sampling, Metropolis sampling, and "hybrid" Monte Carlo) [5, 10, 11, 15]. The goal of all these methods is to simulate drawing a random sample from a target distribution P (x) (generally defined by a Bayesian network or graphical model) that is difficult to sample from directly. This paper investigates a simple modification of importance sampling that demonstrates some advantages over independent and dependent-Markov-chain methods. The idea is to explicitly search for important regions in a target distribution P when sampling from a simpler proposal distribution Q. Some MCMC methods, such as Metropolis and "hybrid" Monte Carlo, attempt to do something like this by biasing a local random search towards higher probability regions, while preserving the asymptotic "fair sampling" properties of the exploration [11, 12]. Here I investigate a simple direct approach where one draws points from a proposal distribution Q but then explicitly searches in P to find points from significant regions. The main challenge is to maintain correctness (i.e., unbiased ness) of the resulting procedure, which we achieve by independently sampling search subsequences and then weighting the sample points so that their expected weight under the proposal distribution Q matches their true probability under the target P. 597 Greedy Importance Sampling Importance sampling ? Draw Xl , ... , X n independently from Q. ? Weight each point Xi by W(Xi) = ~I::l. ? For a random variable, f, estimate E p (.,) f(x) I ",n by f = n L.Ji=I f(Xi)W(Xi). "Indirect" importance sampling ? Draw Xl, ... ,xn independently from Q. .Weighteachpointxibyu(xi) = (3~~i/. ? For a random variable, f, estimate Ep(.,)f(x) A by A f = ",n L.Ji=I f(Xi)U(Xi)/ ",n L.Ji=I U(Xi). Figure 1: Regular and "indirect" importance sampling procedures 2 Generalized importance sampling Many inference problems in graphical models can be cast as determining the expected value of a random variable of interest, f, given observations drawn according to a target distribution P. That is, we are interested in computing the expectation Ep(x) f(x). Usually the random variable f is simple, like the indicator of some event, but the distribution P is generally not in a form that we can sample from efficiently. Importance sampling is a useful technique for estimating Ep(x) f (x) in these cases. The idea is to draw independent points xl, .. " Xn from a simpler "proposal" distribution Q, but then weight these points by w(x) P(x)/Q(x) to obtain a "fair" representation of P. Assuming that we can efficiently evaluate P(x) at each point, the weighted sample can be used to estimate desired expectations (Figure 1). The correctness (i.e., unbiasedness) of this procedure is easy to establish, since the expected weighted value of f under Q is just Eq(x)f(x)w(x) = = EXEX [f(x)w(x)] Q(x) = Ex EX [f(x)~t:n Q(x) = EXEX f(x)P(x) = Ep(x)f(x), This technique can be implemented using "indirect" weights u( x) = f3P (x) / Q ( x) and an alternative estimator (Figure 1) that only requires us to compute a fixed multiple of P (x). This preserves asymptotic correctness because ~ E7=1 f(xdu(xd and ~ E?=l U(Xi) converge to f3Ep(x)f(x) and f3 respectively, which yields j -t Ep(x)f(x) (generally [4]). It will always be possible to apply this extended approach below, but we drop it for now. Importance sampling is an effective estimation technique when Q approximates P over most of the domain, but it fails when Q misses high probability regions of P and systematically yields samples with small weights. In this case, the reSUlting estimator will have high variance because the sample will almost always contain unrepresentative points but is sometimes dominated by a few high weight points. To overcome this problem it is critical to obtain data points from the important regions of P. Our goal is to avoid generating systematically under-weight samples by explicitly searching for significant regions in the target distribution P. To do this, and maintain the unbiased ness of the resulting procedure, we develop a series of extensions to importance sampling that are each provably correct. The first extension is to consider sampling blocks of points instead of just individual points. Let B be a partition of X into finite blocks B, where UBE8 B = X, B n B' = 0, and each B is finite. (Note that B can be infinite.) The "block" sampling procedure (Figure 2) draws independent blocks of points to construct the final sample, but then weights points by their target probability P(x) divided by the total block probability Q(B (x)), For discrete spaces it is easy to verify that this procedure yields unbiased estimates, since Eq(x) [EXjEB(X) f(xj)w(Xj)] EBE8 EXiEB EBE8 [EXjEB [EXjEB = EXEX f(xj)w(Xj)] Q(Xi) = f(xj) ~?~n Q(B) [L:XjEB(X) f(xj)w(Xj)] Q(x) EBE8 [EXjEB = f(xj)w(Xj)] Q(B) = = EBE8 [EXjEB f(xj )P(Xj)] = L:xEx f(x)P(x). 598 D. Schuurmans "Block" importance sampling ? Draw Xl , ... , Xn independently from Q. ? For Xi, recover block Bi = {Xi,l, ... ,Xi,bJ. ? Create a large sample out of the blocks Xl ,1, ... , Xl ,bl , X2 ,1, ... , X2,b2' ??? , Xn,l, ... , Xn ,b" . ? Weighteachx I' ,}' byw(x.') = ",oiP(zi,j) I,} Q(z,,)' L.Jj=l ' ,J ? For a random variable, f, estimate Ep(z) f(x) by j = ~ 2:~=1 2:~~1 f(Xi ,j)W(Xi,j) . "Sliding window" importance sampling ? Draw Xl, ... , xn independently from Q. ? For Xi , recover block Bi , and let Xi ,l = Xi : - Get Xi,l 'S successors Xi,l, Xi,2, ... , Xi ,m by climbing up m - 1 steps from Xi ,l . - Get predecessorsxi,_m+l" ... ,Xi,-l , Xi,O by climbing down m - 1 steps from Xi,l . - Weight W(Xi ,j)= P(Xi ,i)/2:!=i_m+lQ (x; ,k) ? Create final sample from successor points XI , I, ... , Xl , m, X2,1 , ??. , X2 ,m, ?.. , Xn ,l, ... , Xn ,m. ? For a random variable, f, estimate Ep(z) f(x) 1 ",n ",m by f = n 6i=1 6j=1 f(Xi,j)W(Xi,j) . A Figure 2: "Block" and "sliding window" importance sampling procedures Crucially, this argument does not depend on how the partition of X is chosen. In fact, we could fix any partition, even one that depended on the target distribution P, and still obtain an unbiased procedure (so long as the partition remains fixed) . Intuitively, this works because blocks are drawn independently from Q and the weighting scheme still produces a "fair" representation of P . (Note that the results presented in this paper can all be extended to continuous spaces under mild technical restrictions. However, for the purposes of clarity we will restrict the technical presentation in this paper to the discrete case.) The second extension is to allow countably infinite blocks that each have a discrete total order . . . < Xi -1 < Xi < Xi +1 < .. . defined on their elements. This order could reflect the relative probability of Xi and X j under P, but for now we just consider it to be an arbitrary discrete order. To cope with blocks of unbounded length, we employ a "sliding window" sampling procedure that selects a contiguous sub-block of size m from within a larger selected block (Figure 2). This procedure builds each independent subsample by choosing a random point Xl from the proposal distribution Q, determining its containing block B(xt), and then climbing up m - 1 steps to obtain the successors Xl, X2, ??. , X m , and climbing down m - 1 steps to obtain the predecessors X- m +1 , ... , X-I, Xo . The successor points (including Xl) appear in the final sample, but the predecessors are only used to determine the weights of the sample points. Weights are determined by the target probability P (x) divided by the probability that the point X appears in a random reconstruction under Q. This too yields an unbiased estimator [2:~lf(xj)w(Xj)] = 2: XtEX [2:~;~-1 f(xj)2:~=j~~:: Q(Xk)] Q(Xl ) = '" 2: 2: l + m - 1 f(xj)P(Xj)Q(xd - ' " 2: 2: j f(xj)P(Xj)Q(xt} 6BEB x t EB j=l "'J. Q(Xk) - 6BEB Xj EB l=j-m+1 "'J . Q(Xk) 6k=J-m+l 6k=J-m+l 2:BEB2: x j EBf(xj )P(Xj )Ei::=::: ~;::~ = 2:BEB2:xjEBf(xj )P(Xj)= 2: xEx f(x)P(x). sinceEQ(x) (The middle line breaks the sum into disjoint blocks and then reorders the sum so that instead of first choosing the start point Xl and then XL'S successors Xl, .. ? , Xl+m-l. we first choose the successor point Xj and then the start points Xj-m+1 , ... , Xj that could have led to Xj). Note that this derivation does not depend on the particular block partition nor on the particular discrete orderings, so long as they remain fixed. This means that, again, we can use partitions and orderings that explicitly depend on P and still obtain a correct procedure. 599 Greedy Importance Sampling "Greedy" importance sampling (I-D) e Draw Xl , ... , Xn independently from Q. eForeachxi , letxi, l =Xi : - Compute successors Xi ,l, Xi ,2, ... ,Xi,m by taking m - 1 size ? steps in the direction of increase. - Compute predecessors Xi,-m+l, ... ,Xi ,-l ,Xi ,Oby taking m -1 size ? steps in the direction of decrease. - If an improper ascent or descent occurs, truncate paths as shown on the upper right. - Weightw(xi,j) = P(Xi ,j)/L:~=j_m+l Q(Xi ,k) . e Create the final sample from successor points collision If(x)P(x)1 k x? x" ~ if 6 "\, XI ,l, ??? , XI ,m , X'2 , I , ??? , X2 ,m , ??? ,Xn ,l, ??? , ::tn ,m- e For a random variable, f, estimate Ep(z) f(x) A by f ",m = -;;1 ",n L..ti=l L.Jj=l f(Xi ,j)W(Xi ,j) . merge Figure 3: "Greedy" importance sampling procedure; "colliding" and "merging" paths. 3 Greedy importance sampling: I-dimensional case Finally, we apply the sliding window procedure to conduct an explicit search for important regions in X. It is well known that the optimal proposal distribution for importance sampling isjust Q* (x) = If(x )P(x)11 EXEX If(x )P(x) 1 (which minimizes variance [2]). Here we apply the sliding window procedure using an order structure that is determined by the objective If(x )P(x )1 . The hope is to obtain reduced variance by sampling independent blocks of points where each block (by virtue of being constructed via an explicit search) is likely to contain at least one or two high weight points. That is, by capturing a moderate size sample of independent high weight points we intuitively expect to outperform standard methods that are unlikely to observe such points by chance. Our experiments below verify this intuition (Figure 4). The main technical issue is maintaining unbiasedness, which is easy to establish in the 1dimensional case. In the simple I-d setting, the "greedy" importance sampling procedure (Figure 3) first draws an initial point Xl from Q and then follows the direction of increasing If(x)P(x)l, taking fixed size ? steps, until either m - 1 steps have been taken or we encounter a critical point. A single "block" in our final sample is comprised of a complete sequence captured in one ascending search. To weight the sample points we account for all possible ways each point could appear in a subsample, which, as before, entails climbing down m-l steps in the descent direction (to calculate the denominators). The unbiasedness of the procedure then follows directly from the previous section, since greedy importance sampling is equivalent to sliding window importance sampling in this setting. The only nontrivial issue is to maintain disjoint search paths. Note that a search path must terminate whenever it steps from a point x? to a point x** with lower value; this indicates that a collision has occurred because some other path must reach x? from the "other side" of the critical point (Figure 3). At a collision, the largest ascent point x? must be allocated to a single path. A reasonable policy is to allocate x? to the path that has the lowest weight penultimate point (but the only critical issue is ensuring that it gets assigned to a single block). By ensuring that the critical point is included in only one of the two distinct search paths, a practical estimator can be obtained that exhibits no bias (Figure 4). To test the effectiveness of the greedy approach I conducted several I-dimensional experiments which varied the relationship between P, Q and the random variable f (Figure 4). In 600 D. Schuurmans these experiments greedy importance sampling strongly outperformed standard methods, including regular importance sampling and directly sampling from the target distribution P (rejection sampling and Metropolis sampling were not competitive). The results not only verify the unbiasedness of the greedy procedure, but also show that it obtains significantly smaller variances across a wide range of conditions. Note that the greedy procedure actually uses m out of 2m - 1 points sampled for each block and therefore effectively uses a double sample. However, Figure 4 shows that the greedy approach often obtains variance reductions that are far greater than 2 (which corresponds to a standard deviation reduction of V2). 4 Multi-dimensional case Of course, this technique is worthwhile only if it can be applied to multi-dimensional problems. In principle, it is straightforward to apply the greedy procedure of Section 3 to multi-dimensional sample spaces. The only new issue is that discrete search paths can now possibly "merge" as well as "collide"; see Figure 3. (Recall that paths could not merge in the previous case.) Therefore, instead of decomposing the domain into a collection of disjoint search paths, the objective If(x)P(x)1 now decomposes the domain into a forest of disjoint search trees. However, the same principle could be used to devise an unbiased estimator in this case: one could assign a weight to a sample point x that is just its target probability P (x) divided by the total Q-probability of the subtree of points that lead to x in fewer than m steps. This weighting scheme can be shown to yield an unbiased estimator as before. However, the resulting procedure is impractical because in an N-dimensional sample space a search tree will typically have a branching factor of n(N); yielding exponentially large trees. Avoiding the need to exhaustively examine such trees is the critical issue in applying the greedy approach to multi-dimensional spaces. The simplest conceivable strategy is just to ignore merge events. Surprisingly, this turns out to work reasonably well in many circumstances. Note that merges will be a measure zero event in many continuous domains. In such cases one could hope to ignore merges and trust that the probability of "double counting" such points would remain near zero. I conducted simple experiments with a version of greedy importance sampling procedure that ignored merges. This procedure searched in the gradient ascent direction of the objective If{x)p{x)1 and heuristically inverted search steps by climbing in the gradient descent direction. Figures 5 and 6 show that, despite the heuristic nature of this procedure, it nevertheless demonstrates credible performance on simple tasks. The first experiment is a simple demonstration from [12, 10] where the task is to sample from a bivariate Gaussian distribution P of two highly correlated random variables using a "weak" proposal distribution Q that is standard normal (depicted by the elliptical and circular one standard deviation contours in Figure 5 respectively). Greedy importance sampling once again performs very well (Figure 5); achieving unbiased estimates with lower variance than standard Monte Carlo estimators, including common MCMC methods. To conduct a more significant study, I applied the heuristic greedy method to an inference problem in graphical models: recovering the hidden state sequence from a dynamic probabilistic model, given a sequence of observations. Here I considered a simple Kalman filter model which had one state variable and one observation variable per time-step, and used theconditionaldistributionsXtlXt _ l "-' N(Xt_l,O';), ZtlXt "" N(xt,O'~) and initial distribution Xl "-' N(O,O';) . The problem was to infer the value of the final state variable Xt given the observations Zl, Z2, "', Zt. Figure 6 again demonstrates that the greedy approach Greedy Importance Sampling 601 has a strong advantage over standard importance sampling. (In fact, the greedy approach can be applied to "condensation" [6, 8] to obtain further improvements on this task, but space bounds preclude a detailed discussion.) Overall, these preliminary results show that despite the heuristic choices made in this section, the greedy strategy still performs well relative to common Monte Carlo estimators, both in terms of bias and variance (at least on some low and moderate dimension problems). However, the heuristic nature of this procedure makes it extremely unsatisfying. In fact, merge points can easily make up a significant fraction of finite domains. It turns out that a rigorously unbiased and feasible procedure can be obtained as follows. First, take greedy fixed size steps in axis parallel directions (which ensures the steps can be inverted). Then, rather than exhaustively explore an entire predecessor tree to calculate the weights of a sample point, use the well known technique of Knuth [9] to sample a single path from the root and obtain an unbiased estimate of the total Q-probability of the tree. This procedure allows one to formulate an asymptotically unbiased estimator that is nevertheless feasible to implement. It remains important future work to investigate this approach and compare it to other Monte Carlo estimation methods on large dimensional problems-in particular hybrid Monte Carlo [11, 12]. The current results already suggest that the method could have benefits. References [1] P. Dagum and M. Luby. Approximating probabilistic inference in Bayesian belief networks is NP-hard. Artif Intell, 60: 141-153, 1993. [2] M. Evans. Chaining via annealing. Ann Statist, 19:382-393, 1991. [3] B. Frey. Graphical Models for Machine Learning and Digital Communication. MIT Press, Cambridge, MA, 1998. [4] J. Geweke. Baysian inference in econometric models using Monte Carlo integration. Econometrica, 57:1317-1339, 1989. [5] W. Gilks, S. Richardson, and D. Spiegelhalter. Markov chain Monte Carlo in practice. Chapman and Hall, 1996. [6] M. Isard and A. Blake. Coutour tracking by stochastic propagation of conditional density. In ECCV, 1996. [7] M. Jordan, Z. Ghahramani, T. Jaakkola, and L. Saul. An introduction to variational methods for graphical models. In Learning in Graphical Models. Kluwer, 1998. [8] K. Kanazawa, D. Koller, and S. Russell. Stochastic simulation algorithms for dynamic probabilistic networks. In UAl, 1995. [9] D. Knuth. Estimating the efficiency of backtracking algorithms. Math. Comput., 29(129): 121136,1975. [10] D. MacKay. Intro to Monte Carlo methods. In Learning in Graphical Models. Kluwer, 1998. [11] R. Neal. Probabilistic inference using Markov chain Monte Carlo methods. 1993. [12] R. Neal. Bayesian Learning for Neural Networks . Springer, New York, 1996. [13] J. Pearl. Probabilistic Reasoning in Intelligence Systems. Morgan Kaufmann, 1988. [14] R. Shacter and M. Peot. Simulation approaches to general probabilistic inference in belief networks. In Uncertainty in Artificial Intelligence 5. Elsevier, 1990. [15] M. Tanner. Tools for statistical inference: Methods for exploration of posterior distributions and likelihoodfunctions. Springer, New York, 1993. 602 D. Schuurmans . f' ._.11 ? Direc 0.779 0.001 0.071 mean bias stdev Greed Imprt 0.777 0.003 0.065 0.781 0.001 0.Q38 Direc Greed 1m rt Direc Greed 1.038 0.002 0.088 1.044 0.003 0.049 1.032 0.008 0.475 0.258 0.049 0.838 0.208 0.000 0.010 Imrt 0.209 0.001 0.095 . .. {/... Direc 6.024 0.001 0.069 Greed Imprt 6.028 0.004 0.037 6.033 0.009 0.094 Figure 4: I-dimensional experiments: 1000 repetitions on estimation samples of size 100. Problems with varying relationships between P, Q, I and IIPI. / / .. :~:. .. .. .... .. ., 'I:'::'~" ,1 :.~ mean bias stdev Direct Greedy Importance Rejection 0.1884 0.0022 0.07 0.1937 0.0075 0.1374 0.1810 0.0052 0.1762 0.1506 0.0356 0.2868 Gibbs 0.3609 0.1747 0.5464 Metropolis 8.3609 8.1747 22.1212 Figure 5: 2-dimensional experiments: 500 repetitions on estimation samples of size 200. Pictures depict: direct, greedy importance, regular importance, and Gibbs sampling, showing 1 standard deviation countours (dots are sample points, vertical lines are weights). mean bias stdev Importance Greedy 5.2269 2.7731 1.2107 6.9236 1.0764 0.1079 Figure 6: A 6-dimensional experiment: 500 repetitions on estimation samples of size 200. Estimating the value of Xt given the observations Zl, "" Zt. Pictures depict paths sampled by regular versus greedy importance sampling.
1691 |@word mild:1 version:1 middle:1 heuristically:1 simulation:2 crucially:1 reduction:2 initial:2 series:1 elliptical:1 z2:1 current:1 must:3 evans:1 partition:6 drop:1 xex:2 depict:2 greedy:27 selected:1 fewer:1 isard:1 ebf:1 intelligence:2 xk:3 math:1 simpler:2 unbounded:1 constructed:1 direct:3 predecessor:4 prove:1 shacter:1 xt_l:1 peot:1 expected:3 nor:1 examine:1 multi:4 window:6 preclude:1 increasing:1 estimating:3 lowest:1 minimizes:1 impractical:1 ti:1 xd:2 demonstrates:3 zl:2 appear:2 before:2 local:1 frey:1 limit:1 depended:1 despite:2 path:13 merge:5 eb:2 bi:2 range:1 practical:1 gilks:1 practice:1 block:23 implement:1 lf:1 procedure:26 significantly:1 convenient:1 regular:4 suggest:1 get:3 applying:1 restriction:1 equivalent:1 deterministic:1 straightforward:1 independently:7 formulate:1 estimator:10 searching:1 variation:1 target:11 us:2 element:1 ep:8 calculate:2 region:9 ensures:1 improper:1 ordering:2 decrease:1 russell:1 intuition:1 econometrica:1 rigorously:1 exhaustively:2 dynamic:2 depend:3 efficiency:1 easily:2 collide:1 indirect:3 stdev:3 derivation:1 distinct:1 effective:1 monte:13 artificial:1 oby:1 choosing:2 heuristic:4 larger:1 drawing:1 richardson:1 final:6 advantage:2 sequence:3 reconstruction:1 achieve:1 double:2 produce:1 generating:1 develop:1 ex:2 eq:2 strong:1 implemented:1 c:1 recovering:1 direction:7 correct:3 filter:1 stochastic:4 exploration:2 successor:8 assign:1 fix:1 preliminary:1 extension:3 considered:1 hall:1 normal:1 blake:1 bj:1 purpose:1 estimation:5 outperformed:1 dagum:1 waterloo:1 grouped:1 largest:1 correctness:3 create:3 repetition:3 tool:1 weighted:2 hope:2 mit:1 always:2 gaussian:1 e7:1 beb:2 rather:1 avoid:1 varying:1 jaakkola:1 improvement:1 indicates:1 elsevier:1 inference:10 dependent:2 unlikely:1 typically:1 entire:1 hidden:1 koller:1 interested:1 selects:1 provably:1 issue:5 among:1 overall:1 ness:2 integration:1 mackay:1 construct:1 f3:1 once:1 sampling:44 chapman:1 future:1 oflarge:1 np:1 few:1 employ:1 preserve:1 intell:1 individual:1 maintain:3 attempt:1 interest:1 investigate:2 highly:1 circular:1 yielding:1 chain:4 necessary:1 conduct:2 tree:6 desired:1 contiguous:1 deviation:3 comprised:1 successful:1 conducted:2 too:1 unbiasedness:4 density:1 probabilistic:6 tanner:1 again:3 reflect:1 containing:1 choose:1 possibly:1 account:1 b2:1 unsatisfying:1 explicitly:5 break:1 root:1 start:2 recover:2 competitive:1 parallel:1 variance:8 kaufmann:1 efficiently:2 yield:6 climbing:6 weak:1 bayesian:3 carlo:13 reach:1 whenever:1 sampled:2 concentrating:1 recall:1 credible:1 geweke:1 actually:1 appears:1 higher:1 strongly:1 just:5 until:1 ei:1 trust:1 propagation:1 artif:1 contain:2 unbiased:12 true:1 verify:3 assigned:1 neal:2 branching:1 chaining:1 generalized:1 complete:1 tn:1 performs:2 reasoning:1 variational:1 common:2 empirically:1 ji:3 exponentially:1 occurred:1 approximates:1 kluwer:2 significant:5 cambridge:1 gibbs:3 had:1 dot:1 entail:1 something:1 oip:1 posterior:1 moderate:2 devise:1 inverted:2 preserving:1 captured:1 morgan:1 greater:1 converge:2 determine:1 f3p:1 sliding:6 multiple:1 infer:1 technical:3 match:1 long:2 divided:3 ensuring:2 denominator:1 circumstance:1 expectation:2 sometimes:1 achieved:1 proposal:7 condensation:1 annealing:1 allocated:1 ascent:3 effectiveness:1 jordan:1 near:1 counting:1 easy:3 xj:27 zi:1 architecture:1 restrict:1 reduce:1 idea:2 allocate:1 greed:4 york:2 jj:2 ignored:1 generally:3 useful:1 collision:3 detailed:1 statist:1 simplest:1 reduced:1 outperform:1 disjoint:4 per:1 discrete:6 nevertheless:2 achieving:1 drawn:2 clarity:1 econometric:1 asymptotically:1 fraction:1 sum:2 uncertainty:1 family:1 almost:1 reasonable:1 draw:9 investigates:1 capturing:1 bound:1 guaranteed:1 nontrivial:1 x2:6 colliding:1 dominated:1 simulate:1 argument:1 extremely:1 department:1 according:1 truncate:1 remain:2 smaller:1 across:1 metropolis:4 ztlxt:1 modification:1 intuitively:2 restricted:1 xo:1 taken:1 computationally:1 remains:2 overwhelm:1 turn:2 ascending:1 decomposing:1 apply:4 observe:1 worthwhile:1 v2:1 luby:1 alternative:1 encounter:1 graphical:8 maintaining:1 ghahramani:1 build:1 establish:2 approximating:1 bl:1 objective:3 already:1 occurs:1 intro:1 strategy:2 rt:1 exhibit:1 conceivable:1 gradient:2 exex:4 penultimate:1 assuming:1 length:1 kalman:1 relationship:2 demonstration:1 difficult:1 zt:2 policy:1 perform:1 upper:1 vertical:1 observation:5 markov:4 finite:3 descent:3 extended:2 communication:1 varied:1 arbitrary:1 cast:1 baysian:1 merges:3 pearl:1 usually:1 below:2 biasing:1 challenge:1 including:3 belief:2 event:3 critical:6 ual:1 hybrid:3 indicator:1 scheme:2 spiegelhalter:1 picture:2 axis:1 determining:2 asymptotic:2 relative:2 expect:1 versus:1 digital:1 principle:2 systematically:2 eccv:1 course:1 surprisingly:1 side:1 allow:1 bias:5 wide:1 saul:1 taking:3 benefit:1 overcome:1 dimension:1 xn:10 contour:1 dale:2 collection:1 made:1 far:1 cope:1 approximate:1 obtains:2 ignore:2 countably:1 reorder:1 xi:49 subsequence:1 search:16 continuous:2 decomposes:1 terminate:1 reasonably:1 nature:2 ca:1 schuurmans:4 forest:1 complex:1 domain:5 main:2 uwaterloo:1 subsample:2 fair:3 fails:1 sub:1 explicit:2 lq:1 xl:19 comput:1 weighting:3 down:3 xt:5 showing:1 virtue:1 bivariate:1 kanazawa:1 merging:1 effectively:1 importance:34 knuth:2 subtree:1 direc:4 rejection:3 depicted:1 led:1 backtracking:1 likely:1 explore:1 tracking:1 springer:2 corresponds:1 chance:1 ma:1 conditional:1 goal:2 presentation:1 ann:1 towards:1 unrepresentative:1 feasible:2 hard:2 included:1 infinite:2 determined:2 miss:1 total:4 searched:1 evaluate:1 mcmc:3 avoiding:1 correlated:1
756
1,692
Lower Bounds on the Complexity of Approximating Continuous Functions by Sigmoidal Neural Networks Michael Schmitt Lehrstuhl Mathematik und Informatik FakuWit ftir Mathematik Ruhr-Universitat Bochum D-44780 Bochum, Germany mschmitt@lmi.ruhr-uni-bochum.de Abstract We calculate lower bounds on the size of sigmoidal neural networks that approximate continuous functions. In particular, we show that for the approximation of polynomials the network size has to grow as O((logk)1/4) where k is the degree of the polynomials. This bound is valid for any input dimension, i.e. independently of the number of variables. The result is obtained by introducing a new method employing upper bounds on the Vapnik-Chervonenkis dimension for proving lower bounds on the size of networks that approximate continuous functions. 1 Introduction Sigmoidal neural networks are known to be universal approximators. This is one of the theoretical results most frequently cited to justify the use of sigmoidal neural networks in applications. By this statement one refers to the fact that sigmoidal neural networks have been shown to be able to approximate any continuous function arbitrarily well. Numerous results in the literature have established variants of this universal approximation property by considering distinct function classes to be approximated by network architectures using different types of neural activation functions with respect to various approximation criteria, see for instance [1, 2, 3, 5, 6, 11, 12, 14, 15]. (See in particular Scarselli and Tsoi [15] for a recent survey and further references.) All these results and many others not referenced here, some of them being constructive, some being merely existence proofs, provide upper bounds for the network size asserting that good approximation is possible if there are sufficiently many network nodes available. This, however, is only a partial answer to the question that mainly arises in practical applications: "Given some function, how many network nodes are needed to approximate it?" Not much attention has been focused on establishing lower bounds on the network size and, in particular, for the approximation of functions over the reals. As far as the computation of binary-valued Complexity ofApproximating Continuous Functions by Neural Networks 329 functions by sigmoidal networks is concerned (where the output value of a network is thresholded to yield 0 or 1) there are a few results in this direction. For a specific Boolean function Koiran [9] showed that networks using the standard sigmoid u(y) = 1/(1 + e- Y ) as activation function must have size O(nl/4) where n is the number of inputs. (When measuring network size we do not count the input nodes here and in what follows.) Maass [13] established a larger lower bound by constructing a binary-valued function over IRn and showing that standard sigmoidal networks require O(n) many network nodes for computing this function. The first work on the complexity of sigmoidal networks for approximating continuous functions is due to DasGupta and Schnitger [4]. They showed that the standard sigmoid in network nodes can be replaced by other types of activation functions without increasing the size of the network by more than a polynomial. This yields indirect lower bounds for the size of sigmoidal networks in terms of other network types. DasGupta and Schnitger [4] also claimed the size bound AO(I/d) for sigmoidal networks with d layers approximating the function sin(Ax). In this paper we consider the problem of using the standard sigmoid u(y) = 1/(1 + e- Y ) in neural networks for the approximation of polynomials. We show that at least O?logk)1/4) network nodes are required to approximate polynomials of degree k with small error in the loo norm. This bound is valid for arbitrary input dimension, i.e., it does not depend on the number of variables. (Lower bounds can also be obtained from the results on binary-valued functions mentioned above by interpolating the corresponding functions by polynomials. This, however, requires growing input dimension and does not yield a lower bound in terms of the degree.) Further, the bound established here holds for networks of any number of layers. As far as we know this is the first lower bound result for the approximation of polynomials. From the computational point of view this is a very simple class of functions; they can be computed using the basic operations addition and multiplication only. Polynomials also play an important role in approximation theory since they are dense in the class of continuous functions and some approximation results for neural networks rely on the approximability of polynomials by sigmoidal networks (see, e.g., [2, 15]). We obtain the result by introducing a new method that employs upper bounds on the Vapnik-Chervonenkis dimension of neural networks to establish lower bounds on the network size. The first use of the Vapnik-Chervonenkis dimension to obtain a lower bound is due to Koiran [9] who calculated the above-mentioned bound on the size of sigmoidal networks for a Boolean function. Koiran's method was further developed and extended by Maass [13] using a similar argument but another combinatorial dimension. Both papers derived lower bounds for the computation of binary-valued functions (Koiran [9] for inputs from {O, 1}n, Maass [13] for inputs from IRn). Here, we present a new technique to show that and how lower bounds can be obtained for networks that approximate continuous functions. It rests on two fundamental results about the Vapnik-Chervonenkis dimension of neural networks. On the one hand, we use constructions provided by Koiran and Sontag [10] to build networks that have large Vapnik-Chervonenkis dimension and consist of gates that compute certain arithmetic functions. On the other hand, we follow the lines of reasoning of Karpinski and Macintyre [7] to derive an upper bound for the VapnikChervonenkis dimension of these networks from the estimates of Khovanskil [8] and a result due to Warren [16]. In the following section we give the definitions of sigmoidal networks and the VapnikChervonenkis dimension. Then we present the lower bound result for function approximation. Finally, we conclude with some discussion and open questions. 330 2 M Schmitt Sigmoidal Neural Networks and VC Dimension We briefly recall the definitions of a sigmoidal neural network and the VapnikChervonenkis dimension (see, e.g., [7, 10]). We consider /eed/orward neural networks which have a certain number of input nodes and one output node. The nodes which are not input nodes are called computation nodes and associated with each of them is a real number t, the threshold. Further, each edge is labelled with a real number W called weight. Computation in the network takes place as follows: The input values are assigned to the input nodes. Each computation node applies the standard sigmoid u(y) = 1/(1 + e- V ) to the sum W1Xl + ... + WrXr - t where Xl, .?. ,X r are the values computed by the node's predecessors, WI, ??? ,W r are the weights of the corresponding edges, and t is the threshold. The output value of the network is defined to be the value computed by the output node. As it is common for approximation results by means of neural networks, we assume that the output node is a linear gate, i.e., it just outputs the sum WIXI + ... + WrXr - t. (Clearly, for computing functions on finite sets with output range [0, 1] the output node may apply the standard sigmoid as well.) Since u is the only sigmoidal function that we consider here we will refer to such networks as sigmoidal neural networks. (Sigmoidal functions in general need to satisfy much weaker assumptions than u does.) The definition naturally generalizes to networks employing other types of gates that we will make use of (e.g. linear, multiplication, and division gates). The Vapnik-Chervonenkis dimension is a combinatorial dimension of a function class and is defined as follows: A dichotomy of a set S ~ IRn is a partition of S into two disjoint subsets (So, Sl) such that So U SI = S. Given a set F offunctions mapping IRn to {O, I} and a dichotomy (So, Sd of S, we say that F induces the dichotomy (So, Sd on S if there is some f E F such that /(So) ~ {O} and f(Sd ~ {I}. We say further that F shatters S if F induces all dichotomies on S. The VapnikChervonenkis (VC) dimension of F, denoted VCdim(F), is defined as the largest number m such that there is a set of m elements that is shattered by F. We refer to the VC dimension of a neural network, which is given in terms of a "feedforward architecture", i.e. a directed acyclic graph, as the VC dimension of the class of functions obtained by assigning real numbers to all its programmable parameters, which are in general the weights and thresholds of the network or a subset thereof. Further, we assume that the output value of the network is thresholded at 1/2 to obtain binary values. 3 Lower Bounds on Network Size Before we present the lower bound on the size of sigmoidal networks required for the approximation of polynomials we first give a brief outline of the proof idea. We will define a sequence of univariate polynomials (Pn)n>l by means of which we show how to construct neural architectures N n consistmg of various types of gates such as linear, multiplication, and division gates, and, in particular, gates that compute some of the polynomials. Further, this architecture has a single weight as programmable parameter (all other weights and thresholds are fixed). We then demonstrate that, assuming the gates computing the polynomials can be approximated by sigmoidal neural networks sufficiently well, the architecture Nn can shatter a certain set by assigning suitable values to its programmable weight. The final step is to reason along the lines of Karpinski and Macintyre [7] to obtain via Khovanskil's estimates [8] and Warren's result [16] an upper bound on the VC dimension of N n in terms of the number of its computation nodes. (Note that we cannot directly apply Theorem 7 of [7] since it does not deal with division gates.) Comparing this bound with the cardinality of the shattered set we will then be able 331 Complexity ofApproximating Continuous Functions by Neural Networks (3) W 1 n P3 (1) (2) W1 W1 (3) Wi (3) W1 n P2 (2) Wj (1) Wk (1) (2) Wn Wn n P1 Wn j --------------------------------~ k--------------------------------------------------~ Figure 1: The network N n with values k, j, i, 1 assigned to the input nodes Xl, X2, X3, X4 respectively. The weight W is the only programmable parameter of the network. to conclude with a lower bound on the number of computation nodes in N n and thus in the networks that approximate the polynomials. Let the sequence (Pn)n2: l of polynomials over IR be inductively defined by Pn(X) = { 4x(1 - x) P(Pn-dx)) n = 1, n 2:: 2 . Clearly, this uniquely defines Pn for every n 2:: 1 and it can readily be seen that Pn has degree 2n. The main lower bound result is made precise in the following statement. Theorem 1 Sigmoidal neural networks that approximate the polynomials (Pn)n >l on the interval [0,1] with error at most O(2- n ) in the 100 norm must have at least n(nl/4) computation nodes. Proof. For each n a neural architecture N n can be constructed as follows: The network has four input nodes Xl, X2, X3, X4. Figure 1 shows the network with input values assigned to the input nodes in the order X4 = 1, X3 = i, X2 = j, Xl = k. There is one weight which we consider as the (only) programmable parameter of N n . It is associated with the edge outgoing from input node X4 and is denoted by w. The computation nodes are partitioned into six levels as indicated by the boxes in Figure 1. Each level is itself a network. Let us first assume, for the sake of simplicity, that all computations over real numbers are exact. There are three levels labeled with II, having n + 1 input nodes and one output node each, that compute so-called projections 7r : IRnH -+ IR where 7r(YI,"" Yn, a) = Ya for a E {I, ... , n}. The levels labeled P3 , P2 , PI have one input node and n output nodes each. Level P3 receives the constant 1 as input and thus the value W which is the parameter of the network. We define the output values of level P A for>. = 3,2, 1 by (A) wb = Pbon"'-l ( v) , b= 1, ... ,n where v denotes the input value to level P A. This value is equal to w for>. = 3 and (A+l) , .?. , Wn()..+l) ,XA+l ) oth erWlse. . OUT (A) can b id vve observe t h at wb+l e calcu ate f rom 7r (WI 332 M Schmitt w~A) as Pn>'_l(W~A?). Therefore, the computations of level P A can be implemented using n gates each of them computing the function Pn>.-l. We show now that Nn can shatter a set of cardinality n 3 ? Let S = {I, ... ,n p. It has been shown in Lemma 2 of [10] that for each (/31 , ... , /3r) E {O, 1Y there exists some W E [0,1] such that for q = 1, ... ,T pq(w) E [0,1/2) if /3q = 0, and pq(w) E (1/2,1] if /3q = 1. This implies that, for each dichotomy (So, Sd of S there is some that for every (i, j, k) E S Pk (pj.n (Pi.n 2(w))) Pk(Pj.n(Pi.n2(w))) < 1/2 > 1/2 if if W E [0,1] such (i, j, k) E So , (i,j,k)ES1' Note that Pk(Pj.n(Pi.n2 (w))) is the value computed by N n given input values k, j, i, 1. Therefore, choosing a suitable value for w, which is the parameter of Nn , the network can induce any dichotomy on S. In other words, S is shattered by Nn . An such that for each E > weights can be chosen for An such that the function in,? computed by this network satisfies lim?~o in,?(Yl, ... ,Yn, a) = Ya. Moreover, this architecture consists of O(n) computation nodes, which are linear, multiplication, and division gates. (Note that the size of An does not depend on E.) Therefore, choosing E sufficiently small, we can implement the projections 1r in N n by networks of O(n) computation nodes such that the resulting network N~ still shatters S. Now in N~ we have O(n) computation nodes for implementing the three levels labeled II and we have in each level P A a number of O(n) computation nodes for computing Pn>.-l, respectively. Assume now that the computation nodes for Pn>.-l can be replaced by sigmoidal networks such that on inputs from S and with the parameter values defined above the resulting network N:: computes the same functions as N~. (Note that the computation nodes for Pn>.-l have no programmable parameters.) ? It has been shown in Lemma 1 of [10] that there is an architecture N::. We estimate the size of According to Theorem 7 of Karpinski and Macintyre [7] a sigmoidal neural network with I programmable parameters and m computation nodes has VC dimension O((ml)2). We have to generalize this result slightly before being able to apply it. It can readily be seen from the proof of Theorem 7 in [7] that the result also holds if the network additionally contains linear and multiplication gates. For division gates we can derive the same bound taking into account that for a gate computing division, say x/y, we can introduce a defining equality x = z . Y where z is a new variable. (See [7] for how to proceed.) Thus, we have that a network with I programmable parameters and m computation nodes, which are linear, multiplication, division, and sigmoidal gates, has VC dimension O((ml)2). In particular, if m is the number of computation nodes of N::, the VC dimension can shatter a set is O(m 2 ). On the other hand, as we have shown above, of cardinality n 3 ? Since there are O(n) sigmoidal networks in computing the functions Pn>.-l, and since the number of linear, multiplication, and division gates is bounded by O(n), for some value of A a single network computing Pn>.-l must have size at least O(fo). This yields a lower bound of O(nl/4) for the size of a sigmoidal network computing Pn. N:: N:: Thus far, we have assumed that the polynomials Pn are computed exactly. Since polynomials are continuous functions and since we require them to be calculated only on a finite set of input values (those resulting from S and from the parameter values chosen for w to shatter S) an approximation of these polynomials is sufficient. A straightforward analysis, based on the fact that the output value of the network has a "tolerance" close to 1/2, shows that if Pn is approximated with error O(2- n ) Complexity ofApproximating Continuous Functions by Neural Networks 333 in the loo norm, the resulting network still shatters the set S. This completes the proof of the theorem. D The statement of the previous theorem is restricted to the approximation of polynomials on the input domain [0,1]. However, the result immediately generalizes to any arbitrary interval in llt Moreover, it remains valid for multivariate polynomials of arbitrary input dimension. Corollary 2 The approximation of polynomials of degree k by sigmoidal neural networks with approximation error O(ljk) in the 100 norm requires networks of size O((log k)1/4). This holds for polynomials over any number of variables. 4 Conclusions and Open Questions We have established lower bounds on the size of sigmoidal networks for the approximation of continuous functions. In particular, for a concrete class of polynomials we have calculated a lower bound in terms of the degree of the polynomials. The main result already holds for the approximation of univariate polynomials. Intuitively, approximation of multivariate polynomials seems to become harder when the dimension increases. Therefore, it would be interesting to have lower bounds both in terms of the degree and the input dimension. Further, in our result the approximation error and the degree are coupled. Naturally, one would expect that the number of nodes has to grow for each fixed function when the error decreases. At present we do not know of any such lower bound. We have not aimed at calculating the constants in the bounds. For practical applications such values are indispensable. Refining our method and using tighter results it should be straightforward to obtain such numbers. Further, we expect that better lower bounds can be obtained by considering networks of restricted depth. To establish the result we have introduced a new method for deriving lower bounds on network sizes. One of the main arguments is to use the functions to be approximated to construct networks with large VC dimension. The method seems suitable to obtain bounds also for the approximation of other types of functions as long as they are computationally powerful enough. Moreover, the method could be adapted to obtain lower bounds also for networks using other activation functions (e.g. more general sigmoidal functions, ridge functions, radial basis functions). This may lead to new separation results for the approximation capabilities of different types of neural networks. In order for this to be accomplished, however, an essential requirement is that small upper bounds can be calculated for the VC dimension of such networks. Acknowledgments I thank Hans U. Simon for helpful discussions. This work was supported in part by the ESPRIT Working Group in Neural and Computational Learning II, NeuroCOLT2, No. 27150. References [1] A. Barron. Universal approximation bounds for superposition of a sigmoidal function. IEEE Transactions on Information Theory, 39:930--945, 1993. 334 M Schmitt [2J C. K. Chui and X. Li. Approximation by ridge functions and neural networks with one hidden layer. Journal of Approximation Theory, 70:131-141,1992. [3J G. Cybenko. Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals, and Systems, 2:303-314, 1989. [4J B. DasGupta and G. Schnitger. The power of approximating: A comparison of activation functions. In C. L. Giles, S. J. Hanson, and J. D. Cowan, editors, Advances in Neural Information Processing Systems 5, pages 615-622, Morgan Kaufmann, San Mateo, CA, 1993. [5] K. Hornik. Approximation capabilities of multilayer feedforward networks. Neural Networks, 4:251-257, 1991. [6] K. Hornik, M. Stinchcombe, and H. White. Multilayer feedforward networks are universal approximators. Neural Networks, 2:359-366, 1989. [7] M. Karpinski and A. Macintyre. Polynomial bounds for VC dimension of sigmoidal and general Pfaffian neural networks. Journal of Computer and System Sciences, 54:169-176, 1997. [8] A. G. Khovanskil. Fewnomials, volume 88 of Translations of Mathematical Monographs. American Mathematical Society, Providence, RI, 1991. [9] P. Koiran. VC dimension in circuit complexity. In Proceedings of the 11th Annual IEEE Conference on Computational Complexity CCC'96, pages 81-85, IEEE Computer Society Press, Los Alamitos, CA, 1996. [10] P. Koiran and E. D. Sontag. Neural networks with quadratic VC dimension. Journal of Computer and System Sciences, 54:190-198, 1997. [11] V. Y. Kreinovich. Arbitrary nonlinearity is sufficient to represent all functions by neural networks: A theorem. Neural Networks, 4:381-383, 1991. [12] M. Leshno, V. Y. Lin, A. Pinkus, and S. Schocken. Multilayer feedforward networks with a nonpolynomial activation function can approximate any function. Neural Networks, 6:861-867, 1993. [13] W. Maass. Noisy spiking neurons with temporal coding have more computational power than sigmoidal neurons. In M. Mozer, M. 1. Jordan, and T. Petsche, editors, Advances in Neural Information Processing Systems 9, pages 211-217. MIT Press, Cambridge, MA, 1997. [14] H. Mhaskar. Neural networks for optimal approximation of smooth and analytic functions. Neural Computation, 8:164-177, 1996. [15J F. Scarselli and A. C. Tsoi. Universal approximation using feedforward neural networks: A survey of some existing methods and some new results. Neural Networks, 11:15-37, 1998. [16] H. E. Warren. Lower bounds for approximation by nonlinear manifolds. Transactions of the American Mathematical Society, 133:167-178, 1968.
1692 |@word briefly:1 polynomial:28 norm:4 seems:2 open:2 ruhr:2 harder:1 contains:1 chervonenkis:6 existing:1 comparing:1 activation:6 schnitger:3 si:1 must:3 assigning:2 dx:1 readily:2 partition:1 offunctions:1 analytic:1 node:39 sigmoidal:33 mathematical:3 shatter:4 along:1 constructed:1 predecessor:1 become:1 consists:1 introduce:1 p1:1 frequently:1 growing:1 leshno:1 considering:2 increasing:1 cardinality:3 provided:1 moreover:3 bounded:1 circuit:1 what:1 developed:1 temporal:1 every:2 exactly:1 esprit:1 control:1 yn:2 before:2 referenced:1 sd:4 id:1 establishing:1 mateo:1 range:1 directed:1 practical:2 acknowledgment:1 tsoi:2 implement:1 x3:3 universal:5 projection:2 word:1 induce:1 refers:1 radial:1 bochum:3 cannot:1 close:1 straightforward:2 attention:1 independently:1 survey:2 focused:1 simplicity:1 immediately:1 deriving:1 proving:1 construction:1 play:1 exact:1 element:1 approximated:4 labeled:3 role:1 calculate:1 wj:1 decrease:1 mentioned:2 monograph:1 und:1 mozer:1 complexity:7 inductively:1 depend:2 division:8 basis:1 indirect:1 various:2 distinct:1 dichotomy:6 choosing:2 larger:1 valued:4 schocken:1 say:3 itself:1 noisy:1 final:1 sequence:2 los:1 requirement:1 derive:2 p2:2 implemented:1 implies:1 direction:1 vc:13 implementing:1 require:2 vcdim:1 ao:1 cybenko:1 tighter:1 hold:4 sufficiently:3 mapping:1 koiran:7 combinatorial:2 superposition:2 largest:1 mit:1 clearly:2 pn:17 corollary:1 ax:1 derived:1 refining:1 mainly:1 vapnikchervonenkis:4 helpful:1 nn:4 shattered:3 hidden:1 irn:4 germany:1 denoted:2 equal:1 construct:2 having:1 x4:4 others:1 employ:1 few:1 scarselli:2 replaced:2 nl:3 oth:1 edge:3 partial:1 theoretical:1 instance:1 boolean:2 wb:2 giles:1 measuring:1 introducing:2 subset:2 loo:2 universitat:1 providence:1 answer:1 cited:1 fundamental:1 yl:1 michael:1 concrete:1 w1:3 lmi:1 american:2 li:1 account:1 de:1 coding:1 wk:1 satisfy:1 view:1 capability:2 simon:1 ir:2 kaufmann:1 who:1 yield:4 generalize:1 informatik:1 llt:1 fo:1 eed:1 definition:3 thereof:1 naturally:2 proof:5 associated:2 wixi:1 recall:1 lim:1 follow:1 lehrstuhl:1 box:1 just:1 xa:1 hand:3 receives:1 working:1 nonlinear:1 defines:1 indicated:1 equality:1 assigned:3 maass:4 deal:1 white:1 sin:1 uniquely:1 criterion:1 outline:1 ridge:2 demonstrate:1 reasoning:1 sigmoid:5 common:1 spiking:1 volume:1 pinkus:1 refer:2 nonpolynomial:1 cambridge:1 mathematics:1 nonlinearity:1 pq:2 han:1 multivariate:2 recent:1 showed:2 claimed:1 certain:3 indispensable:1 binary:5 arbitrarily:1 approximators:2 yi:1 accomplished:1 seen:2 morgan:1 signal:1 arithmetic:1 ii:3 mhaskar:1 smooth:1 long:1 lin:1 variant:1 basic:1 multilayer:3 karpinski:4 represent:1 addition:1 interval:2 completes:1 grow:2 rest:1 cowan:1 jordan:1 feedforward:5 enough:1 concerned:1 wn:4 architecture:8 idea:1 six:1 sontag:2 proceed:1 programmable:8 aimed:1 induces:2 macintyre:4 sl:1 disjoint:1 dasgupta:3 group:1 four:1 threshold:4 shatters:3 pj:3 thresholded:2 graph:1 merely:1 sum:2 powerful:1 place:1 p3:3 separation:1 bound:44 layer:3 quadratic:1 annual:1 adapted:1 x2:3 ri:1 sake:1 chui:1 argument:2 approximability:1 according:1 ate:1 slightly:1 wi:3 partitioned:1 intuitively:1 restricted:2 computationally:1 mathematik:2 remains:1 count:1 needed:1 know:2 available:1 operation:1 generalizes:2 apply:3 observe:1 barron:1 petsche:1 gate:16 existence:1 denotes:1 pfaffian:1 calculating:1 build:1 establish:2 approximating:4 society:3 question:3 already:1 neurocolt2:1 alamitos:1 ccc:1 thank:1 manifold:1 reason:1 rom:1 assuming:1 statement:3 upper:6 neuron:2 finite:2 defining:1 extended:1 precise:1 arbitrary:4 introduced:1 required:2 hanson:1 established:4 able:3 stinchcombe:1 power:2 suitable:3 rely:1 brief:1 numerous:1 coupled:1 literature:1 multiplication:7 expect:2 interesting:1 acyclic:1 degree:8 sufficient:2 schmitt:4 editor:2 pi:4 translation:1 supported:1 warren:3 weaker:1 taking:1 tolerance:1 calculated:4 dimension:30 valid:3 depth:1 asserting:1 computes:1 made:1 san:1 employing:2 far:3 transaction:2 approximate:9 uni:1 ml:2 conclude:2 assumed:1 continuous:12 additionally:1 erwlse:1 ca:2 hornik:2 interpolating:1 constructing:1 domain:1 pk:3 dense:1 main:3 n2:3 vve:1 xl:4 theorem:7 specific:1 showing:1 consist:1 exists:1 essential:1 vapnik:6 logk:2 univariate:2 orward:1 applies:1 satisfies:1 ma:1 ljk:1 labelled:1 justify:1 lemma:2 called:3 ya:2 arises:1 constructive:1 outgoing:1 es1:1
757
1,693
Dynamics of Supervised Learning with Restricted Training Sets and Noisy Teachers A.C.C. Coolen Dept of Mathematics King's College London The Strand, London WC2R 2LS, UK tcoolen@mth.kc1.ac.uk C.W.H.Mace Dept of Mathematics King's College London The Strand, London WC2R 2LS, UK cmace@mth.kc1.ac.uk Abstract We generalize a recent formalism to describe the dynamics of supervised learning in layered neural networks, in the regime where data recycling is inevitable, to the case of noisy teachers. Our theory generates reliable predictions for the evolution in time of training- and generalization errors, and extends the class of mathematically solvable learning processes in large neural networks to those situations where overfitting can occur. 1 Introduction Tools from statistical mechanics have been used successfully over the last decade to study the dynamics of learning in layered neural networks (for reviews see e.g. [1] or [2]). The simplest theories result upon assuming the data set to be much larger than the number of weight updates made, which rules out recycling and ensures that any distribution of relevance will be Gaussian. Unfortunately, both in terms of applications and in terms of mathematical interest, this regime is not the most relevant one. Most complications and peculiarities in the dynamics of learning arise precisely due to data recycling, which creates for the system the possibility to improve performance by memorizing answers rather than by learning an underlying rule. The dynamics of learning with restricted training sets was first studied analytically in [3] (linear learning rules) and [4] (systems with binary weights). The latter studies were ahead of their time, and did not get the attention they deserved just because at that stage even the simpler learning dynamics without data recycling had not yet been studied. More recently attention has moved back to the dynamics of learning in the recycling regime. Some studies aimed at developing a general theory [5, 6, 7], some at finding exact solutions for special cases [8]. All general theories published so far have in common that they as yet considered realizable scenario's: the rule to be learned was implementable by the student, and overfitting could not yet occur. The next hurdle is that where restricted training sets are combined with unrealizable rules. Again some have turned to non-typical but solvable cases, involving Hebbian rules and noisy [9] or 'reverse wedge' teachers [10]. More recently the cavity method has been used to build a general theory [11] (as yet for batch learning only). In this paper we generalize the general theory launched in [6,5,7], which applies to arbitrary learning rules, to the case of noisy teachers. We will mirror closely the presentation in [6] (dealing with the simpler case of noise-free teachers), and we refer to [5, 7] for background reading on the ideas behind the formalism. A. C. C. Coolen and C. W. H. Mace 238 2 Definitions As in [6, 5] we restrict ourselves for simplicity to perceptrons. A student perceptron operates a linear separation, parametrised by a weight vector J E iRN : S:{-I,I}N -t{-I,I} S(e) = sgn[J?e] It aims to emulate a teacher o~erating a similar rule, which, however, is characterized by a variable weight vector BE iR ,drawn at random from a distribution P(B) such as P(B) = >'6[B+B*] output noise: + (1->')6[B-B*] (1) P(B) = [~~/NrN e- tN (B-B')2/E2 (2) The parameters>. and ~ control the amount of teacher noise, with the noise-free teacher B = B* recovered in the limits>. -t 0 and ~ -t O. The student modifies J iteratively, using examples of input vectors which are drawn at random from a fixed (randomly composed) E {-I, I}N with a> 0, and the corresponding training set containing p = aN vectors values of the teacher outputs. We choose the teacher noise to be consistent, i.e. the answer will remain the same when that particular question given by the teacher to a question re-appears during the learning process. Thus T(e?) = sgn[BJL . e], with p teacher weight vectors BJL, drawn randomly and independently from P(B), and we generalize the training l , B l ), . .. , (e, BP)}. Consistency of teacher noise is natural set accordingly to jj = in terms of applications, and a prerequisite for overfitting phenomena. Averages over the training set will be denoted as ( ... ) b; averages over all possible input vectors E {-I, I}N as ( ... )e. We analyze two classes of learning rules, of the form J (? + 1) = J (?) + f).J (?): Gaussian weight noise: e e e He e = 11 {e(?) 9 [J(?)?e(?), B(?)?e(?)] - ,J(?) } f).J(?) = 11 {(e 9 [J(?)?e, B?eDl> - ,J(m) } on-line: f).J(?) batch : (3) In on-line learning one draws at each step ? a question/answer pair (e (?), B (?)) at random from the training set. In batch learning one iterates a deterministic map which is an average over all data in the training set. Our performance measures are the training- and generalization errors, defined as follows (with the step function O[x > 0] = 1, O[x < 0] = 0): Et(J) = (O[-(J ?e)(B ?em b Eg(J) = (O[-(J ?e)(B* ?e)])e (4) We introduce macroscopic observables, taylored to the present problem, generalizing [5, 6]: Q[J]=J 2, R[J]=J?B*, P[x,y,z;J]=(6[x-J?e]6[y-B*?e]6[z-B?eDl> (5) As in [5, 6] we eliminate technical subtleties by assuming the number of arguments (x, y, z) for which P[x, y, z; J] is evaluated to go to infinity after the limit N -t 00 has been taken. 3 Derivation of Macroscopic Laws Upon generalizing the calculations in [6, 5], one finds for on-line learning: ! ! Q = 2'f} !dXdydZ P[x, y, z] xg[x, z] - 2'f},Q + 'f}2!dXdYdZ P[x, y, z] g2[x, z] (6) R = 'f} !dXdydZ P[x, y, z] y9[x, z]- 'f},R (7) :t P[x, y, z] = ~ ! dx' P[x', y, z] {6[x-x' -'f}G[x', z]] -6[x-x']} -'f}! / dx'dy'dz' / dx'dy'dz'9[x', z]A[x, y, z; x',y', z'] 1 +'i'f}2 ! + 'f}, :x EP2P[x, y, z] dx'dy'dz' P[x', y', z']92[x', z'] 8x {xP[x , y, z]} (8) Supervised Learning with Restricted Training Sets 239 The complexity of the problem is concentrated in a Green's function: A[x, y, Zj x', y', z'] = lim N-+oo (( ([1-6ee , ]6[x-J?e]6[y-B*?e]6[z-B?e] (e?e')6[x' -J?e']6[y' - B*?e']6[y' - B?e'])i?i> )QW;t J It involves a conditional average of the form (K[J])QW;t = dJ Pt(JIQ,R,P)K[J], with Pt(J) 6[Q-Q[J]]6[R- R[J]] nXYZ 6[P[x, y, z] -P[x, y, Zj J]] Pt(JIQ,R,P) JdJ Pt(J) 6[Q - Q[J]]6[R- R[J]] nXYZ 6[P[x, y, z] - P[x, y, z; J]] = in which Pt (J) is the weight probability density at time t. The solution of (6,7,8) can be used to generate the N -+ 00 performance measures (4) at any time: Et =/ dxdydz P[x, y, z]O[-xz] Eg = 11"-1 arccos[RIVQ] (9) Expansion of these equations in powers of"" and retaining only the terms linear in "" gives the corresponding equations describing batch learning. So far this analysis is exact. 4 Closure of Macroscopic Laws As in [6, 5] we close our macroscopic laws (6,7,8) by making the two key assumptions underlying dynamical replica theory: (i) For N -+ 00 our macroscopic observables obey closed dynamic equations. (ii) These equations are self-averaging with respect to the specific realization of D. (i) implies that probability variations within {Q, R, P} subshells are either absent or irrelevant to the macroscopic laws. We may thus make the simplest choice for Pt (J IQ, R, P): Pt(JIQ,R,P) -+ 6[Q-Q[J]] 6[R-R[J]] II 6[P[x,y,z]-P[x,y,ZjJ]] (10) xyz The procedure (10) leads to exact laws if our observables {Q, R, P} indeed obey closed equations for N -+ 00. It is a maximum entropy approximation if not. (ii) allows us to average the macroscopic laws over all training sets; it is observed in simulations, and proven using the formalism of [4]. Our assumptions (10) result in the closure of (6,7,8), since now the Green's function can be written in terms of {Q, R, Pl. The final ingredient of dynamical replica theory is doing the average of fractions with the replica identity / JdJ W[JID]GIJID]) \ JdJ W[JID] = lim sets /dJ I ??? dJn (G[J 1 ID] n-+O IT W[JO<ID])sets a=1 Our problem has been reduced to calculating (non-trivial) integrals and averages. One finds that P[x, y, z] P[x, zly]P[y] with Ply] (211")-!exp[-!y 21With the short-hands Dy = P[y]dy and (f(x, y, z)) = Dydxdz P[x, zly]f(x, y, z) we can write the resulting macroscopic laws, for the case of output noise (1), in the following compact way: = d dt Q = 2",(V - ,Q) [) [)tP[x,zly] = = J + rJ2 Z d dtR = ",(W - ,R) (11) 1 [)x[)22P[x,zIY] a1/dx'P[x',zly] {6[x-x'-",G[x',z]]-6[x-x'] }+2",2Z -",:x {P[x,zly] [U(x-RY)+Wy-,x+[V-RW-(Q-R2)U]~[x,y,z])} (12) with U = (~[x, y, z]9[x, z]), v = (x9[x, z]), W = (y9[x, z]), Z = (9 2[x, z]) The solution of (12) is at any time of the following form: P[x,zly] = (1-,x)6[y-z]P+[xly] + ,x6[y+z]P-[xly] (13) A. C. C. Coolen and C. W. H. Mace 240 Finding the function <I> [x, y, z] (in replica symmetric ansatz) requires solving a saddle-point problem for a scalar observable q and two functions M?[xly]. Upon introducing B = . . :. V. .,. .q.,-Q___R,-2 Q(I-q) (with Jdx M?[xly] Jdx M?[xly]eBxs J[x, y] Jdx M?[xly]eBxs (f[x, y])? = * = 1 for all y) the saddle-point equations acquire the fonn p?[Xly] = for all X, y : ((x-Ry)2) + (qQ-R 2)[I-!:.] a ! Ds (O[X -xl); 2 !DYDS S[(I-A)(X); + A(X);] = qQ+Q-2R ..jqQ_R2 (14) (15) The equations (14) which detennine M?[xly] have the same structure as the corresponding (single) equation in [5, 6], so the proofs in [5, 6] again apply, and the solutions M?[xly], given a q in the physical range q E [R2/Q, 1], are unique. The function <I> [x, y, z] is then given by <I> [X, y, z] =! Ds s {(I-A)O[Z-y](o[X -x)); + AO[Z+Y](o[X -xl);} ..jqQ_R2 P[X, zly] (16) Working out predictions from these equations is generally CPU-intensive, mainly due to the functional saddle-point equation (14) to be solved at each time step. However, as in [7] one can construct useful approximations of the theory, with increasing complexity: (i) Large a approximation (giving the simplest theory, without saddle-point equations) (ii) Conditionally Gaussian approximation for M[xly] (with y-dependent moments) (iii) Annealed approximation of the functional saddle-point equation 5 Benchmark Tests: The Limits a --+ 00 and ,\ --+ 0 We first show that in the limit a --+ 00 our theory reduces to the simple (Q, R) formalism of infinite training sets, as worked out for noisy teachers in [12]. Upon making the ansatz p?[xly] = P[xly] = [27r(Q-R 2)]-t e- t [x- Rv]2/(Q-R 2) (17) one finds <I>[x,y,Z] = (x-Ry)/(Q-R 2) M?[xly] = P[xly], Insertion of our ansatz into (12), followed by rearranging of terms and usage of the above expression for <I> [x, y, z], shows that (12) is satisfied. The remaining equations (11) involve only averages over the Gaussian distribution (17), and indeed reduce to those of [12]: ~! Q = (I-A) { 2(x9[x, y)) 1 d --d R 1} t + 1}{92[x, y)) } + A {2(x9[x,-y)) + 1}(92[x,-y)) } - 2,Q = (I-A)(y9[x,y)) + A(y9[x,-yl) -,R Next we turn to the limit A --+ 0 (restricted training sets & noise-free teachers) and show that here our theory reproduces the fonnalism of [6,5]. Now we make the following ansatz: P+[xly] = P[xly], P[x, zly] = o[z-y]P[xIY] (18) Insertion shows that for A = 0 solutions of this fonn indeed solve our equations, giving <p[x, y, z]--+ <I> [x, y] and M+[xly] M[xly), and leaving us exactly with the fonnalism of [6, 5] describing the case of noise-free teachers and restricted training sets (apart from some new tenns due to the presence of weight decay, which was absent in [6, 5]). = 241 Supervised Learning with Restricted Training Sets 0. , r------~--__, 0..4 ~-------_____I 0..4 11>=0.' 0..3 a=4 0. , 0..0. -- , 0. 0.2 _ __ ___ _____ _ a= 1 0;=1 ------- ---- -- --- - 0. 0;=2 =-= - 0;=2 - - ----- - a=4 a=4 = =-= --=-=--=-=--=-=-=-- -=-=-_oed a=4 , 0;=2 ':::::========:::j 0..3 -- - ---- 0;=1 :::---- - -----1 0;=2 0..2 11>=0.' ~-------~ 0;=1 0., 11>=0, " , no. I 0. , 0. " Figure 1: On-line Hebbian learning: conditionally Gaussian approximation versus exact solution in [9] (.,., = 1, ,X = 0.2). Left: "I = 0.1, right: "I = 0.5. Solid lines: approximated theory, dashed lines: exact result. Upper curves: Eg as functions of time (here the two theories agree), lower curves: E t as functions of time. 6 Benchmark Tests: Hebbian Learning The special case of Hebbian learning, i.e. Q[x, z] = sgn(z), can be solved exactly at any time, for arbitrary {a, ,x, "I} [9], providing yet another excellent benchmark for our theory. For batch execution of Hebbian learning the macroscopic laws are obtained upon expanding (11,12) and retaining only those terms which are linear in.,.,. All integrations can now be done and all equations solved explicitly, resulting in U =0, Z = 1, W = (I-2,X)J2/7r, and Q = Qo e-2rryt + 2Ro(I-2'x) e-17"Yt[I_e-rrrt] "I f{ + [~(I-2,X)2+.!.] V:; 7r a [I-e- 17 "Y tF "12 R = Ro e- 17"Y t +(I-2'x)J2/7r[I-e- 17"Y t ]/"I q = [aR2+(I_e- 17"Yt)2 i'l]/aQ p?[xIY] = [27r(Q-R2)] -t e-tlz-RH sgn(y)[1-e-"..,t]/a"Y]2/(Q-R2) (19) From these results, in tum, follow the performance measures Eg = 7r- 1 arccos[ R/ JQ) and E = ! - !(1-,X)!D 2 t 2 erf[IYIR+[I-e- 77"Y t ]/a"l] + !,X!D erf[IYIR-[I-e- 17"Y t ]/a"l] Y J2(Q-R2) 2 y J2(Q-R2) Comparison with the exact solution, calculated along the lines of [9] or, equivalently, obtained upon putting t ? in [9], shows that the above expressions are all exact. .,.,-2 For on-line execution we cannot (yet) solve the functional saddle-point equation in general. However, some analytical predictions can still be extracted from (11,12,13): Q = Qo e-217"Yt + 2Ro(I-2,X) e-77"Yt[I_e-17"Yt] "I R = Ro e- 17"Y t + (I-2,X)J2/7r[I-e- 17"Y t ]/"I J f{ + [~(I-2,X)2+.!.] V:; 7r a [I_e- 17"Y t ]2 "12 + !L[I_e- 217"Y t ] 2"1 dx xP?[xIY] = Ry ? sgn(y)[I-e- 17"Y t ]/a"l with U =0, W = (I-2,X)J2/7r, V = W R+[I-e- 17"Y t ]/a"l, and Z = 1. Comparison with the results in [9] shows that the above expressions, and thus also that of E g , are all fully exact, at any time. Observables involving P[x, y, z] (including the training error) are not as easily solved from our equations. Instead we used the conditionally Gaussian approximation (found to be adequate for the noiseless Hebbian case [5, 6, 7]). The result is shown in figure 1. The agreement is reasonable, but significantly less than that in [6]; apparently teacher noise adds to the deformation of the field distribution away from a Gaussian shape. 242 A. C. C. Coolen and C. W H. Mac ~ 0.6 000000 0.4 0.4 E ~ 0.2 I i 0.0 0 4 2 6 10 0.0 -3 -2 -I 0 X 0.6 f 0.4 0.4 [ E 0.2 0.2 0.0 L-o!i6iIII."""""',-"--~_~~_ _--' -3 -2 -I 0 2 3 X ,= Figure 2: Large a approximation versus numerical simulations (with N = 10,000), for 0 and A = 0.2. Top row: Perceptron rule, with.,., = ~. Bottom row: Adatron rule, with.,., = ~. Left: training errors E t and generalisation errors Eg as functions of time, for aE {~, 1, 2}. Lines: approximated theory, markers: simulations (circles: E t , squares: Eg) . Right: joint distributions for student field and teacher noise p?[x] = dy P[x, y, z = ?y] (upper: P+[x], lower: P-[x]). Histograms: simulations, lines: approximated theory. J 7 Non-Linear Learning Rules: Theory versus Simulations In the case of non-linear learning rules no exact solution is known against which to test our formalism, leaving numerical simulations as the yardstick. We have evaluated numerically the large a approximation of our theory for Perceptron learning, 9[x, z] = sgn(z)O[-xz], and for Adatron learning, 9[x, z] = sgn(z)lzIO[-xz]. This approximation leads to the following fully explicit equation for the field distributions: 1/ d -p?[xly] = dt a . With U= ' +1 dx' p?[x'ly]{o[x-x'-.,.,.1'[x', ?y]] -o[x-x]} _ ~ {P[ I ] [W _ .,., 8 x y y J X ~ p?[xly] _.,.,2 Z!:I 2 2 uX ,X + U[X?(y)-RY]+(V-RW)[X-X?(y)]]} Q _ R2 Dydx {(I-A)P+[xly][x-P(y)]9[x,Y]+AP-[xly][x-x-(y)]9[x,-y]) V = W= Z= ! 1 1 Dydx x {(I-A)P+[xly]9[x, Y]+AP-[xly]9[x,-y]) Dydx y {(1-A)P+[xly]9[x, Y]+AP-[xly]9[x,-y]) Dydx {(I-A)P+[xly]92[x, Y]+AP-[xly]9 2[x,-yJ) Supervised Learning with Restricted Training Sets 243 J and with the short-hands X?(y) = dx xP?[xly). The result of our comparison is shown in figure 2. Note: E t increases monotonically with a, and Eg decreases monotonically with a, at any t. As in the noise-free formalism [7], the large a approximation appears to capture the dominant terms both for a -7 00 and for a -7 O. The predicting power of our theory is mainly limited by numerical constraints. For instance, the Adatron learning rule generates singularities at x = 0 in the distributions P?[xly) (especially for small "I) which, although predicted by our theory, are almost impossible to capture in numerical solutions. 8 Discussion We have shown how a recent theory to describe the dynamics of supervised learning with restricted training sets (designed to apply in the data recycling regime, and for arbitrary online and batch learning rules) [5, 6, 7] in large layered neural networks can be generalized successfully in order to deal also with noisy teachers. In our generalized approach the joint distribution P[x, y, z) for the fields of student, 'clean' teacher, and noisy teacher is taken to be a dynamical order parameter, in addition to the conventional observables Q and R. From the order parameter set {Q, R, P} we derive the generalization error Eg and the training error E t . Following the prescriptions of dynamical replica theory one finds a diffusion equation for P[x, y, z], which we have evaluated by making the replica-symmetric ansatz. We have carried out several orthogonal benchmark tests of our theory: (i) for a -7 00 (no data recycling) our theory is exact, (ii) for A -7 0 (no teacher noise) our theory reduces to that of [5, 6, 7], and (iii) for batch Hebbian learning our theory is exact. For on-line Hebbian learning our theory is exact with regard to the predictions for Q, R, Eg and the y-dependent conditional averages Jdx xP?[xly), at any time, and a crude approximation of our equations already gives reasonable agreement with the exact results [9] for E t . For non-linear learning rules (Perceptron and Adatron) we have compared numerical solution of a simple large a aproximation of our equations to numerical simulations, and found satisfactory agreement. This paper is a preliminary presentation of results obtained in the second stage of a research programme aimed at extending our theoretical tools in the arena of learning dynamics, building on [5, 6, 7]. Ongoing work is aimed at systematic application of our theory and its approximations to various types of non-linear learning rules, and at generalization of the theory to multi-layer networks. References [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] Mace C.W.H. and Coolen AC.C (1998), Statistics and Computing 8, 55 Saad D. (ed.) (1998), On-Line Learning in Neural Networks (Cambridge: CUP) Hertz J.A., Krogh A and Thorgersson G.I. (1989), J. Phys. A 22, 2133 HomerH. (1992a), Z. Phys. B 86, 291 and Homer H. (1992b), Z. Phys. B 87,371 Coolen A.C.C. and Saad D. (1998), in On-Line Learning in Neural Networks, Saad D. (ed.), (Cambridge: CUP) Coolen AC.C. and Saad D. (1999), in Advances in Neural Information Processing Systems 11, Kearns D., Solla S.A., Cohn D.A (eds.), (MIT press) Coolen A.C.C. and Saad D. (1999), preprints KCL-MTH-99-32 & KCL-MTH-99-33 Rae H.C., Sollich P. and Coolen AC.C. (1999), in Advances in Neural Information Processing Systems 11, Kearns D., Solla S.A., Cohn D.A. (eds.), (MIT press) Rae H.C., Sollich P. and Coolen AC.C. (1999),J. Phys. A 32, 3321 Inoue J.I. (1999) private communication Wong K.YM., Li S. and Tong YW. (1999),preprint cond-mat19909004 Biehl M., Riegler P. and Stechert M. (1995), Phys. Rev. E 52, 4624
1693 |@word private:1 closure:2 simulation:7 fonn:2 solid:1 moment:1 xiy:3 recovered:1 yet:6 dx:8 written:1 numerical:6 dydx:4 shape:1 designed:1 update:1 accordingly:1 short:2 iterates:1 complication:1 simpler:2 mathematical:1 along:1 introduce:1 indeed:3 xz:3 mechanic:1 ry:5 multi:1 cpu:1 increasing:1 underlying:2 qw:2 finding:2 homer:1 adatron:4 exactly:2 ro:4 uk:4 control:1 ly:1 limit:5 id:2 riegler:1 ap:4 studied:2 limited:1 range:1 unique:1 yj:1 procedure:1 significantly:1 get:1 cannot:1 close:1 layered:3 impossible:1 preprints:1 wong:1 conventional:1 map:1 deterministic:1 dz:3 yt:5 modifies:1 go:1 attention:2 annealed:1 l:2 independently:1 simplicity:1 rule:17 variation:1 qq:2 pt:7 exact:13 agreement:3 approximated:3 observed:1 bottom:1 preprint:1 solved:4 capture:2 ensures:1 solla:2 decrease:1 subshells:1 complexity:2 insertion:2 dynamic:10 oed:1 solving:1 upon:6 creates:1 observables:5 zjj:1 easily:1 joint:2 emulate:1 various:1 derivation:1 kcl:2 describe:2 london:4 larger:1 taylored:1 solve:2 biehl:1 erf:2 statistic:1 noisy:7 final:1 online:1 analytical:1 j2:6 relevant:1 turned:1 realization:1 moved:1 extending:1 oo:1 iq:1 ac:6 derive:1 krogh:1 predicted:1 involves:1 implies:1 xly:31 wedge:1 closely:1 peculiarity:1 sgn:7 tlz:1 wc2r:2 ao:1 generalization:4 preliminary:1 unrealizable:1 singularity:1 mathematically:1 pl:1 considered:1 exp:1 coolen:10 tf:1 successfully:2 tool:2 mit:2 gaussian:7 aim:1 rather:1 tcoolen:1 mainly:2 realizable:1 dependent:2 eliminate:1 mth:4 irn:1 jq:1 denoted:1 retaining:2 arccos:2 special:2 integration:1 field:4 construct:1 inevitable:1 randomly:2 composed:1 rj2:1 stechert:1 ourselves:1 interest:1 rae:2 possibility:1 erating:1 arena:1 behind:1 parametrised:1 integral:1 orthogonal:1 re:1 circle:1 deformation:1 theoretical:1 instance:1 formalism:6 tp:1 introducing:1 mac:1 answer:3 teacher:22 combined:1 density:1 systematic:1 yl:1 ansatz:5 ym:1 jo:1 again:2 x9:3 satisfied:1 containing:1 choose:1 li:1 student:5 explicitly:1 closed:2 analyze:1 doing:1 apparently:1 square:1 ir:1 generalize:3 published:1 phys:5 ed:4 definition:1 against:1 e2:1 proof:1 lim:2 back:1 appears:2 tum:1 dt:2 supervised:6 x6:1 follow:1 evaluated:3 done:1 just:1 stage:2 d:2 hand:2 working:1 qo:2 y9:4 cohn:2 marker:1 usage:1 building:1 evolution:1 analytically:1 symmetric:2 iteratively:1 satisfactory:1 eg:9 conditionally:3 deal:1 during:1 self:1 generalized:2 tn:1 recently:2 common:1 functional:3 physical:1 nrn:1 he:1 numerically:1 refer:1 cambridge:2 cup:2 consistency:1 mathematics:2 had:1 dj:2 aq:1 add:1 dominant:1 recent:2 irrelevant:1 apart:1 reverse:1 scenario:1 binary:1 tenns:1 dxdydz:4 monotonically:2 dashed:1 ii:5 rv:1 reduces:2 hebbian:8 technical:1 characterized:1 calculation:1 jiq:3 prescription:1 a1:1 prediction:4 involving:2 ae:1 noiseless:1 jid:2 histogram:1 background:1 hurdle:1 addition:1 leaving:2 macroscopic:9 launched:1 saad:5 ee:1 presence:1 iii:2 restrict:1 reduce:1 idea:1 intensive:1 aproximation:1 absent:2 expression:3 jj:1 adequate:1 useful:1 generally:1 yw:1 aimed:3 involve:1 amount:1 concentrated:1 kc1:2 simplest:3 reduced:1 generate:1 rw:2 zj:2 write:1 key:1 putting:1 drawn:3 clean:1 diffusion:1 replica:6 fraction:1 extends:1 almost:1 jdx:4 reasonable:2 separation:1 draw:1 dy:6 jdj:3 layer:1 followed:1 occur:2 ahead:1 infinity:1 precisely:1 worked:1 bp:1 constraint:1 generates:2 argument:1 developing:1 hertz:1 remain:1 sollich:2 em:1 rev:1 making:3 memorizing:1 restricted:9 taken:2 equation:21 agree:1 describing:2 turn:1 xyz:1 detennine:1 prerequisite:1 apply:2 obey:2 away:1 batch:7 deserved:1 top:1 remaining:1 recycling:7 calculating:1 giving:2 build:1 especially:1 question:3 already:1 fonnalism:2 trivial:1 assuming:2 edl:2 providing:1 acquire:1 equivalently:1 unfortunately:1 dtr:1 upper:2 benchmark:4 implementable:1 situation:1 communication:1 arbitrary:3 pair:1 learned:1 dynamical:4 wy:1 regime:4 reading:1 reliable:1 green:2 including:1 power:2 natural:1 predicting:1 solvable:2 improve:1 djn:1 inoue:1 xg:1 carried:1 review:1 bjl:2 law:8 fully:2 proven:1 versus:3 ingredient:1 ar2:1 consistent:1 xp:4 row:2 zly:8 last:1 free:5 perceptron:4 regard:1 curve:2 calculated:1 made:1 programme:1 far:2 compact:1 observable:1 cavity:1 dealing:1 reproduces:1 overfitting:3 decade:1 rearranging:1 expanding:1 mace:4 expansion:1 excellent:1 did:1 rh:1 noise:14 arise:1 tong:1 explicit:1 xl:2 crude:1 ply:1 specific:1 r2:7 decay:1 mirror:1 execution:2 entropy:1 generalizing:2 saddle:6 strand:2 ux:1 g2:1 scalar:1 subtlety:1 applies:1 extracted:1 conditional:2 identity:1 presentation:2 king:2 typical:1 infinite:1 operates:1 generalisation:1 averaging:1 kearns:2 cond:1 perceptrons:1 college:2 latter:1 yardstick:1 relevance:1 ongoing:1 dept:2 phenomenon:1
758
1,694
Efficient Approaches to Gaussian Process Classification Lehel Csato, Ernest Fokoue, Manfred Opper, Bernhard Schottky Neural Computing Research Group School of Engineering and Applied Sciences Aston University Birmingham B4 7ET, UK. {opperm,csatol}~aston.ac.uk Ole Winther Theoretical Physics II, Lund University, Solvegatan 14 A, S-223 62 Lund, Sweden winther~thep.lu.se Abstract We present three simple approximations for the calculation of the posterior mean in Gaussian Process classification. The first two methods are related to mean field ideas known in Statistical Physics. The third approach is based on Bayesian online approach which was motivated by recent results in the Statistical Mechanics of Neural Networks. We present simulation results showing: 1. that the mean field Bayesian evidence may be used for hyperparameter tuning and 2. that the online approach may achieve a low training error fast. 1 Introduction Gaussian processes provide promising non-parametric Bayesian approaches to regression and classification [2, 1]. In these statistical models, it is assumed that the likelihood of an output or target variable y for a given input x E RN can be written as P(Yla(x)) where a : RN --+ R are functions which have a Gaussian prior distribution, i.e. a is (a priori) assumed to be a Gaussian random field. This means that any finite set of field variables a(xi), i = 1, ... ,l are jointly Gaussian distributed with a given covariance E[a(xi)a(xj)] = K(Xi' Xj) (we will also assume a zero mean throughout the paper). Predictions on a(x) for novel inputs x, when a set D of m training examples (Xi, Yi) i = 1, . . . , m , is given, can be computed from the posterior distribution of the m + 1 variables a(x) and a(xd, ... ,a(xm). A major technical problem of the Gaussian process models is the difficulty of computing posterior averages as high dimensional integrals, when the likelihood is not Gaussian. This happens for example in classification problems. So far, a variety of approximation techniques have been discussed: Monte Carlo sampling [2], the MAP approach [4], bounds on the likelihood [3] and a TAP mean field approach [5]. In this paper, we will introduce three different novel methods for approximating the posterior mean of the random field a(x), which we think are simple enough to be used in practical applications. Two of the techniques L. Csato, E. Fokoue, M Opper, B. Schottky and 0. Winther 252 are based on mean field ideas from Statistical Mechanics, which in contrast to the previously developed TAP approach are easier to implement. They also yield simple approximations to the total likelihood of the data (the evidence) which can be used to tune the hyperparameters in the covariance kernel K (The Bayesian evidence (or MLII) framework aims at maximizing the likelihood of the data). We specialize to the case of a binary classification problem, where for simplicity, the class label Y = ? 1 is assumed to be noise free and the likelihood is chosen as P(Yla) = 8(ya) , (1) where 8(x) is the unit step function, which equals 1 for x > 0 and zero else. We are interested in computing efficient approximations to the posterior mean (a(x)), which we will use for a prediction of the labels via Y = sign(a(x)), where (.. .) denotes the posterior expectation. If the posterior distribution of a(x) is symmetric around its mean, this will give the Bayes optimal prediction. Before starting, let us add two comments on the likelihood (1). First, the MAP approach (i.e. predicting with the fields a that maximize the posterior) would not be applicable, because it gives the trivial result a(x) = O. Second, noise can be easily introduced within a probit model [2], all subsequent calculations will only be slightly altered. Moreover, the Gaussian average involved in the definition of the probit likelihood can always be shifted from the likelihood into the Gaussian process prior, by a redefinition of the fields a (which does not change the prediction), leaving us with the simple likelihood (1) and a modified process covariance [5] . 2 Exact Results At first glance, it may seem that in order to calculate (a(x)) we have to deal with the joint posterior of the fields ai = a(xi) ' i = 1, ... , m together with the field at the test point a(x) . This would imply that for any test point, a different new m + 1 dimensional average has to be performed. Actually, we will show that this is not the case. As above let E denote the expectation over the Gaussian prior. The posterior expectation at any point, say x (a(x)) E [a (x) TI7=l P(Yj laj)] E [TI7=l P(Yj laj)] = --=-~---~--"- (2) can by integration by parts-for any likelihood-be written as (a(x )) ~ K(x, Xj )ajYj = L...J and j )) ",.=y . (CHnp(Yj1a .... J J aa . j (3) J showing that aj is not dependent on the test point x. It is therefore not necessary to compute a m + 1 dimensional average for every prediction. We have chosen the specific definition (3) in order to stress the similarity to predictions with Support Vector Machines (for the likelihood (1), the aj will come out nonnegative) . In the next sections we will develop three approaches for an approximate computation of the aj. 3 Mean Field Method I: Ensemble Learning Our first goal is to approximate the true posterior distribution 1 1 1 T -1 m p( alD ) - e-"2 a K a P(Y ?la?) m - Z J(27r)m detK j=l J J II (4) 253 Efficient Approaches to Gaussian Process Classification of a == (al,"" am) by a simpler, tractable distribution q. Here, K denotes the covariance matrix with elements Kij = K (Xi, Xj). In the variational mean field approach-known as ensemble learning in the Neural Computation Community,the relative entropy distance K L(q,p) = da q(a) In :~:~ is minimized in the family of product distributions q(a) = TI,7=l qj(aj). This is in contrast to [3], where a variational bound on the likelihood is computed. We get J KL(q,p) = ! qi(ai) daiqi(ai) In P(Yilai) L [K-1Lj (ai)O(aj)o ~ , ~ i,j,i#j + +~L [K-1Li (a~)o i where (.. ')0 denotes expectation w.r.t. q. By setting the functional derivative of KL(q,p) with respect to qi(a) equal to zero, we find that the best product distribution is a Gaussian prior times the original Likelihood: 1 (o-"'i)2 qi(a) ex: P(Yila) ~e- hi (5) where mi = -Ai :Ej,#i(K-l)ij(aj)o and Ai = [K-l]:l. Using this specific form for the approximated posterior q(a), replacing the average over the true posterior in (3) by the approximation (5), we get (using the likelihood (I)) a set of m nonlinear equations in the unknowns aj: where D(z) = e- z2 / 2 /..f2i and <I>(z) = J~oo dt D(t). As a useful byproduct of the variational approximation, an upper bound on the Bayesian evidence P(D) = J da 7r(a)P{Dla) can be derived. (71' denotes the Gaussian process prior and P{Dla) = TI,7=l P(Yjlaj)). The bound can be written in terms of the mean field 'free energy' as -lnP(D) < Eqlnq(a) -Eqln[7r(a)P(Dla)] - '''In<l> ~ , (Y', JXi m j ) + ~2 "'y ?a ?(K?'3? -8" A?)y3?a?3 ~ " 'J' (7) ~ which can be used as a yardstick for selecting appropriate hyperparameters in the covariance kernel. The ensemble learning approach has the little drawback, that it requires inversion of the covariance matrix K and, for the free energy (7) one must compute a determinant. A second, simpler approximation avoids these computations. 4 Mean Field Theory II: A 'Naive' Approach The second mean field theory aims at working directly with the variables aj. As a starting point, we consider the partition function (evidence), Z = P(D) = ! dze-tzTKz IT P(Yjlzj) , ;=1 (8) 254 L. Csato. E. Fokoue. M. Opper. B. Schottky and O. Winther which follows from (4) by a standard Gaussian integration, introducing the Fourier transform of the Likelihood .P(Ylz) = I g~ eiaz P(Yla) with i being the imaginary unit. It is tempting to view (8) as a normalizing partition function for a Gaussian process Zi having covariance matrix K-l and likelihood P. Unfortunately, P is not a real number and precludes a proper probabilistic interpretation. N evertheless, dealing formally with the complex measure defined by (8), integration by parts shows that one has YjCl!j = -i(zj)., where the brackets (... ). denote a average over the complex measure. This suggests a simple approximation for calculating the Cl!j. One may think of trying a saddle-point (or steepest descent) approximation to (8) and replace (Zj). by the value of Zj (in the complex Z plane) which makes the integrand stationary thereby neglecting the fluctuations of the Zj. Hence, this approximation would treat expectations of products as (ZiZj). as (Zi).(Zj)*, which may be reasonable for i i= j, but definitely not for the self-correlation i = j. According to the general formalism of mean field theories (outlined e.g. in [6]), one can improve on that idea, by treating the 'self-interactions' separately. This can be done by replacing all Zi (except in the form zi) by a new variable J.Li by inserting a Dirac 8 function representation 8(z - J.L) = J ~r;:e-im(z-J1.) into (8) and integrate over the Z and a variables exactly (the integral factorizes), and finally perform a saddle-point integration over the m and J.L variables. The details of this calculation will be given elsewhere. Within the saddle-point approximation, we get the system of nonlinear equations z; i,i=l=j i,i=l=j which is of the same form as (6) with Aj replaced by the simpler K jj . These equations have also been derived by us in [5] using a Callen identity, but our present derivation allows also for an approximation to the evidence. By plugging the saddlepoint values back into the partition function, we get -lnP(D) ~ - ~ln~ (Y'~) ~Y'Cl!'(K"tJ ~ VKii + ~2 ~ t t t t 8?tJ?K 11.. )y?Cl!? J J ~ which is also simpler to compute than (7) but does not give a bound on the true evidence. 5 A sequential Approach Both previous algorithms do not give an explicit expression for the posterior mean, but require the solution of a set of nonlinear equations. These must be obtained by an iterative procedure. We now present a different approach for an approximate computation of the posterior mean, which is based on a single sequential sweep through the whole dataset giving an explicit update of the posterior. The algorithm is based on a recently proposed Bayesian approach to online learning (see [8] and the articles of Opper and Winther& Solla in [9]). Its basic idea applied to the Gaussian process scenario, is as follows: Suppose, that qt is a Gaussian approximation to the posterior after having seen t examples. This means that we approximate the posterior process by a Gaussian process with mean (a(x))t and covariance Kt(x, y), starting with (a(x))o = 0 and Ko(x, y) = K(x, y). After a new data point Yt+l is observed, the posterior is updated according to Bayes rule. The new non-Gaussian posterior qt is projected back into the family of Gaussians by choosing the closest Gaussian qt+l minimizing the relative entropy K L(qt, qt+d Efficient Approaches to Gaussian Process Classification 255 in order to keep the loss of information small. This projection is equivalent to a matching of the first two moments of lit and qt+1 ' E.g., for the first moment we get (a(x))t+l = (a(x) P(Yt+1la(Xt+d))t (P( I ( ))) Yt+l a Xt+l t = (a(x))t + Kl (t)Kt(x, Xt+l) where the second line follows again from an integration by parts and Kt{t) 1!:!?!. and ~2(t) -- K t (x t+l, x t+l ) . This recursion and u ~ tJ>(zt} with z t -- Yt?l (a(Xt?l)t u(t) the corresponding one for K t can be solved by the ansatz t (a(x))t L K(x, xj)Yjaj(t) (10) j=l L K(x, Xi)Cij (t)K(x, Xj) + K(x, y) (11) i,j where the vector a(t) = (al,"" at, 0, 0, ... ) and the matrix C(t) (which has also only txt nonzero elements) are updated as a(t + 1) a(t) + Kl(t) (C(t)kt+l + et+l) C(t + 1) C(t) where K2 (t) = ~ { ~~~:?) @y + K2(t) (C(t)kt+l + et+1) (C(t)kt+1 + et+1f - (~gt\) r}, (12) k t is the vector with elements K tj , j = 1 ... , t and @ denotes the element-wise product between vectors. The sequen- tial algorithm defined by (10)-(12) has the advantage of not requiring any matrix inversions. There is also no need to solve a numerical optimization problem at each time as in the approach of [11] where a different update of a Gaussian posterior approximation was proposed. Since we do not require a linearization of the likelihood, the method is not equivalent to the extended Kalman Filter approach. Since it is possible to compute the evidence of the new datapoint P(Yt+1) = (P(Yt+1lat+d)t based on the old posterior, we can compute a further approximation to the log evidence for m data via In P(Dm) = 2:~lln(P(Yt+llat+1)k 6 Simulations We present two sets of simulations for the mean field approaches. In the first, we test the Bayesian evidence framework for tuning the hyperparameters of the covariance function (kernel). In the second, we test the ability of the sequential approach to achieve low training error and a stable test error for fixed hyperparameters. For the evidence framework, we give simulation results for both mean field free energies (7) and (10) on a single data set, 'Pima Indian Diabetes (with 200/332 training/test-examples and input dimensionality d = 7) [7]. The results should therefore not be taken as a conclusive evidence for the merits of these approaches, but simply as an indication that they may give reasonable results. We use the radial basis function covariance function K(x,x') = exp (-~ 2:~WI(XI - XD2) . A diagonal term v is added to the covariance matrix corresponding to a Gaussian noise added to the fields with variance v [5]. The free energy, -lnP(D) is minimized by gradient descent with respect to v and the lengthscale parameters WI, ? .. , Wd and the mean field equations for aj are solved by iteration before each update of the hyperparameters (further details will be given elsewhere). Figure 1 shows the evolution of the naive mean free energy and the test error starting from uniform L. Csat6, E. Fokoue, M. Opper, B. &hottky and 0. Winther 256 ws. It typically requires of the order of 10 iteration steps of the a;-equations between each hyperparameter update. We also used hybrid approaches, where the free energy was minimized by one mean field algorithm and the hyperparameters used in the other. As it may be seen from table 1, the naive mean field theory can overestimate the free energy (since the ensemble free energy is an upper bound to the free energy). The overestimation is not nearly as severe at the minimum of the naive mean field free energy. Another interesting observation is that as long as the same hyperparameters are used the actual performance (as measured by the test error) is not very sensitive to the algorithm used. This also seems to be the case for the TAP mean field approach and Support Vector Machines [5]. 74..-----~----~-------,-----__, 115..-----~----~-------,-----__, 0114 \\ Ii:" ' 70 g .E 1. 113 >. ~ ,, W68 ~ Q) c: ~112 1-66 \ , ~ u... 64 '. "'----- 111 o 20 ~ Iterations 60 62 ~ 0~----2~0----~~~--~~~--~~ Iterations Figure 1: Hyperparameter optimization for the Pima Indians data set using the naive mean field free energy. Left figure: The free energy as a function of the number of hyperparameter updates. Right figure: The test error count (out of 332) as a function of the number of hyperparameter updates. Table 1: Pima Indians dataset. Hyperparameters found by free energy minimization. Left column gives the free energy -lnP(D) used in hyperparameter optimization. Test error counts in range 63- 75 have previously been reported [5] Free Energy minimization Ensemble Mean Field, eq. (7) Naive Mean Field, eq. (10) Ensemble MF Error -lnP(D) 72 62 100.6 107.0 Naive MF Error -lnP(D) 70 62 183.2 110.9 For the sequential algorithm, we have studied the sonar [10] and crab [7] datasets. Since we have not computed an approximation to the evidence so far, a simple fixed polynomial kernel was used. Although a probabilistic justification of the algorithm is only valid, when a single sweep through the data is used (the independence of the data is assumed), it is tempting to reuse the same data and iterate the procedure as a heuristic. The two plots show that in this way, only a small improvement is obtained, and it seems that the method is rather efficient in extracting the information from the data in a single presentation. For the sonar dataset, a single sweep is enough to achieve zero training error. Acknowledgements: BS would like to thank the Leverhulme Trust for their support (F /250/K). The work was also supported by EPSRC Grant GR/L52093. 257 Efficient Approaches to Gaussian Process Classification so - Training Error - - Test Error 45 40 35 35 30 ... 30 25 g25 I 20 1,- w 20 ~ 15 15 10 10 \ -" 00 20 40 60 80 1'\ .." \ 140 ___ 160 ,- - .. - - - - - .. 5 ?0L----5~0---F=::::lIQ.-A.1SO'::-----::-2~OO Heration. Figure 2: Training and test errors during learning for the sonar (left) and crab dataset (right). The vertical dash-dotted line marks the end of the training set and the starting point of reusing of it. The kernel function used is K(x, x') (1 + x . x' jm)k with order k = 2 (m is the dimension of inputs) . References [1] Williams C.K.I. and Rasmussen C.E., Gaussian Processes for Regression, in Neural Information Processing Systems 8, Touretzky D.s, Mozer M.C. and Hasselmo M.E. (eds.), 514-520, MIT Press (1996) . [2] Neal R .M, Monte Carlo Implementation of Gaussian Process Models for Bayesian Regression and Classification, Technical Report 9702, Department of Statistics, University of Toronto (1997). [3] Gibbs M.N. and Mackay D.J.C., Variational Gaussian Process Classifiers, Preprint Cambridge University (1997). [4] Williams C.K.I. and Barber D, Bayesian Classification with Gaussian Processes, IEEE Trans Pattern Analysis and Machine Intelligence, 20 1342-1351 (1998). [5] Opper M. and Winther O. Gaussian Processes for Classification: Mean Field Algorithms, Submitted to Neural Computation, http://www.thep.lu.se /tf2/staff/winther/ (1999). [6] Zinn-Justin J, Quantum Field Theory and Critical Phenomena, Clarendon Press Oxford (1990) . [7] Ripley B.D, Pattern Recognition and Neural Networks, Cambridge University Press (1996). [8] Opper M., Online versus Offline Learning from Random Examples: General Results, Phys. Rev. Lett. 77, 4671 (1996). [9] Online Learning in Neural Networks, Cambridge University Press, D. Saad (ed.) (1998). [10] Gorman R.P and Sejnowski T .J, Analysis of hidden units in a layered network trained to classify sonar targets, Neural Networks 1, (1988). [11] Jaakkola T . and Haussler D. Probabilistic kernel regression, In Online Proceedings of 7-th Int. Workshop on AI and Statistics (1999) , http://uncertainty99.microsoft.com/proceedings.htm.
1694 |@word determinant:1 inversion:2 polynomial:1 seems:2 simulation:4 covariance:11 thereby:1 moment:2 selecting:1 imaginary:1 z2:1 wd:1 com:1 written:3 must:2 numerical:1 subsequent:1 partition:3 j1:1 treating:1 plot:1 ti7:2 update:6 stationary:1 intelligence:1 plane:1 steepest:1 manfred:1 toronto:1 simpler:4 specialize:1 introduce:1 mechanic:2 little:1 actual:1 jm:1 moreover:1 developed:1 every:1 y3:1 ti:2 xd:1 exactly:1 k2:2 classifier:1 uk:2 unit:3 grant:1 overestimate:1 before:2 engineering:1 treat:1 oxford:1 fluctuation:1 studied:1 suggests:1 range:1 practical:1 yj:2 implement:1 procedure:2 projection:1 dla:3 matching:1 radial:1 get:5 layered:1 www:1 equivalent:2 map:2 yt:7 maximizing:1 csat6:1 williams:2 starting:5 simplicity:1 rule:1 haussler:1 justification:1 updated:2 target:2 suppose:1 exact:1 diabetes:1 element:4 approximated:1 recognition:1 observed:1 epsrc:1 preprint:1 solved:2 tf2:1 calculate:1 solla:1 mozer:1 overestimation:1 trained:1 basis:1 easily:1 joint:1 htm:1 derivation:1 fast:1 lengthscale:1 monte:2 ole:1 sejnowski:1 choosing:1 heuristic:1 solve:1 say:1 precludes:1 ability:1 statistic:2 think:2 jointly:1 transform:1 online:6 advantage:1 indication:1 interaction:1 product:4 inserting:1 achieve:3 opperm:1 dirac:1 oo:2 develop:1 ac:1 measured:1 ij:1 qt:6 school:1 eq:2 come:1 drawback:1 filter:1 require:2 im:1 around:1 crab:2 exp:1 major:1 birmingham:1 applicable:1 label:2 sensitive:1 hasselmo:1 minimization:2 mit:1 jxi:1 gaussian:30 always:1 aim:2 modified:1 rather:1 ej:1 factorizes:1 jaakkola:1 derived:2 improvement:1 likelihood:18 contrast:2 am:1 dependent:1 lehel:1 lj:1 typically:1 w:1 hidden:1 interested:1 classification:11 priori:1 integration:5 mackay:1 field:30 equal:2 having:2 sampling:1 lit:1 nearly:1 minimized:3 report:1 replaced:1 microsoft:1 severe:1 detk:1 bracket:1 tj:4 kt:6 integral:2 byproduct:1 necessary:1 neglecting:1 sweden:1 old:1 theoretical:1 kij:1 formalism:1 column:1 classify:1 tial:1 introducing:1 uniform:1 gr:1 reported:1 winther:8 definitely:1 probabilistic:3 physic:2 ansatz:1 together:1 again:1 derivative:1 li:2 reusing:1 int:1 performed:1 view:1 bayes:2 solvegatan:1 variance:1 ensemble:6 yield:1 bayesian:9 lu:2 carlo:2 submitted:1 datapoint:1 phys:1 touretzky:1 ed:2 definition:2 energy:15 involved:1 dm:1 mi:1 dataset:4 dimensionality:1 actually:1 back:2 clarendon:1 dt:1 done:1 correlation:1 working:1 replacing:2 trust:1 nonlinear:3 glance:1 aj:10 requiring:1 true:3 evolution:1 hence:1 symmetric:1 nonzero:1 neal:1 deal:1 during:1 self:2 trying:1 stress:1 variational:4 wise:1 novel:2 recently:1 functional:1 b4:1 discussed:1 interpretation:1 cambridge:3 gibbs:1 ai:7 tuning:2 outlined:1 stable:1 similarity:1 zizj:1 gt:1 add:1 posterior:22 closest:1 recent:1 scenario:1 binary:1 yi:1 lnp:6 seen:2 minimum:1 staff:1 maximize:1 tempting:2 ii:4 technical:2 calculation:3 long:1 plugging:1 qi:3 prediction:6 ernest:1 ko:1 regression:4 basic:1 redefinition:1 expectation:5 txt:1 iteration:4 kernel:6 csato:3 laj:2 separately:1 else:1 leaving:1 saad:1 comment:1 seem:1 extracting:1 enough:2 variety:1 xj:6 independence:1 zi:4 iterate:1 idea:4 qj:1 motivated:1 expression:1 reuse:1 jj:1 useful:1 se:2 tune:1 liq:1 http:2 zj:5 shifted:1 dotted:1 sign:1 leverhulme:1 hyperparameter:6 group:1 schottky:3 ylz:1 zinn:1 throughout:1 family:2 reasonable:2 bound:6 hi:1 dash:1 nonnegative:1 fourier:1 integrand:1 department:1 according:2 slightly:1 wi:2 saddlepoint:1 rev:1 b:1 happens:1 taken:1 ln:1 equation:6 previously:2 count:2 merit:1 tractable:1 end:1 gaussians:1 appropriate:1 original:1 denotes:5 lat:1 calculating:1 giving:1 approximating:1 sweep:3 added:2 parametric:1 diagonal:1 gradient:1 distance:1 thank:1 f2i:1 barber:1 trivial:1 kalman:1 minimizing:1 unfortunately:1 cij:1 pima:3 implementation:1 proper:1 zt:1 unknown:1 perform:1 upper:2 vertical:1 observation:1 datasets:1 finite:1 descent:2 extended:1 rn:2 yla:3 community:1 introduced:1 kl:4 conclusive:1 tap:3 trans:1 justin:1 pattern:2 xm:1 lund:2 critical:1 difficulty:1 hybrid:1 predicting:1 recursion:1 altered:1 improve:1 aston:2 imply:1 naive:7 prior:5 acknowledgement:1 relative:2 probit:2 loss:1 interesting:1 versus:1 integrate:1 article:1 elsewhere:2 supported:1 lln:1 free:16 rasmussen:1 offline:1 distributed:1 opper:7 dimension:1 valid:1 avoids:1 lett:1 quantum:1 projected:1 mlii:1 far:2 approximate:4 bernhard:1 keep:1 dealing:1 assumed:4 xi:8 thep:2 ripley:1 iterative:1 sonar:4 table:2 promising:1 complex:3 cl:3 da:2 whole:1 noise:3 hyperparameters:8 fokoue:4 explicit:2 third:1 specific:2 xt:4 showing:2 evidence:13 normalizing:1 workshop:1 sequential:4 linearization:1 gorman:1 easier:1 mf:2 entropy:2 simply:1 saddle:3 aa:1 ald:1 goal:1 identity:1 presentation:1 replace:1 xd2:1 change:1 except:1 total:1 ya:1 la:2 formally:1 support:3 mark:1 yardstick:1 indian:3 phenomenon:1 ex:1
759
1,695
A generative model for attractor dynamics Richard S. Zemel Department of Psychology University of Arizona Tucson, AZ 85721 Michael C. Mozer Department of Computer Science University of Colorado Boulder, CO 80309-0430 zemel@u.arizona.edu mozer@colorado.edu Abstract Attractor networks, which map an input space to a discrete output space, are useful for pattern completion. However, designing a net to have a given set of attractors is notoriously tricky; training procedures are CPU intensive and often produce spurious afuactors and ill-conditioned attractor basins. These difficulties occur because each connection in the network participates in the encoding of multiple attractors. We describe an alternative formulation of attractor networks in which the encoding of knowledge is local, not distributed. Although localist attractor networks have similar dynamics to their distributed counterparts, they are much easier to work with and interpret. We propose a statistical formulation of localist attract or net dynamics, which yields a convergence proof and a mathematical interpretation of model parameters. Attractor networks map an input space, usually continuous, to a sparse output space composed of a discrete set of alternatives. Attractor networks have a long history in neural network research. Attractor networks are often used for pattern completion, which involves filling in missing, noisy, or incorrect features in an input pattern. The initial state of the attractor net is typically determined by the input pattern. Over time, the state is drawn to one of a predefined set of states-the attractors. Attractor net dynamics can be described by a state trajectory (Figure 1a). An attractor net is generally implemented by a set of visible units whose activity represents the instantaneous state, and optionally, a set of hidden units that assist in the computation. Attractor dynamics arise from interactions among the units. In most formulations of afuactor nets,2,3 the dynamics can be characterized by gradient descent in an energy landscape, allowing one to partition the output space into attractor basins. Instead of homogeneous attractor basins, it is often desirable to sculpt basins that depend on the recent history of the network and the arrangement of attractors in the space. In psychological models of human cognition, for example, priming is fundamental: after the model visits an attractor, it should be faster to fall into the same attractor in the near future, i.e., the attractor basin should be broadened. 1 ,6 Another property of attractor nets is key to explaining behavioral data in psychological and neurobiological models: the gang effect, in which the strength of an attractor is influenced by other attractors in its neighborhood. Figure 1b illustrates the gang effect: the proximity of the two rightmost afuactors creates a deeper attractor basin, so that if the input starts at the origin it will get pulled to the right. 81 A Generative Model for Attractor Dynamics k, " , ,/ " "--- ---- .. - Figure 1: (a) A two-dimensional space can be carved into three regions (dashed lines) by an attractor net. The dynamics of the net cause an input pattern (the X) to be mapped to one of the attractors (the O's). The solid line shows the temporal trajectory of the network state. (b) the actual energy landscape for a localist attractor net as a function of y, when the input is fixed at the origin and there are ((-1,0), (1,0), (1, -A)), with a uniform prior. The shapes of three attractors, W attractor basins are influenced by the proximity of attractors to one another (the gang effect). The origin of the space (depicted by a point) is equidistant from the attractor on the left and the attractor on the upper right, yet the origiri clearly lies in the basin of the right attractors. = 1bis effect is an emergent property of the distribution of attractors, and is the basis for interesting dynamics; it produces the mutually reinforcing or inhibitory influence of similar items in domains such as semantics,9 memory,lO,12 and olfaction.4 Training an attract or net is notoriously tricky. Training procedures are CPU intensive and often produce spurious attractors and ill-conditioned attractor basins. 5,11 Indeed, we are aware of no existing procedure that can robustly translate an arbitrary specification of an attractor landscape into a set of weights. These difficulties are due to the fact that each connection participates in the specification of multiple attractors; thus, knowledge in the net is distributed over connections. We describe an alternative attractor network model in which knowledge is localized, hence the name localist attractor network. The model has many virtues, including: a trivial procedure for wiring up the architecture given an attractor landscape; eliminating spurious attractors; achieving gang effects; providing a clear mathematical interpretation of the model parameters, which clarifies how the parameters control the qualitative behavior of the model (e.g., the magnitude of gang effects); and proofs of convergence and stability. A Iocalist attractor net consists of a set of n state units and m attractor units. Parameters associated with an attractor unit i encode the location of the attractor, denoted Wi, and its "pull" or strength, denoted 7ri, which influence the shape of the attractor basin. Its activity at time t, qi(t), reflects the normalized distance from the attractor center to the current state, y(t), weighted by the attractor strength: g(y, w, 0") 7rjg(y(t) , Wi, O"(t)) L:j 7rjg(y(t) , Wj, O"(t)) (1) exp( -\y - w\2/20"2) (2) Thus, the attractors form a layer of normalized radial-basis-function units. The input to the net, &, serves as the initial value of the state, and thereafter the state is pulled toward attractors in proportion to their activity. A straightforward 82 R. S. Zemel and M. C. Mozer expression of this behavior is: (3) where a(l) = Ion the first update and a(t) = 0 fort> 1. More generally, however, one might want to gradually reduce a over time, allowing for a persistent effect of the external input on the asymptotic state. The variables O"(t) and a(t) are not free parameters of the model, but can be derived from the formalism we present below. The localist attractor net is motivated by a generative model of the input based on the attractor distribution, and the network dynamics corresponds to a search for a maximum likelihood interpretation of the observation. In the following section, we derive this result, and then present simulation studies of the architecture. 1 A MAXIMUM LIKELIHOOD FORMULATION The starting point for the statistical formulation of a localist attractor network is a mixture of Gaussians model. A standard mixture of Gaussians consists of m Gaussian density functions in n dimensions. Each Gaussian is parameterized by a mean, a covariance matrix, and a mixture coefficient. The mixture model is generative, i.e., it is considered to have produced a set of observations. Each observation is generated by selecting a Gaussian based on the mixture coefficients and then stochastically selecting a point from the corresponding density function. The model parameters are adjusted to maximize the likelihood of a set of observations. The Expectation-Maximization (EM) algorithm provides an efficient procedure for estimating the parameters.The Expectation step calculates the posterior probability qi of each Gaussian for each observation, and the Maximization step calculates the new parameters based on the previous values and the set of qi . The mixture of Gaussians model can provide an interpretation for a localist attractor network, in an unorthodox way. Each Gaussian corresponds to an attractor, and an observation corresponds to the state. Now, however, instead of fixing the observation and adjusting the Gaussians, we fix the Gaussians and adjust the observation. If there is a single observation, and a = 0 and all Gaussians have uniform spread 0", then Equation 1 corresponds to the Expectation step, and Equation 3 to the Maximization step in this unusual mixture model. Unfortunately, this simple characterization of the localist attractor network does not produce the desired behavior. Many situations produce partial solutions, in which the observation does not end up at an attractor. For example, if two unidimensional Gaussians overlap significantly, the most likely value for the observation is midway between them rather than at the mean of either Gaussian. We therefore extend this mixture-of-Gaussians formulation to better characterize the localist attractor network. As in the simple model, each of the m attractors is a Gaussian generator, the mean of which is a location in the n-dimensional state space. The input to the net, e, is considered to have been generated by a stochastic selection of one of the attractors, followed by the addition of zero-mean Gaussian noise with variance specified by the attractor. Given a particular observation e, the an attractor's posterior probability is the normalized Gaussian probability of e, weighted by its mixing proportion. This posterior distribution for the attractors corresponds to a distribution in state space that is a weighted sum of Gaussians. We then consider the attractor network as encoding this distribution over states implied by the attractor posterior probabilities. At anyone time, however, the attractor network can only represent a single position in state space, rather than 83 A Generative Model for Attractor Dynamics the entire distribution over states. This restriction is appropriate when the state is an n-dimensional point represented by the pattern of activity over n state units. To accommodate this restriction, we change the standard mixture of Gaussians generative model by interjecting an intermediate level between the attractors and the observation. The first generative level consists of the discrete attractors, the second is the state space, and the third is the observation. Each observation is considered to have been generated by moving down this hierarchy: 1. select an attractor x = i from the set of attractors 2. select a state (i.e., a pattern of activity across the state units) based on the preferred location of that attractor: y = Wi + Ny 3. select an observation z = yG + N z The observation z produced by a particular state y depends on the generative weight matrix G. In the networks we consider here, the observation and state spaces are identical, so G is the identity matrix, but the formulation allows for z to lie in some other space. Ny and N z describe the zero-mean, spherical Gaussian noise introduced at the two levels, with deviations (1 y and (1z, respectively. In comparison with the 2-level Gaussian mixture model described above, this 3level model is more complicated but more standard: the observation & is preserved as stable data, and rather than the model manipulating the data here it can be viewed as iteratively manipulating an internal representation that fits the observation and attractor structure. The attractor dynamics correspond to an iterative search through state space to find the most likely single state that: (a) was generated by the mixture of Gaussian attractors, and (b) in tum generated the observation. Under this model, one could fit an observation & by finding the posterior distribution over the hidden states (X and Y) given the observation: (X i Y = ylZ = &) = p(&ly, i)p(y, i) p(&IY) 1riP(yli) (4) p, p(&) Jy p(&IY)L:i1riP(Yli)dy = = where the conditional distributions are Gaussian: p(Y = ylX = i) = 9(ylwi, (1y) and p(&IY = y) = 9(&ly, (1z). Evaluating the distribution in Equation 4 is tractable, because the partition function is a sum of a set of Gaussian integrals. Due to the restriction that the network cannot represent the entire distribution, we do not directly evaluate this distribution but instead adopt a mean-field approach, in which we approximate the posterior by another distribution Q(X, YI&). Based on this approximation, the network dynamics can be seen as minimizing an objective function that describes an upper bound on the negative log probability of the observation given the model and mean-field parameters. In this approach, one can choose any form of Q to estimate the posterior distribution, but a better estimate allows the network to approach a maximum likelihood solution. 13 We select a simple posterior: Q(X, Y) = qid(Y = y), where qi = Q(X = i) is the responsibility assigned to attractor i, and y is the estimate of the state that accounts for the observation. The delta function over Y is motivated by the restriction that the explanation of an input consists of a single state. Given this posterior distribution, the objective for the network is to minimize the free energy F, described here for a particular input example &: F(q,yl&) = J (. , ~ ~ Q(X = i, Y = y) QX=t,Y=y)lnp(&,X=i,Y=y)dy L,. qi In 1r,qi. -lnp(&IY) - L,. qi lnp(yli) R. S. Zemel and M C. Mozer 84 where 7r; is the prior probability (mixture coefficient) associated with attractor i. These priors are parameters of the generative model, as are (fy, u z , and w. F(q, yl&) = 2: qdn 7r,qi. + 2 \ 1& - Yl2 + 2 \ Uz i (f Y 2: qi!y - wd 2 + n In((fy(fz) (5) i Given an observation, a good set of mean-field parameters can be determined by alternating between updating the generative parameters and the mean-field parameters. The update procedure is guaranteed to converge to a minimum of F, as long as the updates are done asynchronously and each update minimizes F with respect to a parameter. s The update equations for the mean-field parameters are: y = qi - 2&+ (f z2E i q,.w,. (f2 + (f2 y z (f y 7riP(Y Ii) E j 7rjp(yli) (6) (7) In our simulations, we hold most of the parameters of the generative model constant, such as the priors 7r, the weights w, and the generative noise in the observation, (f z. The only aspect that changes is the generative noise in the state, (f Y' which is a single parameter shared by all attractors: 2= d 1~ L..Jqi I'Y -Wi 12 (fy (8) i The updates of Equations 6-8 can be in any order. We typically initialize the state y to & at time 0, and then cyclically update the qi, (fy, then y. This generative model avoids the problem of spurious attractors described above for the standard Gaussian mixture model. Intuition into how the model avoids spurious attractors can be gained by inspecting the update equations. These equations effectively tie together two processes: moving y closer to some Wi than the others, and increasing the corresponding responsibility qi. As these two processes evolve together, they act to descrease the noise (fy, which accentuates the pull of the attractor. Thus stable points that do not correspond to the attractors are rare. 2 SIMULATION STUDIES To create an attractor net, we specify the parameters (7ri' w;) associated with the attractors based on the desired structure of the energy landscape (e.g., Figure Ib). The only remaining free parameter, (f z, plays an important role in determining how responsive the system is to the external input. We have conducted several simulation studies to explore properties of localist attractor networks. Systematic investigations with a 200-dimensional state space and 200 attractors, randomly placed at corners of the 200-D hypercube, have demonstrated that spurious responses are exceedingly rare unless more than 85% of an input's features are distorted (Figure 2), and that manipulating parameters such as noise and prior probabilities has the predicted effects. We have also conducted studies of localist attractor networks in the domain of visual images of faces. These simulations have shown that gang effects arise when there is structure among the attractors. For example, when the attractor set consists of a single view of several different faces, and multiple views of one face, then an input that is a morphed face-a linear combination of one of the single-view faces and one view of the gang face-will end up in the gang attractor even when the initial weighting assigned to the gang face was less than 40%. A Generative Modelfor Attractor Dynamics 85 100 ~~~==~~~~--------~~loo % Missing features Figure 2: The input must be severely corrupted before the net makes spurious (final state not at an attractor) or adulterous (final state at a neighbor of the generating attractor) responses. (a) The percentage of spurious responses increases as (Tz is increased. (b) The percentage of adulterous responses increases as (Tz is decreased. To test the architecture on a larger, structured problem, we modeled the domain of three-letter English words. The idea is to use the attractor network as a content addressable memory which might, for example, be queried to retrieve a word with P in the third position and any letter but A in the second position, a word such as HIP. The attractors consist of the 423 three-letter English words, from ACE to ZOO. The state space of the attractor network has one dimension for each of the 26 letters of the English alphabet in each of the 3 positions, for a total of 78 dimensions. We can refer to a given dimension by the letter and position it encodes, e.g., P3 denotes the dimension corresponding to the letter P in the third position of the word. The attractors are at the comers of a [-1, +1]18 hypercube. The attractor for a word such as HIP is located at the state having value -Ion all dimensions except for HI, h, and P3 which have value +1. The external input specifies a state that constrains the solution. For example, one might specify lip in the third position" by setting the external input to +1 on dimension P3 and to -Ion dimensions ll'3, for all letters ll' other than P. One might specify the absence of a constraint in a particular letter position, p, by setting the external input to a on dimensions ll'p, for all letters ll'. The network's task is to settle on a state corresponding to one of the words, given soft constraints on the letters. The interactive-activation model of word perception7 performs a similar computation, and our implementation exhibits the key qualitative properties of their model. If the external input specifies a word, of course the attractor net will select that word. Interesting queries are those in which the external input underconstrains or overconstrains the solution. We illustrate with one example of the network's behavior, in which the external input specifies D 1 , E2 , and G3 ? Because DEG is a nonword, no attractor exists for that state. The closest attractors share two letters with DEG, e.g., PEG, BEG, DEN, and DOG. Figure 3 shows the effect of gangs on the selection of a response, BEG. 3 CONCLUSION Localist attractor networks offer an attractive alternative to standard attractor networks, in that their dynamics are easy to control and adapt. We described a statistical formulation of a type of localist attractor, and showed that it provides a Lyapunov function for the system as well as a mathematical interpretation for the network parameters. The dynamics of this system are derived not from intuitive arguments but from this formal mathematical model. Simulation studies show that the architecture achieves gang effects, and spurious attractors are rare. This approach is inefficient if the attractors have compositional structure, but for many applications of pattern recognition or associative memory, the number of items R. S. Zemel and M C. Mozer 86 Iteration I ~?;s~., ~?; ~: f.? DUG In ~ ~ . . ' j:. . h' f.j.~?.;:: .~:)~.: BEG ;';;,"',' ~.:;?:( ;: _""- '.~ ,'<};~ '';'Uoli :.':(; l;,;)f f:.:: ~+t: ~. <::~'i . ;;,.'.. ,:~;.:,,: >.~~~-t".- r::';"' ~ '.~r.': ,' :: "; ;.U'> .?..?? :; " 3: LEG '~~', u:':,: ; '~: .F? DEN DEW F,.; ~,': :? ' :. :.:':. . ~:. :.: :.< . ,:;'.:: ?'>:'?J 1:/-:: 1:S~ ':?:: DIG ,,':i_ 'a, ~ .... ~ , ~ ~~::, DOG ': L ?<~ ; ,~.: , ;.:~: ! ~ '<f" ' /:.\~~ ~.!.;, ~:, ~ '.~':";:, ;. ':.' ~y'.)' :;',~ :_'>":'.? ':.';.:; ~;.;:: :. ;.. ( ~ .. '.':~ ",j' ":r.:r :.~.!. PEa -'::'.? ,.y Iteration 2 LEG PEG Iteration 3 lEG lEG PEG Iteration 4 lEG Iteration 5 lEG Figure 3: Simulation of the 3-letter word attractor network, queried with DEC. Each frame shows the relative activity of attractor units at various points in processing. Activity in each frame is normalized such that the most active unit is printed in black ink; the lighter the ink color, the less active the unit. Only attractor units sharing at least one letter with DEC are shown. The selection, BEG, is a product of a gang effect. The gangs in this example are formed by words sharing two letters. The most common word beginnings are PE- (7 instances) and DI- (6); the most common word endings are -AC (10) and -ET (10); the most common first-last pairings are B-G (5) and D-G (3). One of these gangs supports B1, two support E2 , and three support G3 , hence BEG is selected. being stored is small. The approach is especially useful in cases where attractor locations are known, and the key focus of the network is the mutual influenre of the attractors, as in many cognitive modelling studies. References [1) Becker, s., Moscovitch, M, Belumann,M, & Joordens, S. (1997). Long-term semantic priming: A computational account and empirical evidence. TOIlTtUIl of Experimental Psychology: Learning, Memory, & Cognition, 23(5), 10591082. [2) Golden, R (1988). Probabilistic characterization ofneura1modelcomputations. In D. Z. Anderson (Ed.), Neural Infrmnation Processing Systems (pp. 310-316). American Institute of PhysiCS. [3) Hopfield, J. J. (1982). Neural networks and physical systems with emergent collective computational abilities. Proceedings of the NationlU Aaulemy of Sciences, 79,2554-2558. [4) Kay, LM., Lancaster. L.R, & Freeman W.J. (1996). Reafference and attractors in the olfactory system during odor recognition. Int JNeural Systems, 7(4),489-95. [5) Mathis, D. (1997). A computational theory of consciousness in cognition. Unpublished Doctoral Dissertation. Boulder. CO: Department of Computer Science, University of Colorado. [6) Mathis, D., & Mozer, M. C. (1996). Conscious and unconscious perception: A computational theory. In G. Cottrell (Ed.), Proceedings of the Eighteenth Annual Conference of the Cognitive Science Society (pp. 324-328). Erlbaum. [7) McC1e11and, J. L. & Rumelhart, D. E. (1981). An interactive activation model of context effects in letter perception: Part L An account of basic findings. Psychological Reuiew, 88,375-407. [8) Neal. R M & Hinton, G. E. (1998). A view of the EM algorithm that justifies incremental, sparse, and other variants. In M I. Jordan (Ed.), Learning in Graphical Models. Kluwer Academic Press. [9) McRae, K., de Sa, V. R, & SeidenbeIg, M S. (1997) On the nature and scope of featural representations of word meaning. TournaI of Experimental Psychology; General, 126(2),99-130. [10) Redish, A. D. & Touretzky, D. S. (1998). The role of the hippocampus in solving the Morris water maze. Neural Computation, 10(1), 73-111. [11) Rodrigues, N. c., & Fontanari, J. F. (1997). Multiva1ley structure of attractor neural networks. Journal of Physics A (Mathematical and General), 30, 7945-7951 . [12) Samsonovich, A. & McNaughton, B. L. (1997) Path integration and cognitive mapping in a continuous attractor neural network model TournaI ofNeumscience, 17(15),5900-5920. [13) Saul, L.K., Jaakkola, T., & Jordan, ML (1996). Mean field theory for sigmoid belief networks. Tounud of AI Resetm:h, 4, 61-76. PART II NEUROSCIENCE
1695 |@word eliminating:1 proportion:2 hippocampus:1 simulation:7 covariance:1 solid:1 accommodate:1 initial:3 selecting:2 rightmost:1 existing:1 current:1 wd:1 activation:2 yet:1 must:1 cottrell:1 visible:1 partition:2 midway:1 shape:2 update:8 generative:15 selected:1 item:2 beginning:1 dissertation:1 provides:2 characterization:2 location:4 mathematical:5 persistent:1 incorrect:1 qualitative:2 consists:5 pairing:1 behavioral:1 olfactory:1 indeed:1 behavior:4 samsonovich:1 uz:1 freeman:1 spherical:1 cpu:2 actual:1 increasing:1 estimating:1 minimizes:1 finding:2 temporal:1 act:1 golden:1 interactive:2 tie:1 tricky:2 control:2 unit:13 broadened:1 ly:2 before:1 local:1 severely:1 encoding:3 path:1 might:4 black:1 doctoral:1 co:2 bi:1 procedure:6 addressable:1 empirical:1 significantly:1 printed:1 word:15 radial:1 get:1 cannot:1 selection:3 context:1 influence:2 restriction:4 map:2 demonstrated:1 missing:2 center:1 eighteenth:1 straightforward:1 starting:1 pull:2 kay:1 retrieve:1 stability:1 mcnaughton:1 hierarchy:1 play:1 colorado:3 carved:1 rip:2 lighter:1 homogeneous:1 unconscious:1 designing:1 rodrigues:1 origin:3 rumelhart:1 recognition:2 updating:1 located:1 role:2 region:1 wj:1 mozer:6 intuition:1 constrains:1 dynamic:16 depend:1 solving:1 creates:1 f2:2 basis:2 comer:1 hopfield:1 emergent:2 represented:1 various:1 alphabet:1 describe:3 query:1 zemel:5 neighborhood:1 lancaster:1 sculpt:1 whose:1 ace:1 larger:1 ability:1 noisy:1 asynchronously:1 final:2 associative:1 net:19 accentuates:1 propose:1 interaction:1 product:1 translate:1 mixing:1 intuitive:1 az:1 convergence:2 produce:5 generating:1 incremental:1 derive:1 illustrate:1 completion:2 fixing:1 ac:1 sa:1 implemented:1 predicted:1 involves:1 beg:5 lyapunov:1 stochastic:1 pea:1 human:1 settle:1 fix:1 investigation:1 inspecting:1 adjusted:1 hold:1 proximity:2 considered:3 exp:1 cognition:3 scope:1 mapping:1 lm:1 achieves:1 adopt:1 create:1 reflects:1 weighted:3 clearly:1 gaussian:15 rather:3 jaakkola:1 encode:1 derived:2 focus:1 modelling:1 likelihood:4 attract:2 typically:2 entire:2 spurious:9 hidden:2 manipulating:3 semantics:1 among:2 ill:2 denoted:2 integration:1 initialize:1 mutual:1 field:6 aware:1 having:1 identical:1 represents:1 filling:1 future:1 others:1 richard:1 randomly:1 composed:1 attractor:117 olfaction:1 adjust:1 mixture:13 predefined:1 integral:1 closer:1 partial:1 unless:1 desired:2 psychological:3 increased:1 formalism:1 hip:2 soft:1 instance:1 maximization:3 localist:13 deviation:1 rare:3 uniform:2 conducted:2 erlbaum:1 loo:1 characterize:1 stored:1 corrupted:1 density:2 fundamental:1 systematic:1 participates:2 yl:2 probabilistic:1 physic:2 michael:1 together:2 iy:4 yg:1 choose:1 external:8 stochastically:1 corner:1 tz:2 inefficient:1 cognitive:3 american:1 account:3 de:1 redish:1 coefficient:3 int:1 depends:1 view:5 responsibility:2 start:1 complicated:1 dug:1 minimize:1 formed:1 variance:1 yield:1 clarifies:1 landscape:5 correspond:2 produced:2 zoo:1 trajectory:2 notoriously:2 dig:1 history:2 influenced:2 touretzky:1 sharing:2 ed:3 energy:4 pp:2 e2:2 proof:2 associated:3 di:1 adjusting:1 knowledge:3 color:1 tum:1 specify:3 response:5 formulation:8 done:1 anderson:1 name:1 effect:13 normalized:4 counterpart:1 hence:2 assigned:2 alternating:1 iteratively:1 consciousness:1 semantic:1 neal:1 yl2:1 attractive:1 wiring:1 ll:4 during:1 performs:1 meaning:1 image:1 instantaneous:1 common:3 sigmoid:1 physical:1 extend:1 interpretation:5 kluwer:1 interpret:1 refer:1 morphed:1 queried:2 ai:1 moving:2 specification:2 stable:2 posterior:9 closest:1 recent:1 showed:1 yi:1 lnp:3 seen:1 minimum:1 dew:1 converge:1 maximize:1 dashed:1 ii:2 multiple:3 desirable:1 faster:1 characterized:1 adapt:1 offer:1 long:3 academic:1 visit:1 jy:1 qi:12 calculates:2 variant:1 basic:1 expectation:3 nonword:1 iteration:5 represent:2 ion:3 dec:2 preserved:1 addition:1 want:1 decreased:1 unorthodox:1 jordan:2 near:1 intermediate:1 easy:1 fit:2 psychology:3 equidistant:1 architecture:4 reduce:1 idea:1 iocalist:1 unidimensional:1 intensive:2 expression:1 motivated:2 assist:1 becker:1 reinforcing:1 cause:1 compositional:1 useful:2 generally:2 clear:1 ylx:1 conscious:1 morris:1 fz:1 specifies:3 percentage:2 peg:3 inhibitory:1 delta:1 neuroscience:1 discrete:3 thereafter:1 key:3 achieving:1 drawn:1 modelfor:1 sum:2 ylz:1 parameterized:1 letter:15 distorted:1 p3:3 dy:2 layer:1 bound:1 hi:1 followed:1 guaranteed:1 arizona:2 annual:1 activity:7 gang:14 occur:1 strength:3 constraint:2 ri:2 encodes:1 aspect:1 anyone:1 argument:1 department:3 structured:1 combination:1 across:1 describes:1 em:2 wi:5 g3:2 leg:6 den:2 gradually:1 boulder:2 equation:7 mutually:1 tractable:1 serf:1 unusual:1 end:2 gaussians:10 appropriate:1 robustly:1 responsive:1 alternative:4 odor:1 denotes:1 remaining:1 graphical:1 especially:1 hypercube:2 society:1 ink:2 implied:1 objective:2 arrangement:1 exhibit:1 gradient:1 distance:1 mapped:1 fontanari:1 fy:5 trivial:1 toward:1 mcrae:1 water:1 modeled:1 providing:1 minimizing:1 optionally:1 unfortunately:1 negative:1 implementation:1 collective:1 allowing:2 upper:2 yli:4 observation:27 i_:1 descent:1 situation:1 hinton:1 frame:2 arbitrary:1 introduced:1 fort:1 dog:2 specified:1 unpublished:1 connection:3 usually:1 pattern:8 below:1 perception:2 tucson:1 including:1 memory:4 explanation:1 belief:1 overlap:1 difficulty:2 featural:1 prior:5 evolve:1 determining:1 asymptotic:1 relative:1 interesting:2 localized:1 generator:1 basin:10 share:1 lo:1 course:1 placed:1 last:1 free:3 english:3 formal:1 deeper:1 pulled:2 institute:1 fall:1 explaining:1 face:7 neighbor:1 saul:1 sparse:2 distributed:3 dimension:9 evaluating:1 avoids:2 ending:1 exceedingly:1 maze:1 qx:1 approximate:1 preferred:1 neurobiological:1 deg:2 ml:1 active:2 b1:1 continuous:2 search:2 iterative:1 lip:1 nature:1 priming:2 mathis:2 domain:3 spread:1 noise:6 arise:2 reafference:1 ny:2 position:8 lie:2 pe:1 ib:1 third:4 weighting:1 cyclically:1 down:1 virtue:1 evidence:1 consist:1 exists:1 effectively:1 gained:1 magnitude:1 conditioned:2 illustrates:1 justifies:1 easier:1 depicted:1 likely:2 explore:1 visual:1 corresponds:5 conditional:1 identity:1 viewed:1 shared:1 absence:1 content:1 change:2 determined:2 except:1 total:1 experimental:2 select:5 internal:1 support:3 moscovitch:1 evaluate:1
760
1,696
Robust Recognition of Noisy and Superimposed Patterns via Selective Attention Soo-Young Lee Brain Science Research Center Korea Advanced Institute of Science & Technology Yusong-gu, Taejon 305-701 Korea Michael C. Mozer Department of Computer Science University of Colorado at Boulder Boulder, CO 80309 USA sylee@ee.kaist.ac.kr mozer@cs.colorado.edu Abstract In many classification tasks, recognition accuracy is low because input patterns are corrupted by noise or are spatially or temporally overlapping. We propose an approach to overcoming these limitations based on a model of human selective attention. The model, an early selection filter guided by top-down attentional control, entertains each candidate output class in sequence and adjusts attentional gain coefficients in order to produce a strong response for that class. The chosen class is then the one that obtains the strongest response with the least modulation of attention. We present simulation results on classification of corrupted and superimposed handwritten digit patterns, showing a significant improvement in recognition rates. The algorithm has also been applied in the domain of speech recognition, with comparable results. 1 Introduction In many classification tasks, recognition accuracy is low because input patterns are corrupted by noise or are spatially or temporally overlapping. Approaches have been proposed to make classifiers more robust to such perturbations, e.g., by requiring classifiers to have low input-to-output mapping sensitivity [1]. We propose an approach that is based on human selective attention. People use selective attention to focus on critical features of a stimulus and to suppress irrelevant features. It seems natural to incorporate a selective-attention mechanism into pattern recognition systems for noisy real world applications. Psychologists have for many years studied the mechanisms of selective attention (e.g., [2]-[4]). However, controversy still exists among competing theories, and only a few models are sufficiently well defmed to apply to engineering pattern recognition problems. Fukushima [5] has incorporated selective attention and attention-switching algorithms into his Neocognitron model, and has demonstrated good recognition performance on superimposed digits. However, the Neocognitron model has many unknown parameters which must be determined heuristically, and its performance is sensitive to the parameter values. Also, its computational requirements are prohibitively expensive for many realtime applications. Rao [6] has also recently introduced a selective attention model based 32 s. -Y. Lee and M C. Mozer on Kalman filters and demonstrated classifications of superimposed patterns. However, his model is based on linear systems, and a nonlinear extension is not straightforward. There being no definitive approach to incorporating selective attention into pattern recognition, we propose a novel approach and show it can improve recognition accuracy. 2 Psychological Views of Selective Attention The modern study of selective attention began with Broadbent [7]. Broadbent presented two auditory channels to subjects, one to each ear, and asked subjects to shadow one channel. He observed that although subjects could not recall most of what took place in the unshadowed channel, they could often recall the last few seconds of input on that channel. Therefore, he suggested that the brain briefly stores incoming stimuli but the stimulus information fades and is neither admitted to the conscious mind nor is encoded in a way that would permit later recollection, unless attention is directed toward it. This view is known as an early filtering or early selection model. Treisman [8] proposed a modification to this view in which the filter merely attenuates the input rather than absolutely preventing further analysis. Although late-selection and hybrid views of attention have been proposed, it is clear that early selection plays a significant role in human information processing [3]. The question about where attention acts in the stream of processing is independent of another important issue: what factors drive attention to select one ear or one location instead of another. Attention may be directed based on low-level stimulus features, such as the amplitude of a sound or the color of a visual stimulus. This type of attentional control is often called bottom up. Attention may also be directed based on expectations and object knowledge, e.g., to a location where critical task-relevant information is expected. This type of attentional control is often called top down. 3 A Multilayer Perceptron Architecture for Selective Attention We borrow the notion of an early selection filter with top-down control and integrate it into a multilayer perceptron (MLP) classifier, as depicted in Figure 1. The dotted box is a standard MLP classifier, and an attention layer with one-to-one connectivity is added in front of the input layer. Although we have depicted an MLP with a single hidden layer, our approach is applicable to general MLP architectures. The kth element of the input vector, denoted xk, is gated to the kth input of the MLP by an attention gain or filtering coefficient ak. Previously, the first author has shown a benefit of treating the ak's like ordinary adaptive parameters during training [9]-[12]. In the present work, we fix the attention gains at 1 during training, causing the architecture to behave as an ordinary MLP. However, we allow the gains to be adjusted during classification of test patterns. Our basic conjecture is that recognition accuracy may be improved if attention can suppress noise along irrelevant dimensions and enhance a weak signal along relevant dimensions. "Relevant" and "irrelevant" are determined by topdown control of attention. Essentially, we use knowledge in the trained MLP to determine which input dimensions are critical for classifying a test pattern. To be concrete, consider an MLP trained to classify handwritten digits. When a test pattern is presented, we can adjust the attentional gains via gradient descent so as to make the input as good an example of the class "0" as possible. We do this for each of the different output classes, "0" through "9", and choose the class for which the strongest response is obtained with the smallest 33 Robust Pattern Recognition via Selective Attention attentional modulation (the exact quantitative rule is presented below). The conjecture is that if the net can achieve a strong response for a class by making a small attentional modulation, that class is more likely to be correct than whichever class would have been selected without applying selective attention . .........................................................................................................................................................._................,: ! ~ ~ Xl~----~~~~~--~~~~~~~~~~~~~yl 1 1 ~ ~--~-+~~~~~~~~~~~~~~ ~ ~ li ~------~~>E~~~~~~~~~~~~~~~~~. Y3 ! ! ~ ? ? x ~ i: a ; 'l: h w(l) vJ2) /M 1 i ................................................ __ ..........................................................................................................._... _ .....! Figure 1: MLP architecture for selective attention The process of adjusting the attentional gains to achieve a strong response from a particular class-<:all it the attention class-proceeds as follows. First, a target output vector t S= [1'1 1'2? ??t'M f is defmed. For bipolar binary output representations, t,S = 1 is for the attention class and -1 for the others. Second, the attention gain ak's are set to I. Third, the attention gain ak's are adapted to minimize error E S == ~ L (t,S - y,)2 with the given , input x = [XI X2??? XN f and pre-trained and frozen synaptic weights W. The update rule is based on a gradient-descent algorithm with error back-propagation. At the (n+ J)'th iterative epoch, the attention gain a k is updated as (Ia) (lb) where E denotes the attention output error, ojl) thej'th attribute of the back-propagated error at the first hidden-layer, and WJ~) the synaptic weight between the input x k and the j'th neuron at the first hidden layer. Finally, " is a step size. The attention gains are thresholded to lie in [0, 1]. The application of selective attention to a test example is summarized as follows: Step 1: Apply a test input pattern to the trained MLP and compute output values. Step 2: For each of the classes with top m activation values, (I) Initialize all attention gain ak' s to 1 and set the target vector tS. (2) Apply the test pattern and attention gains to network and compute output. (3) Apply the selective attention algorithm in Eqs.( 1) to adapt the attention gains. (4) Repeat steps (2) and (3) until the attention process converges. (5) Compute an attention measure M on the asymptotic network state. s.-Y. Lee and M. C. Mozer 34 Step 3: Select the class with a minimum attention measure M as the recognized class. The attention measure is defined as (2a) Dj =:f(Xk -xk )2/2N = Eo 2 :fxk(l-ak) =4[/i 2 (2b) /2N - y;(i)]2/2M , (2c) where D/ is the square of Euclidean distance between two input patterns before and after the application of selective attention and Eo is the output error after the application of selective attention. Here, D/ and Eo are normalized with the number of input pixels and number of output classes, respectively. The superscript s for attention classes is omitted for simplicity. To make the measure M a dimensionless quantity, one may normalize the D/ and Eo with the input energy (~kX~ ) and the training output error, respectively. However, it does not affect the selection process in Step 3. One can think of the attended input i as the minimal deformation of the test input needed to trigger the attended class, and therefore the Euclidean distance between x and i is a good measure for the classification confidence. In fact, D/ is basically the same quantity minimized by Rao [6]. However, the MLP classifier in our model is capable of nonlinear mapping between the input and output patterns. A nearest-neighbor classifier, with the training data as examples, could also be used to find the minimum-distance class. Our model with the MLP classifier computes a similar function without the large memory and computational requirements. The proposed selective attention algorithm was tested on recognition of noisy numeral patterns. The numeral database consists of samples of the handwritten digits (0 through 9) collected from 48 people, for a total of 480 samples. Each digit is encoded as a 16x16 binary pixel array. Roughly 16% of the pixels are black and coded as 1; white pixels are coded as O. Four experiments were conducted with different training sets of 280 training patterns each. A one hidden-layer MLP was trained by back propagation. The numbers of input, hidden, and output neurons were 256, 30, and to, respectively. Three noisy test patterns were generated from each training pattern by randomly flipping each pixel value with a probability Pt, and the 840 test patterns were presented to the network for classification. In Figure 2, the false recognition rate is plotted as a function of the number of candidates considered for the attentional manipulation, m. (Note that the run time of the algorithm is proportional to m, but that increasing m does not imply a more lax classification criterion, or additional external knowledge playing into the classification.) Results are shown for three different pixel inversion probabilities, Pt =0.05, 0.1, and 0.15. Considering the average 16% of black pixels in the data, the noisy input patterns with Pt= 0.15 correspond to a SNR of approximately 0 dB. For each condition in the figure, the false recognition rates for the four different training sets are marked with an '0', and the means are connected by the solid curve. 35 Robust Pattern Recognition via Selective Attention A standard MLP classifier corresponds to m = I (i.e., only the most active output of the MLP is considered as a candidate response). The false recognition rate is clearly lower when the attentional manipulation is used to select a response from the MLP (m > I). It appears that performance does not improve further by considering more than the top three candidates. 4 Attention Switching Superimposed Patterns 2 I/) Q) a; a: c:: 1.5 o -..::; ?c Ol o o Q) a: 3l ~ 0.5 o for Suppose that we superimpose the binary input patterns for two different handwritten digits using the logical OR operator (the pixels corresponding to the black ink have logical value 1). Can we use attention to recognize the two patterns in sequence? This is an extreme case of a situation that is common in visual pattern recognition-where two patterns are spatially overlapping. We explore the following algorithm. First, one pattern is recognized with the selective attention process used in Section 3. Second, attention is switched from the recognized pattern to the remaining pixels in the image. Switching is accomplished by removing attention from the pixels of the recognized pattern: the attentional gain of an input is clamped to 0 following switching if and only if its value after the first-stage selective attention process was 1 (Le., that input was attended during the recognition of the first pattern); all other gains are set to 1. Third, the recognition process with selective attention is performed again to recognize the second pattern. The proposed selective attention and attention switching algorithm was tested for recognition of 2 superimposed numeral data. Again, four experiments were conducted with l\ r---.. 2 3 4 5 Number of Candidates (a) Pr=O.05, 20r-~----~----~----~----~-. I/) Q) Cii a: c:: 15 ,g ?c g> o 10 Q) a: 3l ~ 5 OL-~----~----~----~----~~ 2 3 4 5 Number of Candidates (b) Pr=O.IO, 40r-~----~----~----~----~-. I/) Q) a; a: c:: 30 o :t: c:: Ol 8 20 Q) a: 3l ~ 10 OL-~----~----~----~----~~ 2 3 4 5 Number of Candidates (c) Pr=O.15, Figure 2: False recognition rates for noisy patterns as a function of the number of top candidates. Each binary pixel of training patterns is randomly inverted with a probability Pr. s.-Y. 36 Lee and M C. Mozer different training sets. For each experiment, 40 patterns were selected from 280 training patterns, and 720 test patterns were generated by superimposing pairs of patterns from different output classes. The test patterns were still binary. .-. ~ I??? ~ ~~ ... ~ ~ [kJ ~ [9] [9] Et][3] ~[2J0[J ~~g]m Figure 3: Examples of Selective Attention and Attention Switching Figure 3 shows six examples of the selective attention and attention switching algorithm in action, each consisting of four panels in a horizontal sequence. The six examples were fonned by superimposing instances of the following digit pairs: (6,3), (9,0), (6,4), (9,3), (2,6), and (5,2). The fIrst panel for each example shows the superimposed pattern. The second panel shows the attended input i for the fIrst round classifIcation; because this input has continuous values, we have thresholded the values at 0.5 to facilitate viewing in the fIgure. The third panel shows the masking pattern for attention switching, generated by thresholding the input pattern at 1.0. The fourth panel sho~s the residual input pattern for the second round classifIcation. The attended input x has analog values, but thresholded by 0.5 to be shown in the second rectangles. Figure 3 shows that attention switching is done effectively, and the remaining input patterns to the second classifIer are quite visible. We compared perfonnance for three different methods. First, we simply selected the two MLP outputs with highest activity; this method utilizes neither selective attention. Second, we perfonned attention switching but did not apply selective attention (i.e., m=I). Third, we perfonned both attention switching and selective attention (with m=3). Table I summarizes the recognition rates for the fITst and the second patterns read out of the MLP for the three methods. As hypothesized, attention switching increases the recognition rate for the second pattern, and selective attention increases the recognition rate for both the fITSt and the second pattern. Table I: Recognition Rates (%) of Two Superimposed Numeral Patterns No selective attention or switching Switching only Switching & selective attention First Pattern Second Pattern 91. 3 91. 3 95.9 62.7 75.4 77.4 Robust Pattern Recognition via Selective Attention 37 5 Conclusion In this paper, we demonstrated a selective-attention algorithm for noisy and superimposed patterns that obtains improved recognition rates. We also proposed a simple attention switching algorithm that utilizes the selective-attention framework to further improve performance on superimposed patterns. The algorithms are simple and easily implemented in feedforward MLPs. Although our experiments are preliminary, they suggest that attention-based algorithms will be useful for extracting and recognizing multiple patterns in a complex background. We have conducted further simulation studies supporting this conjecture in the domain of speech recognition, which we will integrate into this presentation if it is accepted at NIPS. Acknowledgements S.Y. Lee acknowledges supports from the Korean Ministry of Science and Technology. We thank Dr. Y. Le Cun for providing the handwritten digit database. References [1] Jeong D.G., and Lee, S.Y. (1996). Merging backpropagation and Hebbian learning rules for robust classification, Neural Networks, 9:1213-1222. [2] Cowan, N. (1997). Attention and Memory: An Integrated Framework, Oxford Univ. Press. [3] Pashler, H.E. (1998). The Psychology ofAttention, MIT Press. [4] Parasuraman, R. (ed.) (1998). The Attentive Brain, MIT Press. [5] Fukushima, K. (1987). Neural network model for selective attention in visual pattern recognition and associative recall, Applied Optics, 26:4985-4992. [6] Rao, R.P.N. (1998). Correlates of attention in a model of dynamic visual recognition. In Neural Information Processing Systems 10, MIT Press. [7] Broadbent, D.E. (1958). Perception and Communication. Pergamon Press. [8] Treisman, A. (1960). Contextual cues in selective listening, Quarterly Journal of Experimental Psychology, 12:242-248. [9] Lee, H.J., Lee, S.Y. Lee, Shin, S.Y., and Koh, B.Y. (1991). TAG: A neural network model for large-scale optical implementation, Neural Computation, 3:135-143. [1O]Lee, S.Y., Jang, J.S., Shin, S.Y., & Shim, C.S. (1988). Optical Implementation of Associative Memory with Controlled Bit Significance, Applied Optics, 27:19211923. [11 ] Kruschke, J.K. (1992). ALCOVE: An Examplar-Based Connectionist Model of Category Learning, Psychological Review, 99:22-44. [12]Lee, S.Y., Kim, D.S., Abn, K.H., Jeong, J.H., Kim, H., Park, S.Y., Kim, L.Y., Lee, J.S., & Lee, H.Y. (1997). Voice Command II: a DSP implementation of robust speech recognition in real-world noisy environments, International Conference on Neural Information Processing, pp. 1051-1054, Dunedin, New Zealand.
1696 |@word briefly:1 inversion:1 seems:1 heuristically:1 simulation:2 attended:5 solid:1 contextual:1 activation:1 must:1 visible:1 treating:1 update:1 cue:1 selected:3 xk:3 location:2 along:2 consists:1 expected:1 roughly:1 nor:1 brain:3 ol:4 considering:2 increasing:1 panel:5 what:2 quantitative:1 y3:1 act:1 bipolar:1 prohibitively:1 classifier:9 control:5 before:1 engineering:1 io:1 switching:16 ak:6 oxford:1 modulation:3 approximately:1 black:3 studied:1 co:1 directed:3 backpropagation:1 digit:8 shin:2 j0:1 pre:1 confidence:1 suggest:1 selection:6 operator:1 applying:1 dimensionless:1 pashler:1 demonstrated:3 center:1 straightforward:1 attention:81 kruschke:1 zealand:1 simplicity:1 fade:1 adjusts:1 rule:3 array:1 borrow:1 his:2 notion:1 updated:1 target:2 trigger:1 play:1 colorado:2 exact:1 pt:3 suppose:1 element:1 recognition:32 expensive:1 database:2 observed:1 role:1 bottom:1 wj:1 connected:1 highest:1 mozer:5 environment:1 asked:1 dynamic:1 controversy:1 trained:5 gu:1 examplar:1 easily:1 univ:1 quite:1 encoded:2 kaist:1 tested:2 think:1 noisy:8 superscript:1 associative:2 sequence:3 frozen:1 net:1 took:1 propose:3 causing:1 relevant:3 achieve:2 normalize:1 requirement:2 produce:1 converges:1 object:1 ac:1 nearest:1 eq:1 strong:3 implemented:1 c:1 shadow:1 guided:1 correct:1 attribute:1 filter:4 human:3 viewing:1 numeral:4 fix:1 preliminary:1 adjusted:1 extension:1 sufficiently:1 considered:2 mapping:2 sho:1 early:5 smallest:1 omitted:1 applicable:1 sensitive:1 alcove:1 mit:3 clearly:1 rather:1 command:1 focus:1 dsp:1 improvement:1 superimposed:10 kim:3 integrated:1 hidden:5 selective:38 pixel:11 issue:1 classification:12 among:1 denoted:1 initialize:1 park:1 minimized:1 others:1 stimulus:5 connectionist:1 few:2 modern:1 randomly:2 recognize:2 consisting:1 fukushima:2 mlp:18 adjust:1 extreme:1 capable:1 korea:2 perfonnance:1 unless:1 euclidean:2 plotted:1 deformation:1 minimal:1 psychological:2 instance:1 classify:1 rao:3 ordinary:2 snr:1 recognizing:1 conducted:3 front:1 corrupted:3 international:1 sensitivity:1 lee:13 yl:1 michael:1 treisman:2 enhance:1 concrete:1 connectivity:1 again:2 ear:2 choose:1 dr:1 external:1 li:1 summarized:1 coefficient:2 stream:1 later:1 view:4 performed:1 masking:1 minimize:1 square:1 mlps:1 accuracy:4 correspond:1 weak:1 handwritten:5 basically:1 drive:1 strongest:2 synaptic:2 ed:1 attentive:1 energy:1 pp:1 propagated:1 gain:15 auditory:1 adjusting:1 logical:2 recall:3 color:1 knowledge:3 amplitude:1 back:3 appears:1 response:7 improved:2 done:1 box:1 stage:1 until:1 horizontal:1 broadbent:3 nonlinear:2 overlapping:3 propagation:2 facilitate:1 usa:1 hypothesized:1 requiring:1 normalized:1 yusong:1 spatially:3 read:1 white:1 round:2 during:4 defmed:2 criterion:1 neocognitron:2 image:1 novel:1 recently:1 began:1 common:1 analog:1 he:2 significant:2 dj:1 irrelevant:3 manipulation:2 store:1 binary:5 accomplished:1 inverted:1 minimum:2 additional:1 ministry:1 cii:1 eo:4 recognized:4 determine:1 signal:1 ii:1 multiple:1 sound:1 hebbian:1 adapt:1 coded:2 controlled:1 basic:1 multilayer:2 essentially:1 expectation:1 background:1 subject:3 db:1 cowan:1 extracting:1 ee:1 feedforward:1 affect:1 psychology:2 architecture:4 competing:1 listening:1 six:2 speech:3 action:1 useful:1 clear:1 conscious:1 category:1 dotted:1 four:4 neither:2 thresholded:3 rectangle:1 merely:1 year:1 run:1 fourth:1 place:1 utilizes:2 realtime:1 summarizes:1 taejon:1 comparable:1 bit:1 layer:6 activity:1 adapted:1 optic:2 x2:1 tag:1 thej:1 optical:2 conjecture:3 department:1 cun:1 modification:1 making:1 psychologist:1 pr:4 boulder:2 koh:1 previously:1 mechanism:2 needed:1 mind:1 whichever:1 lax:1 permit:1 apply:5 quarterly:1 jang:1 voice:1 top:6 denotes:1 remaining:2 ink:1 pergamon:1 added:1 question:1 quantity:2 flipping:1 gradient:2 kth:2 distance:3 attentional:11 thank:1 recollection:1 collected:1 sylee:1 toward:1 kalman:1 providing:1 korean:1 suppress:2 attenuates:1 implementation:3 unknown:1 gated:1 neuron:2 descent:2 behave:1 t:1 supporting:1 situation:1 incorporated:1 communication:1 vj2:1 perturbation:1 lb:1 overcoming:1 superimpose:1 introduced:1 pair:2 jeong:2 nip:1 suggested:1 topdown:1 below:1 pattern:56 proceeds:1 perception:1 soo:1 memory:3 ia:1 critical:3 perfonned:2 natural:1 hybrid:1 residual:1 advanced:1 improve:3 technology:2 imply:1 temporally:2 acknowledges:1 entertains:1 kj:1 epoch:1 review:1 acknowledgement:1 asymptotic:1 shim:1 limitation:1 filtering:2 proportional:1 integrate:2 switched:1 thresholding:1 classifying:1 playing:1 repeat:1 last:1 allow:1 perceptron:2 institute:1 neighbor:1 benefit:1 curve:1 dimension:3 xn:1 world:2 computes:1 preventing:1 author:1 adaptive:1 fonned:1 correlate:1 obtains:2 ojl:1 incoming:1 active:1 xi:1 continuous:1 iterative:1 table:2 channel:4 robust:7 complex:1 domain:2 did:1 significance:1 noise:3 definitive:1 x16:1 xl:1 candidate:8 lie:1 clamped:1 late:1 third:4 young:1 down:3 removing:1 fxk:1 showing:1 exists:1 incorporating:1 false:4 merging:1 effectively:1 kr:1 kx:1 depicted:2 admitted:1 simply:1 likely:1 explore:1 visual:4 corresponds:1 marked:1 presentation:1 determined:2 called:2 total:1 accepted:1 superimposing:2 experimental:1 select:3 people:2 support:1 absolutely:1 incorporate:1 fitst:2
761
1,697
Statistical Dynamics of Batch Learning s. Li and K. Y. Michael Wong Department of Physics, Hong Kong University of Science and Technology Clear Water Bay, Kowloon, Hong Kong {phlisong, phkywong}@ust.hk Abstract An important issue in neural computing concerns the description of learning dynamics with macroscopic dynamical variables. Recent progress on on-line learning only addresses the often unrealistic case of an infinite training set. We introduce a new framework to model batch learning of restricted sets of examples, widely applicable to any learning cost function, and fully taking into account the temporal correlations introduced by the recycling of the examples. For illustration we analyze the effects of weight decay and early stopping during the learning of teacher-generated examples. 1 Introduction The dynamics of learning in neural computing is a complex multi-variate process. The interest on the macroscopic level is thus to describe the process with macroscopic dynamical variables. Recently, much progress has been made on modeling the dynamics of on-line learning, in which an independent example is generated for each learning step [1, 2]. Since statistical correlations among the examples can be ignored, the dynamics can be simply described by instantaneous dynamical variables. However, most studies on on-line learning focus on the ideal case in which the network has access to an almost infinite training set, whereas in many applications, the collection of training examples may be costly. A restricted set of examples introduces extra temporal correlations during learning, and the dynamics is much more complicated. Early studies briefly considered the dynamics of Adaline learning [3, 4, 5], and has recently been extended to linear perceptrons learning nonlinear rules [6, 7}. Recent attempts, using the dynamical replica theory, have been made to study the learning of restricted sets of examples, but so far exact results are published for simple learning rules such as Hebbian learning, beyond which appropriate approximations are needed [8]. In this paper, we introduce a new framework to model batch learning of restricted sets of examples, widely applicable to any learning rule which minimizes an arbitrary cost function by gradient descent. It fully takes into account the temporal correlations during learning, and is therefore exact for large networks. 287 Statistical Dynamics ofBatch Learning 2 Formulation Consider the single layer perceptron with N ? 1 input nodes {~j} connecting to a single output node by the weights {Jj }. For convenience we assume that the inputs ~j are Gaussian variables with mean 0 and variance 1, and the output state 5 is a function f(x) of the activation x at the output node, i.e. 5=f(x); x=J?{. (1) = The network is assigned to "learn" p aN examples which map inputs {{j} to the outputs {5~} (p = 1, ... ,p). 5~ are the outputs generated by a teacher percept ron {Bj }, namely (2) Batch learning by gradient descent is achieved by adjusting the weights {Jj } iteratively so that a certain cost function in terms of the student and teacher activations {x~} and {y~} is minimized. Hence we consider a general cost function (3) The precise functional form of g(x, y) depends on the adopted learning algorithm. For the case of binary outputs, f(x) = sgnx. Early studies on the learning dynamics considered Adaline learning [3, 4, 5], where g(x, y) = -(5 - x)2/2 with 5 = sgny. For recent studies on Hebbian learning [8], g(x,y) = x5. To ensure that the perceptron is regularized after learning, it is customary to introduce a weight decay term. Furthermore, to avoid the system being trapped in local minima, noise is often added in the dynamics. Hence the gradient descent dynamics is given by dJj (t) -_ N1 " ", ( ) ) I' ) ( ) (4) -;u~g (x~ t ,yl' ~j -).Jj(t +"1j t, ~ where, here and below, g' (x, y) and gil (x, y) respectively represent the first and second partial derivatives of g(x, y) with respect to x. ). is the weight decay strength, and "1j(t) is the noise term at temperature T with (5) 3 The Cavity Method Our theory is the dynamical version of the cavity method [9, 10, 11]. It uses a self-consistency argument to consider what happens when a new example is added to a training set. The central quantity in this method is the cavity activation, which is the activation of a new example for a perceptron trained without that example. Since the original network has no information about the new example, the cavity activation is stochastic. Specifically, denoting the new example by the label 0, its cavity activation at time t is ho(t) = J(t) . f1. (6) For large N and independently generated examples, ho(t) is a Gaussian variable. Its covariance is given by the correlation function G(t, s) of the weights at times t and s, that is, (7) (ho(t)ho(s?) = J(t). J(s) == G(t,s), 288 S. Li and K. Y. M. Wong where ~J and ~2 are assumed to be independent for j i- k. The distribution is further specified by the teacher-student correlation R(t), given by (ho(t)yo) = j(t) . jj = R(t). (8) Now suppose the perceptron incorporates the new example at the batch-mode learning step at time s. Then the activation of this new example at a subsequent time t > s will no longer be a random variable. Furthermore, the activations of the original p examples at time t will also be adjusted from {xJl(t)} to {x~(t)} because of the newcomer, which will in turn affect the evolution of the activation of example 0, giving rise to the so-called Onsager reaction effects. This makes the dynamics complex, but fortunately for large p '" N, we can assume that the adjustment from xJl(t) to x2(t) is small, and perturbative analysis can be applied. Suppose the weights of the original and new perceptron at time t are {Jj (t)} and {JJ(t)} respectively. Then a perturbation of (4) yields (! + ,x) (.tj(t) - Jj(t? ~g'(xo(t),yo)~J = + ~ 2: ~fgll(XJl(t), YJl)~r(J'2(t) - Jk(t?. (9) Jlk The first term on the right hand side describes the primary effects of adding example the training set, and is the driving term for the difference between the two perceptrons. The second term describes the secondary effects due to the changes to the original examples caused by the added example, and is referred to as the On sager reaction term. One should note the difference between the cavity and generic activations of the added example. The former is denoted by ho(t) and corresponds to the activation in the perceptron {Jj (t) }, whereas the latter, denoted by Xo (t) and corresponding to the activation in the percept ron {.tj (t)}, is the one used in calculating the gradient in the driving term of (9). Since their notations are sufficiently distinct, we have omitted the superscript 0 in xo(t), which appears in the background examples x~(t). o to The equation can be solved by the Green's function technique, yielding .tj(t) - Jj(t) = 2:! dsGjk(t, s) (~g~(s)~2) , (10) k where g~(s) = g'(xo(s),yo) and Gjk(t, s) is the weight Green's function satisfying Gjk(t,S) = G(O)(t - S)6jk + ~ ~! dt'G(O)(t - t')~fg~(t')~rGik(t' - s), (11) Jl\ = G(O)(t - s) e(t - s) exp( -,x(t - s? is the bare Green's function, and e is the step function. The weight Green's function describes how the effects of example 0 propagates from weight Jk at learning time s to weight Jj at a subsequent time t, including both primary and secondary effects. Hence all the temporal correlations have been taken into account. For large N, the equation can be solved by a diagrammatic approach similar to [5]. The weight Green's function is self-averaging over the distribution of examples and is diagonal, i.e. limN-+ooGjk(t,s) = G(t,s)6jk , where G(t, s) = G(O)(t - s) +a ! ! dt 1 dt2G(O)(t - td(g~(tdDJl(tl' t2))G(t2' s). (12) 289 Statistical Dynamics ofBatch Learning D ~ (t, s) is the example Green's function given by D~(t,s) = c5(t - s) + Jdt'G(t,t')g~(t')D~(t',s). (13) This allows us to express the generic activations of the examples in terms of their cavity counterparts. Multiplying both sides of (10) and summing over j, we get xo(t) - ho(t) = J dsG(t, s)g~(s). (14) This equation is interpreted as follows. At time t, the generic activation xo{t) deviates from its cavity counterpart because its gradient term g&(s) was present in the batch learning step at previous times s. This gradient term propagates its influence from time s to t via the Green's function G(t, s). Statistically, this equation enables us to express the activation distribution in terms of the cavity activation distribution, thereby getting a macroscopic description of the dynamics. To solve for the Green's functions and the activation distributions, we further need the fluctuation-response relation derived by linear response theory, C(t, s) =a J dt'G(O) (t - t')(g~(t')x~(s? + 2T J dt'G(O)(t - t')G(s, t'). (15) Finally, the teacher-student correlation is given by R(t) = a 4 J dt'G(O)(t - t')(g~(t')y~}. (16) A Solvable Case The cavity method can be applied to the dynamics of learning with an arbitrary cost function. When it is applied to the Hebb rule, it yields results identical to the exact results in [8]. Here we present the results for the Adaline rule to illustrate features of learning dynamics derivable from the study. This is a common learning rule and bears resemblance with the more common back-propagation rule. Theoretically, its dynamics is particularly convenient for analysis since g" (x) = -1, rendering the weight Green's function time translation invariant, Le. G(t, s) = G(t - s). In this case, the dynamics can be solved by Laplace transform. To monitor the progress of learning, we are interested in three performance measures: (a) Training error ft, which is the probability of error for the training examples. It is given by ft = (9 (-xsgnY?:q, , where the average is taken over the joint distribution p(x, y) of the training set. (b) Test error ftest, which is the probof the training examples are corrupted by an ability of error when the inputs additive Gaussian noise of variance ~ 2 . This is a relevant performance measure when the percept ron is applied to process data which are the corrupted versions of the training data. It is given by ftest = (H(xsgny/~JC(t,t?):r;y. When ~2 = 0, the test error reduces to the training error. (c) Generalization error fg, which is the probability of error for an arbitrary input ~j when the teacher and student outputs are compared. It is given by fg = arccos[R(t)/ JC(t, t?)/7r. e; Figure l(a) shows the evolution of the generalization error at T = O. When the weight decay strength varies, the steady-state generalization error is minimized at the optimum 7r (17) Aopt = '2 - 1, S. Li and K. Y. M Wong 290 which is independent of Q. It is interesting to note that in the cases of the linear percept ron , the optimal weight decay strength is also independent of Q and only determined by the output noise and unlearn ability of the examples [5, 7]. Similarly, here the student is only provided the coarse-grained version of the teacher's activation in the form of binary bits. For A < Aopt, the generalization error is a non-monotonic function in learning time. Hence the dynamics is plagued by overtraining, and it is desirable to introduce early stopping to improve the perceptron performance. Similar behavior is observed in linear perceptrons [5, 6, 7]. To verify the theoretical predictions, simulations were done with N = 500 and using 50 samples for averaging. As shown in Fig. l(a), the agreement is excellent. Figure 1 (b) compares the generalization errors at the steady-state and the early stopping point. It shows that early stopping improves the performance for A < Aopt, which becomes near-optimal when compared with the best result at A = Aopt. Hence early stopping can speed up the learning process without significant sacrifice in the generalization ability. However, it cannot outperform the optimal result at steadystate. This agrees with a recent empirical observation that a careful control of the weight decay may be better than early stopping in optimizing generalization [12]. 0.40 0.38 ,0.3s l 0.36 w~ 1..=10 e~ 0.32 L~:::::::::=:::;::;;;;;;;mEE:=::: 1..=0.1 1..=1.. .... a=O.5 c .Q ? ~ a=O.5 t ~ 0.32 oc " ~ 0.30 iii N ~ I ~~0.34~ 0.28 i .t:! l.-..a~=.....I~.2.....-.............................~.....-.................._ 1..=10 ~ 0.28 I..=Ql ~ 0.26 Q) c 1..=\", 0.24 0.24 00 a=1.2 [--r-hh-e; 0.20 o -----c... 2 - '--'--- ~---"--'--~-'--- 4 6 timet 8 10 12 0.22 0.0 ---'- - _.- 0.5 1.0 1.5 weight decay ).. Figure 1: (a) The evolution of the generalization error at T = 0 for Q = 0.5,1.2 and different weight decay strengths A. Theory: solid line, simulation: symbols. (b) Comparing the generalization error at the steady state (00) and at the early stopping point (t es ) for Q = 0.5,1.2 and T = O. In the search for optimal learning algorithms, an important consideration is the environment in which the performance is tested. Besides the generalization performance, there are applications in which the test examples have inputs correlated with the training examples. Hence we are interested in the evolution of the test error for a given additive Gaussian noise 6, in the inputs. Figure 2(a) shows, again, that there is an optimal weight decay parameter Aopt which minimizes the test error. Furthermore, when the weight decay is weak, early stopping is desirable. Figure 2(b) shows the value of the optimal weight decay as a function of the input noise variance 6, 2 . To the lowest order approximation, Aopt ex: 6, 2 for sufficiently large 6, 2 . The dependence of Aopt on input noise is rather general since it also holds in the case of random examples [13]. In the limit of small 6,2, Aopt vanishes as 6,2 for Q < 1, whereas Aopt approaches a nonzero constant for Q > 1. Hence for 291 Statistical Dynamics of Batch Learning a < 1, weight decay is not necessary when the training error is optimized, but when the percept ron is applied to process increasingly noisy data, weight decay becomes more and more important in performance enhancement. Figure 2(b) also shows the phase line Aot(~2) below which overtraining occurs. Again, to the lowest order approximation, Aot ex ~2 for sufficiently large ~2 . However, unlike the case of generalization error, the line for the onset of overtraining does not coincide exactly with the line of optimal weight decay. In particular, for an intermediate range of input noise, the optimal line lies in the region of overtraining, so that the optimal performance can only be attained by tuning both the weight decay strength and learning time. However, at least in the present case, computational results show that the improvement is marginal. 0.30 , 20 I 028 i ~A=Ol c . l a=1.2 ... 0.26 Q) 0.24 g ~ , :--s r A=lO A=3.6 0 I t t 15 .-< 10 l o.oo ~ 0.22 ~ 5 -0.05 ' - - - - - - - - o 0.18 0---'--~2~~~ 3----"4~--'5----'6 Time Figure 2: (a) The evolution of the test error for ~ 2 = 3, T = 0 and different weight decay strengths A (Aopt ::::: 1.5,3.6 for a = 0.5, 1.2 respectively). (b) The lines of the optimal weight decay and the onset of overtraining for a = 5. Inset: The same data with Aot - Aopt (magnified) versus ~2. 5 Conclusion Based on the cavity method, we have introduced a new framework for modeling the dynamics of learning, which is applicable to any learning cost function, making it a versatile theory. It takes into full account the temporal correlations generated by the use of a restricted set of examples, which is more realistic in many situations than theories of on-line learning with an infinite training set. While the Adaline rule is solvable by the cavity method, it is still a relatively simple model approachable by more direct methods. Hence the justification of the method as a general framework for learning dynamics hinges on its applicability to less trivial cases. In general, g~(t') in (13) is not a constant and DJl(t, s) has to be expanded as a series. The dynamical equations can then be considered as the starting point of a perturbation theory, and results in various limits can be derived, e.g. the limits of small a, large a, large A, or the asymptotic limit. Another area for the useful application of the cavity method is the case of batch learning with very large learning steps. Since it has been shown recently that such learning converges in a few steps [6], the dynamical equations remain simple enough for a meaningful study. Preliminary results along this direction are promising and will be reported elsewhere. S. Li and K. Y. M Wong 292 An alternative general theory for learning dynamics, the dynamical replica theory, has recently been developed [8]. It yields exact results for Hebbian learning, and approximate results for more non-trivial cases. Based on certain self-averaging assumptions, the theory is able to approximate the dynamics by the evolution of single-time functions , at the expense of having to solve a set of saddle point equations in the replica formalism at every learning instant. On the other hand, our theory retains the functions G(t,s) and C(t, s) with double arguments, but develops naturally from the stochastic nature of the cavity activations. Contrary to a suggestion [14], the cavity method can also be applied to the on-line learning with restricted sets of examples. It is hoped that by adhering to an exact formalism, the cavity method can provide more fundamental insights when the studies are extended to more sophisticated multilayer networks of practical importance. The method enables us to study the effects of weight decay and early stopping. It shows that the optimal strength of weight decay is determined by the imprecision in the examples, or the level of input noise in anticipated applications. For weaker weight decay, the generalization performance can be made near-optimal by early stopping. Furthermore, depending on the performance measure, optimality may only be attained by a combination of weight decay and early stopping. Though the performance improvement is marginal in the present case, the question remains open in the more general context. We consider the present work as the beginning of an in-depth study of learning dynamics. Many interesting and challenging issues remain to be explored. Acknowledgments We thank A. C. C. Coolen and D. Saad for fruitful discussions during NIPS. This work was supported by the grant HKUST6130/97P from the Research Grant Council of Hong Kong. References [1] D. Saad and S. Solla, Phys. Rev. Lett. 74,4337 (1995) . [2] [3] [4] [5] [6] [7] [8] D. Saad and M. Rattray, Phys. Rev. Lett. 79, 2578 (1997). J. Hertz, A. Krogh and G. I. Thorbergssen, J. Phys. A 22, 2133 (1989) . M. Opper, Europhys. Lett. 8, 389 (1989) . A. Krogh and J. A. Hertz, J. Phys. A 25, 1135 (1992) . S. Bos and M. Opper, J. Phys. A 31, 4835 (1998). S. Bos, Phys. Rev. E 58, 833 (1998) . A. C. C. Coolen and D. Saad, in On-line Learning in Neural Networks, ed. D. Saad (Cambridge University Press, Cambridge, 1998). [9] M. Mezard, G. Parisi and M. Virasoro, Spin Glass Theory and Beyond (World Scientific, Singapore) (1987). [10] K. Y. M. Wong, Europhys. Lett. 30, 245 (1995). [11] K. Y. M. Wong, Advances in Neural Information Processing Systems 9, 302 (1997) . [12] L. K. Hansen, J. Larsen and T. Fog, IEEE Int. Conf. on Acoustics, Speech, and Signal Processing 4,3205 (1997). [13] Y. W. Tong, K. Y. M. Wong and S. Li, to appear in Proc. of IJCNN'99 (1999) . [14] A. C. C. Cool en and D. Saad, Preprint KCL-MTH-99-33 (1999).
1697 |@word kong:3 version:3 briefly:1 open:1 mee:1 simulation:2 covariance:1 thereby:1 versatile:1 solid:1 series:1 denoting:1 reaction:2 comparing:1 activation:19 perturbative:1 ust:1 additive:2 realistic:1 subsequent:2 enables:2 beginning:1 coarse:1 node:3 ron:5 along:1 direct:1 introduce:4 theoretically:1 sacrifice:1 behavior:1 multi:1 gjk:2 ol:1 td:1 sager:1 provided:1 becomes:2 notation:1 lowest:2 what:1 interpreted:1 minimizes:2 developed:1 magnified:1 onsager:1 temporal:5 every:1 exactly:1 control:1 grant:2 appear:1 local:1 limit:4 fluctuation:1 challenging:1 range:1 statistically:1 practical:1 acknowledgment:1 area:1 empirical:1 convenient:1 djl:1 get:1 convenience:1 cannot:1 context:1 influence:1 wong:7 fruitful:1 map:1 starting:1 independently:1 adhering:1 rule:8 insight:1 justification:1 laplace:1 suppose:2 xjl:3 exact:5 us:1 agreement:1 satisfying:1 jk:4 particularly:1 observed:1 ft:2 preprint:1 solved:3 region:1 solla:1 environment:1 vanishes:1 dynamic:25 trained:1 joint:1 various:1 distinct:1 kcl:1 describe:1 europhys:2 widely:2 solve:2 ability:3 transform:1 noisy:1 superscript:1 timet:1 parisi:1 relevant:1 adaline:4 description:2 yjl:1 getting:1 enhancement:1 optimum:1 double:1 converges:1 illustrate:1 oo:1 depending:1 progress:3 krogh:2 cool:1 direction:1 stochastic:2 f1:1 generalization:12 preliminary:1 adjusted:1 hold:1 sufficiently:3 considered:3 exp:1 plagued:1 bj:1 driving:2 early:13 omitted:1 proc:1 applicable:3 label:1 coolen:2 hansen:1 council:1 agrees:1 kowloon:1 gaussian:4 rather:1 avoid:1 derived:2 focus:1 yo:3 improvement:2 hk:1 bos:2 glass:1 stopping:11 mth:1 relation:1 interested:2 issue:2 among:1 denoted:2 arccos:1 marginal:2 having:1 identical:1 anticipated:1 minimized:2 t2:2 develops:1 jdt:1 few:1 approachable:1 phase:1 n1:1 attempt:1 interest:1 introduces:1 yielding:1 fog:1 tj:3 partial:1 necessary:1 theoretical:1 virasoro:1 formalism:2 modeling:2 retains:1 cost:6 applicability:1 reported:1 teacher:7 corrupted:2 varies:1 fundamental:1 physic:1 yl:1 michael:1 connecting:1 again:2 central:1 conf:1 derivative:1 li:5 account:4 student:5 int:1 jc:2 caused:1 depends:1 onset:2 analyze:1 complicated:1 spin:1 variance:3 percept:5 yield:3 weak:1 multiplying:1 published:1 overtraining:5 phys:6 ed:1 larsen:1 naturally:1 adjusting:1 improves:1 sophisticated:1 back:1 appears:1 attained:2 dt:5 response:2 formulation:1 done:1 though:1 furthermore:4 correlation:9 hand:2 nonlinear:1 propagation:1 mode:1 resemblance:1 scientific:1 effect:7 verify:1 counterpart:2 evolution:6 hence:8 assigned:1 former:1 imprecision:1 iteratively:1 nonzero:1 x5:1 during:4 self:3 steady:3 djj:1 oc:1 hong:3 temperature:1 steadystate:1 instantaneous:1 consideration:1 recently:4 common:2 unlearn:1 functional:1 jl:1 significant:1 cambridge:2 tuning:1 consistency:1 similarly:1 access:1 longer:1 recent:4 optimizing:1 certain:2 binary:2 minimum:1 fortunately:1 signal:1 full:1 desirable:2 reduces:1 hebbian:3 prediction:1 multilayer:1 represent:1 achieved:1 whereas:3 background:1 limn:1 macroscopic:4 extra:1 saad:6 unlike:1 contrary:1 incorporates:1 near:2 ideal:1 intermediate:1 iii:1 enough:1 rendering:1 affect:1 variate:1 speech:1 jj:10 ignored:1 useful:1 clear:1 outperform:1 aot:3 singapore:1 gil:1 trapped:1 rattray:1 express:2 monitor:1 replica:3 aopt:11 almost:1 bit:1 layer:1 strength:7 ijcnn:1 x2:1 speed:1 argument:2 optimality:1 expanded:1 relatively:1 department:1 combination:1 hertz:2 describes:3 remain:2 increasingly:1 rev:3 making:1 happens:1 restricted:6 invariant:1 xo:6 jlk:1 taken:2 equation:7 remains:1 turn:1 hh:1 needed:1 adopted:1 generic:3 appropriate:1 batch:8 alternative:1 ho:7 customary:1 original:4 ensure:1 hinge:1 recycling:1 instant:1 calculating:1 giving:1 added:4 quantity:1 occurs:1 question:1 costly:1 primary:2 dependence:1 diagonal:1 gradient:6 thank:1 trivial:2 water:1 besides:1 illustration:1 dsg:1 ql:1 expense:1 rise:1 observation:1 descent:3 situation:1 extended:2 precise:1 perturbation:2 arbitrary:3 introduced:2 namely:1 specified:1 optimized:1 acoustic:1 nip:1 address:1 beyond:2 able:1 dynamical:8 below:2 green:9 including:1 unrealistic:1 regularized:1 solvable:2 improve:1 technology:1 bare:1 deviate:1 asymptotic:1 fully:2 bear:1 interesting:2 suggestion:1 versus:1 propagates:2 dsgjk:1 translation:1 lo:1 elsewhere:1 supported:1 side:2 weaker:1 perceptron:7 taking:1 fg:3 depth:1 lett:4 opper:2 world:1 made:3 collection:1 c5:1 coincide:1 far:1 approximate:2 derivable:1 cavity:16 summing:1 assumed:1 search:1 bay:1 promising:1 learn:1 nature:1 correlated:1 excellent:1 complex:2 noise:9 fig:1 referred:1 tl:1 en:1 hebb:1 tong:1 mezard:1 lie:1 grained:1 diagrammatic:1 phkywong:1 inset:1 symbol:1 explored:1 decay:21 concern:1 adding:1 importance:1 hoped:1 simply:1 saddle:1 adjustment:1 monotonic:1 corresponds:1 careful:1 change:1 infinite:3 specifically:1 determined:2 averaging:3 called:1 secondary:2 e:1 meaningful:1 perceptrons:3 latter:1 tested:1 ex:2
762
1,698
Bayesian Reconstruction of 3D Human Motion from Single-Camera Video Nicholas R. Howe Department of Computer Science Cornell University Ithaca, NY 14850 nihowe@cs.comell.edu Michael E. Leventon Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 leventon@ai.mit.edu William T. Freeman MERL - a Mitsubishi Electric Research Lab 201 Broadway Cambridge, MA 02139 freeman@merL.com Abstract The three-dimensional motion of humans is underdetermined when the observation is limited to a single camera, due to the inherent 3D ambiguity of 2D video. We present a system that reconstructs the 3D motion of human subjects from single-camera video, relying on prior knowledge about human motion, learned from training data, to resolve those ambiguities. After initialization in 2D, the tracking and 3D reconstruction is automatic; we show results for several video sequences. The results show the power of treating 3D body tracking as an inference problem. 1 Introduction We seek to capture the 3D motions of humans from video sequences. The potential applications are broad, including industrial computer graphics, virtual reality, and improved human-computer interaction. Recent research attention has focused on unencumbered tracking techniques that don't require attaching markers to the subject's body [4, 5], see [12] for a survey. Typically, these methods require simultaneous views from multiple cameras. Motion capture from a single camera is important for several reasons. First, though underdetermined, it is a problem people can solve easily, as anyone viewing a dancer in a movie can confirm. Single camera shots are the most convenient to obtain, and, of course, apply to the world's film and video archives. It is an appealing computer vision problem that emphasizes inference as much as measurement. This problem has received less attention than motion capture from multiple cameras. Goncalves et.al. rely on perspective effects to track only a single arm, and thus need not deal with complicated models, shadows, or self-occlusion [7]. Bregler & Malik develop a body tracking system that may apply to a single camera, but performance in that domain is Bayesian Reconstruction of 3D Human Motion from Single-Camera Video 821 not clear; most of the examples use multiple cameras [4]. Wachter & Nagel use an iterated extended Kalman filter, although their body model is limited in degrees of freedom [12l Brand [3] uses an learning-based approach, although with representational expressiveness restricted by the number of HMM states. An earlier version of the work reported here [10] required manual intervention for the 2D tracking. This paper presents our system for single-camera motion capture, a learning-based approach, relying on prior information learned from a labeled training set. The system tracks joints and body parts as they move in the 2D video, then combines the tracking information with the prior model of human motion to form a best estimate of the body's motion in 3D. Our reconstruction method can work with incomplete information, because the prior model allows spurious and distracting information to be discarded . The 3D estimate provides feedback to influence the 2D tracking process to favor more likely poses. The 2D tracking and 3D reconstruction modules are discussed in Sections 3 and 4, respectively. Section 4 describes the system operation and presents performance results. Finally, Section 5 concludes with possible improvements. 2 2D Tracking The 2D tracker processes a video stream to determine the motion of body parts in the image plane over time. The tracking algorithm used is based on one presented by Ju et. al. [9], and performs a task similar to one described by Morris & Rehg [11]. Fourteen body parts are modeled as planar patches, whose positions are controlled by 34 parameters. Tracking consists of optimizing the parameter values in each frame so as to minimize the mismatch between the image data and a projection of the body part maps. The 2D parameter values for the first frame must be initialized by hand, by overlaying a model onto the 2D image of the first frame. We extend Ju et. al.'s tracking algorithm in several ways. We track the entire body, and build a model of each body part that is a weighted average of several preceding frames, not just the most recent one. This helps eliminate tracking errors due to momentary glitches that last for a frame or two. We account for self-occlusions through the use of support maps [4, 1]. It is essential to address this problem, as limbs and other body parts will often partly or wholly obscure one another. For the single-camera case, there are no alternate views to be relied upon when a body part cannot be seen. The 2D tracker returns the coordinates of each limb in each successive frame. These in tum yield the positions of joints and other control points needed to perform 3D reconstruction. 3 3D Reconstruction 3D reconstruction from 2D tracking data is underdetermined. At each frame, the algorithm receives the positions in two dimensions of 20 tracked body points, and must to infer the correct depth of each point. We rely on a training set of 3D human motions to determine which reconstructions are plausible. Most candidate projections are unnatural motions, if not anatomically impossible, and can be eliminated on this basis. We adopt a Bayesian framework, and use the training data to compute prior probabilities of different 3D motions. We model plausible motions as a mixture of Gaussian probabilities in a high-dimensional space. Motion capture data gathered in a professional studio provide the training data: frame-by-frame 3D coordinates for 20 tracked body points at 20-30 frames per second. We want to model the probabilities of human motions of some short duration, long enough be N. R. Howe, M. E. Leventon and W T. Freeman 822 informative, but short enough to characterize probabilistically from our training data. We assembled the data into short motion elements we caJled snippets of 11 successive frames, about a third of a second. We represent each snippet from the training data as a large column vector of the 3D positions of each tracked body point in each frame of the snippet. We then use those data to build a mixture-of-Gaussians probability density model [2]. For computational efficiency, we used a clustering approach to approximate the fitting of an EM algorithm. We use k-means clustering to divide the snippets into m groups, each of which will be modeled by a Gaussian probability cloud. For each cluster, the matrix M j is formed, where the columns of M j are the nj individual motion snippets after subtracting the mean J.l j. The singular value decomposition (SVD) gives M j = U j Sj where Sj contains the singular values along the diagonal, and Uj contains the basis vectors. (We truncate the SVD to include only the 50 largest singular values.) The cluster can be modeled by a multidimensional Gaussian with covariance Aj = UjSJUJ. The prior probability of a snippet x over all the models is a sum of the Gaussian probabilities weighted by the probability of each model. VI, ;j m P(x) = Lk7fje- !(x-llj)TA- 1 (X-llj) (1) j=1 Here k is a normalization constant, and 7fj is the a priori probability of model j, computed as the fraction of snippets in the knowledge base that were originally placed in cluster j . Given this approximately derived mixture-of-factors model [6], we can compute the prior probability of any snippet. To estimate the data term (likelihood) in Bayes' law, we assume that the 2D observations include some Gaussian noise with variance (T. Combined with the prior, the expression for the probability of a given snippet x given an observation ybecomes p(x,e,s,vly) = k' (e-IIY-R6 ,?.v(XlII 2/(2tr2)) (f k7fj e-!(X-llj)T A _l(X-llj)) (2) J=l In this equation, Rn ,s,ii(X) is a rendering function which maps a 3D snippet x into the image coordinate system, performing scaling s, rotation about the vertical axis and image-plane translation v. We use the EM algorithm to find the probabilities of each Gaussian in the mixture and the corresponding snippet x that maximizes the probability given the observations [6]. This allows the conversion of eleven frames of 2D tracking measurements into the most probable corresponding 3D snippet. In cases where the 2D tracking is poor, the reconstruction may be improved by matching only the more reliable points in the likelihood term of Equation 2. This adds a second noise process to explain the outlier data points in the likelihood term. e, To perform the full 3D reconstruction, the system first divides the 2D tracking data into snippets, which provides the y values of Eq. 2, then finds the best (MAP) 3D snippet for each of the 2D observations. The 3D snippets are stitched together, using a weighted interpolation for frames where two snippets overlap. The result is a Bayesian estimate of the subject's motion in three dimensions. 4 Performance The system as a whole will track and successfully 3D reconstruct simple, short video clips with no human intervention, apart from 2D pose initialization. It is not currently reliable enough to track difficult footage for significant lengths of time. However, analysis of short clips demonstrates that the system can successfully reconstruct 3D motion from ambiguous Bayesian Reconstruction of 3D Human Motion from Single-Camera Video 823 2D video. We evaluate the two stages of the algorithm independently at first, and then consider their operation as a system. 4.1 Performance of the 3D reconstruction The 3D reconstruction stage is the heart of the system. To our knowledge, no similar 2D to 3D reconstruction technique relying on prior infonnation has been published. ([3], developed simultaneously, also uses an inference-based approach). Our tests show that the module can restore deleted depth infonnation that looks realistic and is close to the ground truth, at least when the knowledge base contains some examples of similar motions. This makes the 3D reconstruction stage itself an important result, which can easily be applied in conjunction with other tracking technologies. To test the reconstruction with known ground truth, we held back some of the training data for testing. We artificially provided perfect 2D marker position data, yin Eq. 2, and tested the 3D reconstruction stage in isolation. After removing depth information from the test sequence, the sequence is reconstructed as if it had come from the 2D tracker. Sequences produced in this manner look very much like the original. They show some rigid motion error along the line of sight. An analysis of the uncertainty in the posterior probability predicts high uncertainty for the body motion mode of rigid motion parallel to the orthographic projection [10]. This slipping can be corrected by enforcing ground-contact constraints. Figure 1 shows a reconstructed running sequence corrected for rigid motion error and superimposed on the original. The missing depth information is reconstructed well, although it sometimes lags or anticipates the true motion slightly. Quantitatively, this error is a relatively small effect. After subtracting rigid motion error, the mean residual 3D errors in limb position are the same order of magnitude as the small frame-to frame changes in those positions. - ' ~ . -, " _. ~ .~ ..? Figure 1: Original and reconstructed running sequences superimposed (frames 1, 7, 14, and 21). 4.2 Performance of the 2D tracker The 2D tracker performs well under constant illumination, providing quite accurate results from frame to frame. The main problem it faces is the slow accumulation of error. On longer sequences, the errors can build up to the point where the module is no longer tracking the body parts it was intended to track. The problem is worsened by low contrast, occlusion and lighting changes. More careful body modeling [5], lighting models, and modeling of the background may address these issues. The sequences we used for testing were several seconds long and had fairly good contrast. Although adequate to demonstrate the operation of our system, the 2D tracker contains the most open research issues. 4.3 Overall system performance Three example reconstructions are given, showing a range of different tracking situations. The first is a reconstruction of a stationary figure waving one arm, with most of the motion 824 N. R. Howe. M E. Leventon and W. T. Freeman in the image plane. The second shows a figure bringing both arms together towards the camera, resulting in a significant amount of foreshortening. The third is a reconstruction of a figure walking sideways, and includes significant self-occlusion Figure 2: First clip and its reconstruction (frames 1, 2S, SO, and 7S). The first video is the easiest to track because there is little or no occlusion and change in lighting. The reconstruction is good, capturing the stance and motion of the arm. There is some rigid motion error, which is corrected through ground friction constraints. The knees are slightly bent; this may be because the subject in the video has different body proportions than those represented in the training database. Figure 3: Second clip and its reconstruction (frames 1, 2S, SO, and 7S). The second video shows a figure bringing its arms together towards the camera. The only indication of this is in the foreshortening of the limbs, yet the 3D reconstruction correctly captures this in the right arm. (Lighting changes and contrast problems cause the 2D tracker to lose the left arm partway through, confusing the reconstruction of that limb, but the right arm is tracked accurately throughout.) The third video shows a figure walking to the right in the image plane. This clip is the hardest for the 2D tracker, due to repeated and prolonged occlusion of some body parts. The tracker loses the left arm after IS frames due to severe occlusion, yet the remaining tracking information is still sufficient to perform an adequate reconstruction. At about frame 4S, the left leg has crossed behind the right several times and is lost, at which point the reconstruction quality begins to degrade. The key to a more reliable reconstruction on this sequence is better tracking. Bayesian Reconstruction of 3D Human Motion from Single-Camera Video .. .. ,t 825 :'."...., ~ 1. Figure 4: Third clip and its reconstruction (frames 6, 16, 26, and 36). 5 Conclusion We have demonstrated a system that tracks human figures in short video sequences and reconstructs their motion in three dimensions. The tracking is unassisted, although 2D pose initialization is required. The system uses prior information learned from training data to resolve the inherent ambiguity in going from two to three dimensions, an essential step when working with a single-camera video source. To achieve this end, the system relies on prior knowledge, extracted from examples of human motion. Such a learning-based approach could be combined with more sophisticated measurement-based approaches to the tracking problem [12, 8, 4]. References [1] J. R. Bergen, P. Anandan, K. J. Hanna, and R. Hingorani. Hierarchical model-based motion estimation. In European Conference on Computer Vision, pages 237-252, 1992. [2] C. M. Bishop. Neural networks for pattern recognition. Oxford, 1995. [3] M. Brand. Shadow puppetry. In Proc. 7th IntI. Con! on Computer Vision, pages 1237-1244. IEEE, 1999. [4] c. Bregler and 1. Malik. Tracking people with twists and exponential maps. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Santa Barbera, 1998. [5] D . M. Gavrila and L. S. Davis. 3d model-based tracking of humans in action: A multi-view approach. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, 1996. [6] Z. Ghahramani and G. E. Hinton. The EM algorithm for mixtures offactor analyzers. Technical report, Department of Computer Science, University of Toronto, May 21 1996. (revised Feb. 27, 1997). [7] L. Goncalves, E. Di Bernardo, E. Ursella, and P. Perona. Monocular tracking of the human arm in 3D. In Proceedings of the Third International Conference on Computer Vision, 1995. [8] M. Isard and A. Blake. Condensation - conditional density propagation for visual tracking. International Journal of Computer Vision, 29( 1):5-28, 1998. [9] S. X. Ju, M. J. Black, and Y. Yacoob. Cardboard people: A parameterized model of articulated image motion. In 2nd International Conference on Automatic Face and Gesture Recognition, 1996. 826 N. R. Howe, M. E. Leventon and W T. Freeman [10] M. E. Leventon and W. T. Freeman. Bayesian estimation of 3-d human motion from an image sequence. Technical Report TR98-06, Mitsubishi Electric Research Lab, 1998. [11] D. D. Morris and 1. Rehg. Singularity analysis for articulated object tracking. In IEEE Computer Societ), Conference on Computer Yz'sion and Pattern Recognition, Santa Barbera, 1998. [12] S. Wachter and H.-H. Nagel. Tracking of persons in monocular image sequences. In Nonrigid and ArticuLated Motion Workshop, 1997.
1698 |@word version:1 proportion:1 nd:1 open:1 seek:1 mitsubishi:2 decomposition:1 covariance:1 shot:1 contains:4 com:1 comell:1 yet:2 must:2 realistic:1 informative:1 eleven:1 treating:1 stationary:1 intelligence:1 isard:1 plane:4 short:6 provides:2 toronto:1 successive:2 along:2 consists:1 combine:1 fitting:1 manner:1 multi:1 footage:1 freeman:6 relying:3 resolve:2 little:1 prolonged:1 provided:1 begin:1 maximizes:1 easiest:1 developed:1 nj:1 multidimensional:1 bernardo:1 demonstrates:1 control:1 intervention:2 overlaying:1 puppetry:1 oxford:1 interpolation:1 approximately:1 black:1 initialization:3 limited:2 range:1 camera:17 testing:2 lost:1 orthographic:1 wholly:1 convenient:1 projection:3 matching:1 onto:1 cannot:1 close:1 influence:1 impossible:1 accumulation:1 map:5 demonstrated:1 missing:1 attention:2 duration:1 independently:1 focused:1 survey:1 knee:1 rehg:2 coordinate:3 us:3 element:1 recognition:5 walking:2 predicts:1 labeled:1 database:1 cloud:1 module:3 capture:6 upon:1 efficiency:1 basis:2 easily:2 joint:2 worsened:1 represented:1 articulated:3 artificial:1 whose:1 lag:1 film:1 solve:1 plausible:2 quite:1 reconstruct:2 favor:1 itself:1 sequence:13 indication:1 reconstruction:31 subtracting:2 interaction:1 achieve:1 representational:1 cluster:3 perfect:1 object:1 help:1 develop:1 pose:3 received:1 eq:2 c:1 shadow:2 come:1 correct:1 filter:1 human:18 viewing:1 virtual:1 require:2 probable:1 underdetermined:3 singularity:1 bregler:2 tracker:9 ground:4 blake:1 adopt:1 estimation:2 proc:1 lose:1 currently:1 infonnation:2 largest:1 successfully:2 sideways:1 weighted:3 mit:1 gaussian:6 sight:1 cornell:1 sion:1 yacoob:1 probabilistically:1 conjunction:1 derived:1 improvement:1 likelihood:3 superimposed:2 industrial:1 contrast:3 inference:3 bergen:1 rigid:5 typically:1 entire:1 eliminate:1 spurious:1 perona:1 going:1 issue:2 overall:1 priori:1 fairly:1 eliminated:1 broad:1 look:2 hardest:1 report:2 quantitatively:1 inherent:2 foreshortening:2 simultaneously:1 individual:1 intended:1 occlusion:7 william:1 freedom:1 severe:1 mixture:5 behind:1 stitched:1 held:1 iiy:1 accurate:1 incomplete:1 divide:2 cardboard:1 initialized:1 merl:2 column:2 earlier:1 modeling:2 leventon:6 graphic:1 characterize:1 reported:1 anticipates:1 combined:2 ju:3 density:2 international:3 person:1 michael:1 tr2:1 together:3 ambiguity:3 reconstructs:2 return:1 account:1 potential:1 includes:1 vi:1 stream:1 crossed:1 view:3 lab:3 relied:1 bayes:1 complicated:1 parallel:1 waving:1 minimize:1 formed:1 vly:1 variance:1 yield:1 gathered:1 bayesian:7 iterated:1 accurately:1 emphasizes:1 produced:1 lighting:4 published:1 simultaneous:1 explain:1 manual:1 dancer:1 di:1 con:1 massachusetts:1 knowledge:5 sophisticated:1 back:1 tum:1 ta:1 originally:1 planar:1 improved:2 though:1 just:1 stage:4 hand:1 receives:1 working:1 marker:2 propagation:1 mode:1 quality:1 aj:1 effect:2 true:1 stance:1 deal:1 self:3 ambiguous:1 davis:1 distracting:1 nonrigid:1 demonstrate:1 performs:2 motion:40 fj:1 image:10 rotation:1 tracked:4 twist:1 fourteen:1 extend:1 discussed:1 measurement:3 significant:3 cambridge:2 ai:1 automatic:2 analyzer:1 had:2 longer:2 base:2 add:1 feb:1 posterior:1 recent:2 perspective:1 optimizing:1 apart:1 slipping:1 seen:1 anandan:1 preceding:1 determine:2 ii:1 multiple:3 full:1 infer:1 technical:2 gesture:1 long:2 bent:1 controlled:1 vision:7 represent:1 normalization:1 sometimes:1 background:1 want:1 condensation:1 singular:3 howe:4 source:1 ithaca:1 bringing:2 archive:1 subject:4 nagel:2 gavrila:1 enough:3 rendering:1 isolation:1 expression:1 unnatural:1 cause:1 action:1 adequate:2 clear:1 santa:2 amount:1 morris:2 clip:6 track:8 per:1 correctly:1 group:1 key:1 deleted:1 fraction:1 sum:1 offactor:1 parameterized:1 uncertainty:2 throughout:1 patch:1 confusing:1 scaling:1 capturing:1 constraint:2 anyone:1 friction:1 performing:1 relatively:1 llj:4 department:2 alternate:1 truncate:1 poor:1 describes:1 slightly:2 em:3 appealing:1 leg:1 anatomically:1 restricted:1 outlier:1 unassisted:1 inti:1 heart:1 equation:2 monocular:2 needed:1 end:1 operation:3 gaussians:1 apply:2 limb:5 hierarchical:1 nicholas:1 hingorani:1 professional:1 original:3 clustering:2 include:2 running:2 remaining:1 wachter:2 build:3 uj:1 ghahramani:1 society:2 yz:1 contact:1 malik:2 move:1 diagonal:1 hmm:1 degrade:1 reason:1 enforcing:1 kalman:1 length:1 modeled:3 glitch:1 providing:1 difficult:1 broadway:1 perform:3 conversion:1 vertical:1 observation:5 revised:1 discarded:1 snippet:16 situation:1 extended:1 hinton:1 frame:24 rn:1 expressiveness:1 required:2 learned:3 assembled:1 address:2 pattern:4 mismatch:1 including:1 reliable:3 video:19 power:1 overlap:1 rely:2 restore:1 residual:1 arm:10 movie:1 technology:2 axis:1 concludes:1 prior:11 law:1 goncalves:2 degree:1 sufficient:1 obscure:1 translation:1 course:1 placed:1 last:1 institute:1 face:2 attaching:1 feedback:1 dimension:4 depth:4 world:1 san:1 sj:2 approximate:1 reconstructed:4 confirm:1 francisco:1 don:1 reality:1 hanna:1 european:1 artificially:1 electric:2 domain:1 main:1 whole:1 noise:2 repeated:1 body:21 ny:1 slow:1 position:7 momentary:1 exponential:1 candidate:1 r6:1 third:5 removing:1 bishop:1 showing:1 essential:2 workshop:1 magnitude:1 illumination:1 studio:1 yin:1 likely:1 visual:1 tracking:30 truth:2 loses:1 relies:1 extracted:1 ma:2 conditional:1 careful:1 towards:2 change:4 corrected:3 partly:1 svd:2 brand:2 support:1 people:3 evaluate:1 tested:1
763
1,699
Transductive Inference for Estimating Values of Functions Olivier Chapelle*, Vladimir Vapnik*,t, Jason Westontt.t,* * AT&T Research Laboratories, Red Bank, USA. t Royal Holloway, University of London, Egham, Surrey, UK. tt Barnhill BioInformatics.com, Savannah, Georgia, USA. { chapelle, vlad, weston} @research.att.com Abstract We introduce an algorithm for estimating the values of a function at a set of test points Xe+!, ... , xl+m given a set of training points (XI,YI), ... ,(xe,Ye) without estimating (as an intermediate step) the regression function . We demonstrate that this direct (transducti ve) way for estimating values of the regression (or classification in pattern recognition) can be more accurate than the traditionalone based on two steps, first estimating the function and then calculating the values of this function at the points of interest. 1 Introduction Following [6] we consider a general scheme of transductive inference. Suppose there exists a function y* = fo(x) from which we observe the measurements corrupted with noise ((Xl, YI)," . (xe, Ye)), Yi = Y; + ~i' (1) Find an algorithm A that using both the given set of training data (1) and the given set of test data (2) (Xl+!,' .. , XHm) selects from a set of functions {x t--+ f (x)} a function Y = f(x) = fA(xlxl,YI, ... ,Xl,Yl,XHI"",XHm) and minimizes at the points of interest the functional R(A) = E (~ (y; - fA(Xilxl,Yl, ... ,Xl,Ye,Xl+l, . .. ,Xl+m))2) (3) (4) i=l+l where expectation is taken over X and~. For the training data we are given the vector X and the value Y, for the test data we are only given x. Usually, the problem of estimating values of a function at points of interest is sol ved in two steps: first in a given set of functions f (x, a), a E A one estimates the regression, i.e the function which minimizes the functional R(a) = J ((y - f(x, a))2dF(x, Y), (5) 422 0. Chapelle, V. N. Vapnik and J. Weston (the inductive step) and then using the estimated function Y = f(x,al) we calculate the values at points of interest yi = (6) f(x;, ae) (the deductive step). Note, however, that the estimation of a function is equivalent to estimating its values in the continuum points of the domain of the function. Therefore, by solving the regression problem using a restricted amount of information, we are looking for a more general solution than is required. In [6] it is shown that using a direct estimation method one can obtain better bounds than through the two step procedure. In this article we develop the idea introduced in [5] for estimating the values of a function only at the given points. The material is organized as follows. In Section 1 we consider the classical (inductive) Ridge Regression procedure, and the leave-one--out technique which is used to measure the quality of its solutions. Section 2 introduces the transductive method of inference for estimation of the values of a function based on this leave-one- out technique. In Section 3 experiments which demonstrate the improvement given by transductive inference compared to inductive inference (in both regression and pattern recognition) are presented. Finally, Section 4 summarizes the results. 2 Ridge Regression and the Leave-One-Out procedure In order to describe our transductive method, let us first discuss the classical twostep (inductive plus deductive) procedure of Ridge Regression. Consider the set of functions linear in their parameters n f(x, a) = L aicPi(x). (7) i=1 To minimize the expected loss (5), where F(x, y) is unknown, we minimize the following empirical functional (the so-called Ridge Regression functional [1]) l Remp(a) = 1~ eL)Yi - f(Xi, a)) 2 + 1'110.11 2 (8) i=1 where l' is a fixed positive constant, called the regularization parameter. The minimum is given by the vector of coefficients ae = a(xl, Yl, ... , Xl, Yl) = (KT K + 1'1)-1 KTy (9) where y = (Y1, ... ,Ylf, and K is a matrix with elements: Kij=cPj(Xi), i=I, ... ,?, j=I, ... ,n. (10) (11) The problem is to choose the value l' which provides small expected loss for training on a sample Sl = {(Xl,Yl), .. . ,(Xl,Yl)}. For this purpose, we would like to choose l' such that f"f minimizing (8) also minimizes R= J (Y* - f"f(x* I Sl))2dF(x*, y*)dF(Se). (12) 423 Transductive Inference for Estimating Values of Functions Since F(x, y) is unknown one cannot estimate this minimum directly. To solve this problem we instead use the leave-one-out procedure, which is an almost unbiased estimator of (12). The leave-one-out error of an algorithm on the training sample Sf. is (13) The leave-one-out procedure consists of removing from the training data one element (say (Xi, Yi)), constructing the regression function only on the basis of the remaining training data and then testing the removed element. In this fashion one tests all f elements of the training data using f different decision rules. The minimum over, of (13) we consider as the minimum over, of (12) since the expectation of (13) coincides with (12) [2]. For Ridge Regression, one can derive a dosed form expression for the leave- one- out error. Denoting (14) the error incurred by the leave-one-out procedure is [6] -_1 T. loo(r) - f f. L (Y'_k TA- 1KTy)2 ~ ~=1 ~ 'Y (15) 1 _ kT A- 1k. ~ ~ 'Y where (16) k t = (i>I(xd??? ,l/>n(Xt)f? Let, = ,0 be the minimum of (15). Then the vector yO = K*(KT K +,0 I)-I KTy (17) where I/>(XHI) K*- ( . (18) 1/>1 (XHm) is the Ridge Regression estimate of the unknown values (Ye+l' ... ,Ye+m)' 3 Leave-One-Out Error for Transductive Inference In transductive inference, our goal is to find an algorithm A which minimizes the functional (4) using both the training data (1) and the test data (2). We suggest the following method: predict (Ye+l' ... 'Ye+m) by finding those values which minimize the leave-one-out error of Ridge Regression training on the joint set (Xl, yd,?? . ,(Xl, Yl), (Xl+l, ye+l),?? ., (XHm, Ye+m)' (19) This is achieved in the following way. Suppose we treat the unknown values (Ye+l" .. ,Ye+m) as variables and for some fixed value of these variables we minimize the following empirical functional Remp(aly;, .. ?, y~) = f: m ( ~(Yi f. ~=l f(xi,a))2 Hm + .L (y; - f(xi, a))2 ) +,llaI1 2 . ~=l+1 (20) This functional differs only in the second term from the functional (8) and corresponds to performing Ridge Regression with the extra pairs (21) 424 O. Chapel/e, V. N. Vapnik and J. Weston Suppose that vector Y" = (Yi, ... , y:n) is taken from some set Y" E Y such that the pairs (21) can be considered as a sample drawn from the same distribution as the pairs (Xl, yi), ... , (Xl, yi)? In this case the leave-one-out error of minimizing (20) over the set (19) approximates the functional (4). We can measure this leaveone-out error using the same technique as in Ridge Regression. Using the closed form (15) one obtains (Y:'~ _kTt~TA-I~-1~kTY) 2 L +m 1 - k A-y k 1 7loo(rly~, .. ?,y~) = -f-- l+m i=l where we denote x = (Xl, ... , Xl+ m), Y = (YI, ... , Yl, Yi+1" (22) i i .. , Yi+m)T, and (23) Kij=<pj(Xi), i=I, ... ,i+m, j=I, ... ,n. (24) kt = (<PI(Xt} ... ,<Pn(xt)f? (25) Now let us rewrite the expression (22) in an equivalent form to separate the terms with Y from the terms with x. Introducing (26) and the matrix M with elements l+m M tJ.. -_ "~ k=l CikC kj Cu 2 (27) we obtain the equivalent expression of (22) (28) In order for the Y" which minimize the leave-one-out procedure to be valid it is required that the pairs (21) are drawn from the same distribution as the pairs (Xl, yi), ... , (Xl, yi)? To satisfy this constraint we choose vectors Y" from the set Y = {Y" : IIY" - y011 -s: R} (29) where the vector yO is the solution obtained from classical Ridge Regression. To minimize (28) under constraint (29) we use the functional (30) where 'Y" is a constant depending on R. Now, to find the values at the given points of interest (2) all that remains is to find the minimum of (30) in Y". Note that the matrix M is obtained using only the vectors X. Therefore, to find the minimum of this functional we rewrite Equation (30) as where (32) Transductive Inference for Estimating Values of Functions 425 and Mo is a ex e matrix, Ml is a e x m matrix and M2 is a m x m matrix. Taking the derivative of (31) in y* we obtain the condition for the solution 2M1 Y + 2M2 Y* - 2),*Y o + 2),*Y* = 0 (33) which gives the predictions Y* = ()'* 1+ M 2)-1 (-MIY + ),*yO) . (34) In this algorithm (which we will call Transductive Regression) we have two parameters to control: )' and )'*. The choice of )' can be found using the leave-one-out estimator (15) for Ridge Regression. This leaves )'* as the only free parameter. 4 Experiments To compare our one- step transductive approach with the classical two- step approach we performed a series of experiments on regression problems. We also describe experiments applying our technique to the problem of pattern recognition . 4.1 Regression We conducted computer simulations for the regression problem using two datasets from the DELVE repository: boston and kin-32th. The boston dataset is a well- known problem where one is required to estimate house prices according to various statistics based on 13 locational, economic and structural features from data collected by the U.S Census Service in the Boston Massachusetts area. The kin-32th dataset is a realistic simulation of the forward dynamics of an 8 link all-revolute robot arm. The task is to predict the distance of the end-effector from a target, given 32 inputs which contain information on the joint positions, twist angles and so forth. Both problems are nonlinear and contain noisy data. Our objective is to compare our transductive inference method directly with the inductive method of Ridge Regression. To do this we chose the set of basis functions ?i(X) = exp (-llx - xiI 12/2(2), i = 1, ... , e, and found the values of )' and a for Ridge Regression which minimized the leave-one-out bound (15). We then used the same values of these parameters in our transductive approach , and using the basis functions ?i(X) = exp (-llx - XiW /2(72) , i = 1, . . . , e+ m, we then chose a fixed value of)'* . For the boston dataset we followed the same experimental setup as in [4], that is, we partitioned the training set of 506 observations randomly 100 times into a training set of 481 observations and a testing set of 25 observations. We chose the values of)' and a by taking the minimum average leave- one-out error over five more random splits of the data stepping over the parameter space. The minimum was found at )' = 0.005 and log a = 0.7. For our transductive method, we also chose the parameter),* = 10. In Figure la we plot mean squared error (MSE) on the test set averaged over the 100 runs against log a for Ridge Regression and Transductive Regression. Transductive Regression outperforms Ridge Regression, especially at the minimum. To observe the influence of the number of test points m on the generalization ability of our transductive method, we ran further experiments, setting )'* = e/2m for 0. Chapel/e, V. N. Vapnik and J. Weston 426 1- 9.2 ,, 9 Tranaductive Regression - _ . Ridge Regress.,n 1 8r---~----~----~---------, 7.95 " 8.8 7.9 8.6 g7.85 UJ 8.4 W ~ 7.8 8.2 7.75 8 7.7 7.8 7.6 0.4 0.6 0.8 7.65;L--~5:--=~10===-15==:"2=0-----:!25 1.2 Log sigma Test Set Size (a) (b) 015 0.14 ,, 1- ,, 1:' ~0. 12 Q) O.13,r---~--=--- ,, ~ 0.1 0.12 t: ~0. 115 ;l ,, ,, ---- -------- ------ ----- -- ---- 0.125 ,, ,, , I- 0.11 0.135 Transductive Regression - - - RIdge RegreSSion / 0.13 g , ,, ,, , I- " 0.11 0.105 0.1 , - --- , 0.095 0.09 1 1.5 2 2.5 Log sigma 3 1 0~ 0----1~50,------.,.200 ~---:2=-=-' 50 0.09''-------:'':50-----.,. Test Set Size (c) (d) Figure 1: A comparison of Transductive Regression to Ridge Regression on the boston dataset: (a) error rates for varying (J', (b) varying the test set size, m, and on the kin-32fh dataset: (c) error rates for varying (J', (d) varying the test set size. different values of m . In Figure 1b we plot m against MSE on the testing set, at log (J' = 0.7. The results indicate that increasing the test set size gives improved performance in Transductive Regression. For Ridge Regression, of course, the size of the testing set has no influence on the generalization ability. We then performed similar experiments on the kin-32fh dataset. This time, as we were interested in large testing sets giving improved performance for Transductive Regression we chose 100 splits where we took a subset of only 64 observations for training and 256 for testing. Again the leave-one-out estimator was used to find the values, = 0.1 and log (J' = 2 for Ridge Regression, and for Transductive Regression we also chose the parameter = 0.1. We plotted MSE on the testing set against = 50/m) log (J' (Figure 1c) and the size of the test set m for log (J' = 2.75 (also, (Figure 1d) for the two algorithms. For large test set sizes our method outperforms Ridge Regression. ,* 4.2 ,* Pattern Recognition This technique can also be applied for pattern recognition problems by solving them based on minimizing functional (8) with y = ?1. Such a technique is known as a Linear Discriminant (LD) technique. Transductive Inference for Estimating Values of Functions Postal Banana Diabetes Titanic Breast Cancer Heart Thyroid AB ABR - - 12.3 26.5 22.6 30.4 20.3 4.4 10.9 23.8 22.6 26.5 16.6 4.6 427 SVM 5.5 11.5 23.5 22.4 26.0 16.0 4.8 TLD 4.7 1l.4 23.3 22.4 25.7 15.7 4.0 Table 1: Comparison of percentage test error of AdaBoost (A B) , Regularized AdaBoost (ABR), Support Vector Machines (SVM) and Tmnsductive Linear Discrimination (TLD) on seven datasets. Table 1 describes results of experiments on classification in the following problems: 2 class digit recognition (0 - 4 versus 5 - 9) splitting the training set into 23 runs of 317 observations and considering a testing set of 2000 observations, and six problems from the UCI database. We followed the same experimental setup as in [3]: the performance of a classifier is measured by its average error over one hundred partitions of the datasets into training and testing sets. Free parameter(s) are chosen via validation on the first five training datasets. The performance of the transductive LD technique was compared to Support Vector Machines, AdaBoost and Regularized AdaBoost [3]. It is interesting to note that in spite of the fact that LD technique is one of the simplest pattern recognition techniques, transductive inference based upon this method performs well compared to state of the art methods of pattern recognition . 5 Summary In this article we performed transductive inference in the problem of estimating values of functions at the points of interest. We demonstrate that estimating the unknown values via a one- step (transductive) procedure can be more accurate than the traditional two-step (inductive plus deductive) one. References [1] A. Hoerl and R. W. Kennard . Ridge regression: Biased estimation for nonorthogonal problems. Technometrics, 12(1):55-67, 1970. [2] A. Luntz and V. Brailovsky. On the estimation of characters obtained in statistical procedure of recognition,. Technicheskaya Kibernetica, 1969. [In Russian]. [3] G. Witsch, T. Onoda, and K.-R. Muller. Soft margins for adaboost. Technical report, Royal Holloway, University of London, 1998. TR-98-2l. [4] C. Saunders, A. Gammermann, and V. Vovk. Ridge regression learning algorithm in dual variables. In Proccedings of the 15th International Conference on Machine Learning, pages 515-52l. Morgan Kaufmann, 1998. [5] V. Vapnik. Estimating of values of regression at the point of interest. In Method of Pattern Recognition. Sovetskoe Radio, 1977. [In Russian]. [6] V. Vapnik. Estimation of Dependences Based on Empirical Data. SpringerVerlag, New York, 1982.
1699 |@word cu:1 repository:1 simulation:2 tr:1 ld:3 series:1 att:1 denoting:1 outperforms:2 com:2 realistic:1 partition:1 plot:2 discrimination:1 leaf:1 provides:1 postal:1 five:2 direct:2 consists:1 introduce:1 expected:2 considering:1 increasing:1 estimating:14 minimizes:4 finding:1 xd:1 classifier:1 uk:1 control:1 positive:1 service:1 xhm:4 treat:1 yd:1 plus:2 chose:6 delve:1 g7:1 averaged:1 testing:9 differs:1 digit:1 procedure:10 area:1 empirical:3 spite:1 suggest:1 cannot:1 applying:1 influence:2 equivalent:3 splitting:1 chapel:2 estimator:3 rule:1 m2:2 target:1 suppose:3 olivier:1 diabetes:1 element:5 recognition:10 database:1 calculate:1 sol:1 removed:1 ran:1 dynamic:1 solving:2 rewrite:2 upon:1 basis:3 joint:2 various:1 describe:2 london:2 kty:4 saunders:1 solve:1 say:1 ability:2 statistic:1 transductive:27 noisy:1 took:1 uci:1 forth:1 leave:16 derive:1 develop:1 depending:1 measured:1 indicate:1 material:1 generalization:2 considered:1 exp:2 nonorthogonal:1 predict:2 mo:1 continuum:1 fh:2 purpose:1 estimation:6 hoerl:1 radio:1 deductive:3 pn:1 varying:4 yo:3 improvement:1 ved:1 inference:13 savannah:1 el:1 selects:1 interested:1 classification:2 dual:1 art:1 minimized:1 report:1 xhi:2 randomly:1 ve:1 ab:1 technometrics:1 interest:7 introduces:1 tj:1 iiy:1 accurate:2 kt:4 plotted:1 twostep:1 effector:1 kij:2 soft:1 introducing:1 subset:1 hundred:1 conducted:1 loo:2 corrupted:1 international:1 yl:8 squared:1 again:1 choose:3 derivative:1 rly:1 coefficient:1 satisfy:1 performed:3 jason:1 closed:1 red:1 minimize:6 kaufmann:1 proccedings:1 barnhill:1 fo:1 against:3 surrey:1 regress:1 dataset:6 massachusetts:1 remp:2 vlad:1 organized:1 ta:2 adaboost:5 improved:2 nonlinear:1 quality:1 russian:2 usa:2 ye:11 contain:2 unbiased:1 inductive:6 regularization:1 laboratory:1 coincides:1 tt:1 demonstrate:3 ridge:23 performs:1 functional:12 twist:1 stepping:1 approximates:1 m1:1 measurement:1 llx:2 chapelle:3 ylf:1 robot:1 xe:3 yi:16 muller:1 morgan:1 minimum:10 technical:1 prediction:1 regression:41 ae:2 breast:1 expectation:2 df:3 achieved:1 extra:1 biased:1 call:1 structural:1 intermediate:1 split:2 ktt:1 brailovsky:1 economic:1 idea:1 expression:3 six:1 york:1 se:1 amount:1 simplest:1 sl:2 percentage:1 estimated:1 xii:1 drawn:2 pj:1 run:2 angle:1 almost:1 decision:1 summarizes:1 bound:2 followed:2 constraint:2 locational:1 thyroid:1 performing:1 according:1 describes:1 character:1 partitioned:1 restricted:1 census:1 taken:2 heart:1 equation:1 remains:1 discus:1 abr:2 end:1 observe:2 egham:1 remaining:1 calculating:1 giving:1 especially:1 uj:1 classical:4 objective:1 xiw:1 fa:2 dependence:1 traditional:1 distance:1 separate:1 link:1 seven:1 collected:1 discriminant:1 minimizing:3 vladimir:1 setup:2 sigma:2 unknown:5 observation:6 datasets:4 looking:1 banana:1 y1:1 aly:1 introduced:1 pair:5 required:3 usually:1 pattern:8 royal:2 regularized:2 arm:1 scheme:1 titanic:1 hm:1 kj:1 revolute:1 loss:2 interesting:1 versus:1 validation:1 incurred:1 article:2 tld:2 bank:1 pi:1 cancer:1 course:1 summary:1 free:2 taking:2 leaveone:1 valid:1 forward:1 obtains:1 ml:1 xi:7 table:2 onoda:1 mse:3 constructing:1 domain:1 noise:1 kennard:1 georgia:1 fashion:1 position:1 sf:1 xl:20 house:1 kin:4 removing:1 xt:3 svm:2 exists:1 vapnik:6 cpj:1 margin:1 boston:5 corresponds:1 weston:4 goal:1 price:1 springerverlag:1 vovk:1 called:2 experimental:2 la:1 holloway:2 support:2 bioinformatics:1 ex:1
764
17
622 LEARNING A COLOR ALGORITHM FROM EXAMPLES Anya C. Hurlbert and Tomaso A. Poggio Artificial Intelligence Laboratory and Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA ABSTRACT A lightness algorithm that separates surface reflectance from illumination in a Mondrian world is synthesized automatically from a set of examples, pairs of input (image irradiance) and desired output (surface reflectance). The algorithm, which resembles a new lightness algorithm recently proposed by Land, is approximately equivalent to filtering the image through a center-surround receptive field in individual chromatic channels. The synthesizing technique, optimal linear estimation, requires only one assumption, that the operator that transforms input into output is linear. This assumption is true for a certain class of early vision algorithms that may therefore be synthesized in a similar way from examples. Other methods of synthesizing algorithms from examples, or "learning", such as backpropagation, do not yield a significantly different or better lightness algorithm in the Mondrian world. The linear estimation and backpropagation techniques both produce simultaneous brightness contrast effects. The problems that a visual system must solve in decoding two-dimensional images into three-dimensional scenes (inverse optics problems) are difficult: the information supplied by an image is not sufficient by itself to specify a unique scene. To reduce the number of possible interpretations of images, visual systems, whether artificial or biological, must make use of natural constraints, assumptions about the physical properties of surfaces and lights. Computational vision scientists have derived effective solutions for some inverse optics problems (such as computing depth from binocular disparity) by determining the appropriate natural constraints and embedding them in algorithms. How might a visual system discover and exploit natural constraints on its own? We address a simpler question: Given only a set of examples of input images and desired output solutions, can a visual system synthesize. or "learn", the algorithm that converts input to output? We find that an algorithm for computing color in a restricted world can be constructed from examples using standard techniques of optimal linear estimation. The computation of color is a prime example of the difficult problems of inverse optics. We do not merely discriminate betwN'n different wavelengths of light; we assign @ American Institute of Physics 1988 623 roughly constant colors to objects even though the light signals they send to our eyes change as the illumination varies across space and chromatic spectrum. The computational goal underlying color constancy seems to be to extract the invariant surface spectral reflectance properties from the image irradiance, in which reflectance and iI-" lumination are mixed 1 ? Lightness algorithms 2-8, pioneered by Land, assume that the color of an object can be specified by its lightness, or relative surface reflectance, in each of three independent chromatic channels, and that lightness is computed in the same way in each channel. Computing color is thereby reduced to extracting surface reflectance from the image irradiance in a single chromatic channel. The image irra.diance, s', is proportional to the product of the illumination intensity e' and the surface reflectance r' in that channel: s' (x, y) = r' (x, y )e' (x, y). (1 ) This form of the image intensity equation is true for a Lambertian reflectance model, in which the irradiance s' has no specular components, and for appropriately chosen color channels 9. Taking the logarithm of both sides converts it to a sum: s(x, y) = rex, y) + e(x,y), (2) where s = loges'), r = log(r') and e = log(e'). Given s(x,y) alone, the problem of solving Eq. 2 for r(x,y) is underconstrained. Lightness algorithms constrain the problem by restricting their domain to a world of Mondrians, two-dimensional surfaces covered with patches of random colors 2 and by exploiting two constraints in that world: (i) r'(x,y) is unifonn within patches but has sharp discontinuities at edges between patches and (ii) e' (x, y) varies smoothly across the Mondrian. Under these constraints, lightness algorithms can recover a good approximation to r( x, y) and so can recover lightness triplets that label roughly constant colors 10. We ask whether it is possible to synthesize from examples an algorithm that ex? tracts reflectance from image irradiance. and whether the synthesized algorithm will resemble existing lightness algorithms derived from an explicit analysis of the constraints. We make one assumption, that the operator that transforms irradiance into reflectance is linear. Under that assumption, motivated by considerations discussed later, we use optimal linear estimation techniques to synthesize an operator from examples. The examples are pairs of images: an input image of a Mondrian under illumination that varies smoothly across space and its desired output image that displays the reflectance of the Mondrian without the illumination. The technique finds the linear estimator that best maps input into desired output. in the least squares sense. For computational convenience we use one-dimensional "training vectors" that represent vertical scan lines across the ~londrian images (Fig. 1). We generate many 624 1S0t? -~ ? ~ 100 , '. a , ~~~--------------~~--~--~~ SO 100 110 100 llO 100 lilt., Input d.t. i;[:2:=:0 a SO 101 ISO ZOO UI :~kfhJfEirQ b 0 )01 II 100 ISO 100 110 p/Jte' f~l o 50 100 UO ZOO ZSO 100 I"'" c )00 .1 p,xe' 100 ISO 100 110 lot p'." OlltPllt lIlll.l . .ll.a Fig. 1. (a) The input data, a one-dimensional vector 320 pixels long. Its random Mondrian reflectance pattern is superimposed on a linear illumination gradient with a random slope and offset. (b) shows the corresponding output solution, on the left the illumination and on the right reBectance. We used 1500 such pairs of inputoutput examples (each different from the others) to train the operator shown in Fig. 2. (c) shows the result obtained by the estimated operator when it acts on the input data (a), not part of the training set. On the left is the illumination and on the right the reflectance, to be compared with (b). This result is fairly typical: in some cases the prediction is even better, in others it is worse. different input vectors s by adding together different random T and e vectors, according to Eq. 2. Each vector r represents a pattern of step changes across space, corresponding to one column of a reHectance image. The step changes occur at random pixels and are of random amplitude between set minimum and maximum values. Each vector t represents a smooth gradient across space with a random offset and slope, correspondin~ to one column of an illumination image. We th~n arrange the training vectors sand r as the columns of two matrices Sand R, resp~ti?.. ely. Our goal is then to compute the optimal solution L of LS =R where L is a linear operator represented as a matrix. 625 It is well known that the solution of this equation that is optimal in the least squares sense is ( 4) where S+ is the Moore-Penrose pseudoinverse 11. We compute the pseudoinverse by overconstraining the problem - using many more training vectors than there are number of pixels in each vector - and using the straightforward formula that applies in the overconstrained case 12: S+ ST(SST)-l. The operator L computed in this way recovers a good approximation to the correct output vector r when given a new s, not part of the training set, as input (Fig. Ic). A second operator, estimated in the same way, recovers the illumination e. Acting on a random two-dimensional Mondrian L also yields a satisfactory approximation to the correct output image. Our estimation scheme successfully synthesizes an algorithm that performs the lightness computation in a Mondrian world. What is the algorithm and what is its relationship to other lightness algorithms? To answer these questions we examine the structure of the matrix L. We assume that, although the operator is not a convolution operator, it should approximate one far from the boundaries of the image. That is, in its central part, the operator should be space-invariant, performing the same action on each point in the image. Each row in the central part of L should therefore be the same as the row above but displaced by one element to the right. Inspection of the matrix confirmes this expectation. To find the form of L in its center, we thus average the rows there, first shifting them appropriately. The result, shown in Fig. 2, is a space-invariant filter with a narrow positive peak and a broad, shallow, negative surround. Interestingly, the filter our scheme synthesizes is very similar to Land's most recent retinex operator 5, which divides the image irradiance at each pixel by a weighted average of the irradiance at all pixels in a large surround and takes the logarithm of that result to yield lightness 13. The lightness triplets computed by the retinex operator agree well with human perception in a Mondrian world. The retinex operator and our matrix L both differ from Land's earlier retinex algorithms, which require a non-linear thresholding step to eliminate smooth gradients of illumination. The shape of the filter in Fig. 2, particularly of its large surround, is also suggestive of the "nonclassical" receptive fields that have been found in V4, a cortical area implicated in mechanisms underlying color constancy 14-17. The form of the space-invariant filter is similar to that derived in our earlier formal analysis of the lightness problem 8. It is qualitatively the same as that which results from the direct application of regularization methods exploiting the spatial constraints on reflectance and illumination described above 9.18.19. The Fourier transform of the filter of Fig. 2 is approximately a bandpass filter that cuts out low frequencies due = 626 ~ 0 ( .) ~ C' C - -80 s::. 0 Pi xe Is .2' ~ a -80 ----------- o +80 Pixels Fig. 2. The space-invariant part of the estimated operator, obtained by shifting and averaging the rows of a 160-pixel-wide central square of the matrix L, trained on a set of 1500 examples with linear illumination gradients (see Fig. 1). When logarithmic illumination gradients are used , a qualitatively similar receptive field is obtained. In a separate experiment we use a training set of one-dimensional Mondrians with either linear illumination gradients or slowly varying sinusoidal illumination components with random wavelength, phase and amplitude. T he resulting filter is shown in the inset. The surrounds of both filters extend beyond the range we can estimate reliably, the range we show here. to slow gradients of illumination and preserves intennediate frequencies due to step changes in reflectance. In contrast, the operator that recovers the illumination, e. takes the form of a low-pass filter. \Ve stress that the entire operator L is not a space-invariant filter. In this context, it is clear that the shape of the estimated operator should vary with the type of illumination gradient in the training set. We synthesize a second operator using a new set of examples that contain equal numbers of vectors with random, sinusoidally varying illumination components and VE"(tors with random, linear illumination gradients. Whereas the first operator, synthE.>Sized from examples with strictly linear illumination gradients, has a broad negative surround that remains virtually constant throughout its extent, the new operator's surround (Fig . 2, inset) has a smaller ext(,111 627 and decays smoothly towards zero from its peak negative value in its center. We also apply the operator in Fig. 2 to new input vectors in which the density and amplitude of the step changes of reflectance differ greatly from those on which the operator is trained. The operator performs well, for example, on an input vector representing one column of an image of a small patch of one reflectance against a uniform background of a different reflectance, the entire image under a linear illumination gradient. This result is consistent with psychophysical experiments that show that color constancy of a patch holds when its Mondrian background is replaced by an equivalent grey background 20. The operator also produces simultaneous brightness contrast, as expected from the shape and sign of its surround. The output reflectance it computes for a patch of fixed input reflectance decreases linearly with increasing average irradiance of the input test vector in which the patch appears. Similarly, to us, a dark patch appears darker when against a light background than against a dark one. This result takes one step towards explaining such illusions as the Koffka Ring 21. A uniform gray annulus against a bipartite background (Fig. 3a) appears to split into two halves of different lightnesses when the midline between the light and dark halves of the background is drawn across the annulus (Fig. 3b). The estimated operator acting on the Koffka Ring of Fig. 3b reproduces our perception by assigning a lower output reflectance to the left half of the annulus (which appears darker to us) than to the right half 22. Yet the operator gives this brightness contrast effect whether or not the midline is drawn across the annulus (Fig. 3c). Becau~e the opf'rator can perform only a linear transformation between the input and output images, it is not surprising that the addition of the midline in the input evokes so little change in the output. These results demonstrate that the linear operator alone cannot compute lightness in all worlds and suggest that an additional operator might be necessary to mark and guide it within bounded regions. Our estimation procedure is motivated by our previous observation 9.23,18 that standard regularization algorithms 19 in early vision define linear mappings between input and output and therefore can be estimated associatively under certain condi? tions. The technique of optimal linear estimation that we use is closely related to optimal Bayesian estimation 9. If we were to assume from the start that the optimal linear operator is space-invariant, we could considerably simplify (and streamline) the computation by using standard correlation te<:hniques 9.24. How does our estimation technique compare with other methods of "learning" a lightness algorithm? We can compute the r~ularized pseudoinverse using gradient descent on a "neural" network 25 with linf'ar units. Since the pseudoinverse is lhf" unique best linear approximation in the L1 norm. a gradient descent method that 628 minimizes the square error between the actual output and desired output of a fully connected linear network is guaranteed to converge, albeit slowly. Thus gradient descent in weight space converges to the same result as our first technique, the global minimum. b a c .n sa 0 .It .sa _ ut _ input data pixel ...~~ ~:== :iu II ... ,, e. sa . ~ - ~ . _ . ut _ output reflectance - with edge ~'~ ...:i'I~~ I' _ . .' - .. =.. e. " ~. . _ I ~ ~ ( ut , _ output reflectance - without edge Fig. 3. (a) Koffka Ring. (b) Koftka Ring with midline drawn across annulus. (c) Horizontal scan lines across Koffka Ring. Top: Scan line starting at arrow in (b). Middle: Scan line at corresponding location in the output of linear operator acting on (b). Bottom: Scan line at same location in the output of operator acting on (a). 629 We also compare the linear estimation technique with a "backpropagation" network: gradient descent on a 2-layer network with sigmoid units 25 (32 inputs, 32 "hidden units", and 32 linear outputs), using training vectors 32 pixels long. The network requires an order of magnitude more time to converge to a stable configuration than does the linear estimator for the same set of 32-pixel examples. The network's performance is slightly, yet consistently, better, measured as the root-mean-square error in output, averaged over sets of at least 2000 new input vectors. Interestingly, the backpropagation network and the linear estimator err in the same way on the same input vectors. It is possible that the backpropagation network may show considerable inprovement over the linear estimator in a world more complex than the Mondrian one. We are presently examining its performance on images with real-world features such as shading, shadows, and highlights26. We do not think that our results mean that color constancy may be learned during a critical period by biological organisms. It seems more reasonable to consider them simply as a demonstration on a toy world that in the course of evolution a visual system may recover and exploit natural constraints hidden in the physics of the world. The significance of our results lies in the facts that a simple statistical technique may be used to synthesize a lightness algorithm from examples; that the technique does as well as other techniques such as backpropagation; and that a similar technique may be used for other problems in early vision. Furthermore, the synthesized operator resembles both Land's psychophysically-tested retinex operator and a neuronal nonclassical receptive field. The operator's properties suggest that simultaneous color (or brightness) contrast might be the result of the visual system's attempt to discount illumination gradients 27 REFERENCES AND NOTES 1. Since we do not have perfect color constancy, our visual system must not extract reflectance exactly. The limits on color constancy might reveal limits on the underlying computation. 2. 3. 4. and S. 5. 6. E.H. Land, Am. Sci. 52,247 (1964). E.H. Land and J.J. McCann, J. Opt. Soc. Am. 61, 1 {1971}. E.H. Land, in Central and Peripheral Mechanisms of Colour Vision, T. Ottoson Zeki, Eds., (Macmillan, New York, 1985), pp. 5-17. E.H. Land, Proc. Nat. Acad. Sci. USA 83, 3078 (1986). B.K.P. Hom, Computer Graphics and Image Processing 3, 277 (1974). 630 7. A. Blake, in Central and Peripheral Mechanisms of Colour Vision, T. Ottoson and S. Zeki, Eds., (Macmillan, New York, 1985), pp. 45-59. 8. A. Hurlbert, J. Opt. Soc. Am. A 3,1684 (1986). 9. A. Hurlbert and T. Poggio, ArtificiaLIntelligence Laboratory Memo 909, (M.LT., Cambridge, MA, 1987). 10. r'{x,y) can be recovered at best only to within a constant, since Eq. 1 is invariant under the transformation of r' int.o ar' and e' into a-ie', where a is a constant. 11. A. Albert, Regression and the Moore-Penrose Pseudoinllerse, (Academic Press, New York, 1972). 12. The pseudoinverse, and therefore L, may also be computed by recursive techniques that improve its form as more data become available l l . 13. Our synthesized filter is not exactly identical with Land's: the filter of Fig. 2 subtracts from the value at each point the average value of the logarithm of irradiance at all pixels, rather than the logarithm of the average values. The estimated operator is therefore linear in the logarithms, whereas Land's is not. The numerical difference between the outputs of the two filters is small in most cases (Land, personal communication), and both agree well with psychophysical results. 14. R. Desimone, S.J. Schein, J. Moran and L.G. Ungerleider, Vision Res. 25, 441 (1985). 15. H.M. Wild, S.R. Butler, D. Carden and J.J. Kulikowski, Nature (London) 313, 133 (1985). 16. S.M. Zeki, Neuroscience 9, 741 (1983). 17. S.M. Zeki, Neuroscience 9, 767 (1983). 18. T. Poggio, et. al, in Proceedings Image Understanding Workshop, L. Baumann, Ed., (Science Applications International Corporation, McLean, VA, 1985), pp.' 25-39. 19. T. Poggio, V. Torre and C. Koch, Nature (London) 317,314 (1985). 20. A. Valberg and B. Lange-Malecki, Investigative Ophthalmology and Visual Science Supplement 28, 92 (1987). 21. K. Koffka, Principles of Gestalt Psychology, (Harcourt, Brace and Co., New York, 1935). 22. Note that the operator achieves this effect by subtracting a non-existent illumination gradient from the input signal. 23. T. Poggio and A. Hurlbert, Artificial Intelligence Laboratory Working Paper 264, (M.LT., Cambridge, MA, 1984). 24. Estimation of the operator on two-dimensional examples is possible, but computationally very expensive if done in the same way. The present computer simulations require several hours when run on standard serial computers. The two-dimensional case 631 will need much more time (our one-dimensional estimation scheme runs orders of magnitude faster on a CM-1 Connection Machine System with 16K-processors). 25. D. E. Rumelhart, G.E. Hinton and R.J. Williams, Nature (London) 323, 533 (1986 ). 26. A. Hurlbert, The Computation of Color, Ph.D. Thesis, M.l. T., Cambridge, MA, in preparation. 2i. We are grateful to E. Land, E. Hildreth, .J. Little, F. Wilczek and D. Hillis for reading the draft and for useful discussions. A. Rottenberg developed the routines for matrix operations that we used on the Connection Machine. T. Breuel wrote the backpropagation simulator.
17 |@word middle:1 seems:2 norm:1 grey:1 simulation:1 llo:1 brightness:4 thereby:1 shading:1 configuration:1 disparity:1 correspondin:1 interestingly:2 existing:1 err:1 recovered:1 surprising:1 assigning:1 yet:2 must:3 numerical:1 shape:3 alone:2 intelligence:2 half:4 inspection:1 iso:3 draft:1 location:2 simpler:1 constructed:1 direct:1 become:1 wild:1 mccann:1 expected:1 roughly:2 tomaso:1 examine:1 simulator:1 brain:1 automatically:1 little:2 actual:1 increasing:1 discover:1 underlying:3 bounded:1 intennediate:1 what:2 cm:1 minimizes:1 developed:1 transformation:2 corporation:1 act:1 ti:1 exactly:2 unit:3 uo:1 positive:1 scientist:1 limit:2 acad:1 ext:1 approximately:2 might:4 resembles:2 co:1 range:2 averaged:1 unique:2 recursive:1 backpropagation:7 illusion:1 procedure:1 rator:1 area:1 significantly:1 suggest:2 convenience:1 cannot:1 operator:38 context:1 equivalent:2 map:1 center:3 send:1 straightforward:1 williams:1 starting:1 l:1 estimator:4 embedding:1 resp:1 pioneered:1 synthesize:5 element:1 expensive:1 particularly:1 rumelhart:1 cut:1 bottom:1 constancy:6 region:1 connected:1 decrease:1 ui:1 personal:1 existent:1 trained:2 grateful:1 solving:1 mondrian:13 bipartite:1 represented:1 train:1 investigative:1 effective:1 london:3 artificial:3 solve:1 think:1 transform:1 itself:1 breuel:1 nonclassical:2 subtracting:1 product:1 inputoutput:1 exploiting:2 produce:2 tract:1 ring:5 converges:1 object:2 tions:1 perfect:1 measured:1 sa:3 eq:3 soc:2 streamline:1 resemble:1 shadow:1 differ:2 closely:1 correct:2 torre:1 filter:13 human:1 sand:2 require:2 assign:1 opt:2 biological:2 strictly:1 hold:1 koch:1 ic:1 blake:1 ungerleider:1 mapping:1 tor:1 arrange:1 early:3 vary:1 achieves:1 estimation:12 proc:1 label:1 successfully:1 weighted:1 rather:1 chromatic:4 varying:2 derived:3 consistently:1 superimposed:1 greatly:1 contrast:5 sense:2 am:3 eliminate:1 entire:2 hidden:2 pixel:11 iu:1 spatial:1 fairly:1 field:4 equal:1 identical:1 represents:2 broad:2 others:2 simplify:1 preserve:1 ve:2 koffka:5 individual:1 midline:4 replaced:1 phase:1 attempt:1 light:5 edge:3 desimone:1 necessary:1 poggio:5 divide:1 logarithm:5 loge:1 desired:5 re:1 schein:1 column:4 earlier:2 sinusoidally:1 ar:2 uniform:2 examining:1 graphic:1 rex:1 answer:1 varies:3 considerably:1 psychophysically:1 st:1 density:1 peak:2 international:1 ie:1 v4:1 physic:2 decoding:1 together:1 thesis:1 central:5 slowly:2 worse:1 cognitive:1 mclean:1 american:1 toy:1 sinusoidal:1 int:1 ely:1 later:1 root:1 lot:1 start:1 recover:3 slope:2 square:5 yield:3 bayesian:1 annulus:5 zoo:2 processor:1 simultaneous:3 ed:3 against:4 frequency:2 pp:3 hurlbert:5 recovers:3 massachusetts:2 ask:1 color:17 ut:3 overconstrained:1 amplitude:3 routine:1 irradiance:10 appears:4 specify:1 done:1 though:1 furthermore:1 binocular:1 correlation:1 working:1 horizontal:1 harcourt:1 wilczek:1 hildreth:1 gray:1 reveal:1 usa:2 effect:3 contain:1 true:2 evolution:1 regularization:2 laboratory:3 moore:2 satisfactory:1 ll:1 during:1 stress:1 demonstrate:1 performs:2 l1:1 image:27 consideration:1 recently:1 sigmoid:1 physical:1 discussed:1 interpretation:1 he:1 extend:1 synthesized:5 organism:1 cambridge:4 surround:8 similarly:1 stable:1 surface:8 own:1 recent:1 prime:1 certain:2 xe:2 minimum:2 additional:1 converge:2 period:1 signal:2 ii:4 smooth:2 faster:1 academic:1 long:2 serial:1 va:1 prediction:1 regression:1 vision:7 expectation:1 albert:1 represent:1 condi:1 whereas:2 background:6 addition:1 appropriately:2 brace:1 associatively:1 virtually:1 extracting:1 split:1 specular:1 psychology:1 reduce:1 lange:1 whether:4 motivated:2 colour:2 york:4 action:1 useful:1 covered:1 sst:1 clear:1 transforms:2 dark:3 discount:1 ph:1 reduced:1 generate:1 baumann:1 supplied:1 unifonn:1 sign:1 estimated:7 neuroscience:2 zeki:4 drawn:3 merely:1 convert:2 sum:1 run:2 inverse:3 evokes:1 throughout:1 reasonable:1 patch:8 lilt:1 layer:1 hom:1 guaranteed:1 display:1 occur:1 optic:3 constraint:8 constrain:1 scene:2 synthe:1 fourier:1 performing:1 department:1 according:1 peripheral:2 across:10 smaller:1 slightly:1 ophthalmology:1 shallow:1 presently:1 restricted:1 invariant:8 computationally:1 equation:2 agree:2 remains:1 mechanism:3 available:1 operation:1 apply:1 lambertian:1 appropriate:1 spectral:1 top:1 kulikowski:1 exploit:2 reflectance:24 psychophysical:2 question:2 receptive:4 gradient:17 separate:2 sci:2 extent:1 relationship:1 demonstration:1 difficult:2 negative:3 memo:1 synthesizing:2 reliably:1 perform:1 vertical:1 convolution:1 displaced:1 observation:1 descent:4 hinton:1 communication:1 sharp:1 intensity:2 pair:3 specified:1 connection:2 learned:1 narrow:1 hour:1 hillis:1 discontinuity:1 address:1 beyond:1 pattern:2 perception:2 reading:1 shifting:2 critical:1 natural:4 representing:1 scheme:3 improve:1 technology:1 lightness:19 eye:1 linf:1 extract:2 understanding:1 determining:1 relative:1 anya:1 opf:1 fully:1 mixed:1 filtering:1 proportional:1 sufficient:1 consistent:1 thresholding:1 principle:1 pi:1 land:13 row:4 course:1 koftka:1 implicated:1 side:1 formal:1 guide:1 institute:2 wide:1 explaining:1 taking:1 boundary:1 depth:1 cortical:1 world:12 computes:1 qualitatively:2 subtracts:1 far:1 gestalt:1 approximate:1 wrote:1 pseudoinverse:5 suggestive:1 reproduces:1 global:1 butler:1 spectrum:1 triplet:2 channel:6 learn:1 nature:3 synthesizes:2 complex:1 domain:1 significance:1 linearly:1 arrow:1 neuronal:1 fig:17 slow:1 darker:2 s0t:1 explicit:1 bandpass:1 lie:1 formula:1 inset:2 moran:1 offset:2 decay:1 workshop:1 underconstrained:1 restricting:1 adding:1 albeit:1 supplement:1 magnitude:2 te:1 illumination:25 nat:1 smoothly:3 logarithmic:1 lt:2 wavelength:2 simply:1 visual:8 penrose:2 macmillan:2 applies:1 ma:3 goal:2 sized:1 towards:2 considerable:1 change:6 typical:1 acting:4 averaging:1 discriminate:1 pas:1 mark:1 retinex:5 scan:5 ularized:1 preparation:1 tested:1 ex:1
765
170
795 SONG LEARNING IN BIRDS M. Konishi Division of Biology California Institute of Technology Birds sing to communicate. Male birds use song to advertise their territories and attract females. Each bird species has a unique song or set of songs. Song conveys both species and individual identity. In most species, young birds learn some features of adult song. Song develops gradually from amorphous to fixed patterns of vocalization as if crystals form out of liquid. Learning of a song proceeds in two steps; birds commit the song to memory in the first stage and then they vocally reproduce it in the second stage. The two stages overlap each other in some species, while they are separated by several months in other species. The ability of a bird to commit a song to memory is restricted to a period known as the sensitive phase. Vocal reproduction of the memorized song requires auditory feedback. Birds deafened before the second stage cannot reproduce the memorized song. Birds change vocal output until it matches with the memorized song, which thus serves as a template. Birds use a built-in template when a tutor model is not available. Exposure to a tutor model modifies this innate template. A series of brain nuclei controls song production and patterning. Recording multiand single neurons from this nuclei in the singing bird is possible. The learned temporal pattern of song is recognizable in the neural discharge of these nuclei. The need for auditory feedback for song learning suggests the presence of links between the auditory and vocal control systems. One such link is found in the HVc, one of the forebrain song nuclei. This nucleus contains neurons sensitive to sound in addition to those which control song production. In the white-crowned sparrow, the HVc contains neurons selective for the bird's own individual song. The stimulus selectivity of these neurons are thus shaped by the bird's hearing of its own voice during song development. [1] Konishi, M. (1985) Birdson: from behavior to neuron. Ann. Rev. Neurosci. 8:125-170. [2] Konishi, M. (1985) The role of auditory feedback in the control of vocalization in the white-crowned sparrow. Z. Tierpsychol. 22:770-783. [3] McCasland, J. S. (1987) Neuronal control of bird song production. J. Neurosci., 723-739. [4] Margoliash, D. (1983) Acoustic parameters underlying the responses of songspecific neurons in the white-crowned sparrow. J. Neurosci. 3:10389-1057. [5] Nottebohm, F. T., Stokes, M., & Leonard, C. M. (1976) Central control of song in the canary Serinus canarius. J. Compo Neurol. 165:457-486.
170 |@word tutor:2 white:3 during:1 memorized:3 link:2 series:1 contains:2 liquid:1 crystal:1 patterning:1 neuron:6 sensitive:2 sing:1 compo:1 stokes:1 crowned:3 own:2 female:1 recognizable:1 selectivity:1 acoustic:1 advertise:1 california:1 learned:1 behavior:1 adult:1 brain:1 attract:1 proceeds:1 pattern:2 reproduce:2 period:1 built:1 selective:1 memory:2 underlying:1 sound:1 overlap:1 sparrow:3 match:1 development:1 shaped:1 technology:1 biology:1 temporal:1 canary:1 stimulus:1 develops:1 control:6 addition:1 before:1 individual:2 phase:1 recording:1 nucleus:5 bird:14 suggests:1 male:1 presence:1 production:3 unique:1 institute:1 template:3 feedback:3 hvc:2 vocal:3 song:22 cannot:1 hearing:1 modifies:1 exposure:1 nottebohm:1 learn:1 konishi:3 discharge:1 margoliash:1 central:1 neurosci:3 role:1 neuronal:1 communicate:1 singing:1 young:1 amorphous:1 division:1 neurol:1 reproduction:1 territory:1 separated:1 deafened:1 rev:1 conveys:1 ability:1 gradually:1 commit:2 restricted:1 auditory:4 vocalization:2 forebrain:1 identity:1 month:1 ann:1 leonard:1 serf:1 change:1 available:1 response:1 specie:5 stage:4 until:1 voice:1 innate:1
766
1,700
Regular and Irregular Gallager-type Error-Correcting Codes Y. Kabashirna and T. Murayarna Dept. of Compt. IntI. & Syst. Sci. Tokyo Institute of Technology Yokohama 2268502, Japan D. Saad and R. Vicente Neural Computing Research Group Aston University Birmingham B4 7ET, UK Abstract The performance of regular and irregular Gallager-type errorcorrecting code is investigated via methods of statistical physics. The transmitted codeword comprises products of the original message bits selected by two randomly-constructed sparse matrices; the number of non-zero row/column elements in these matrices constitutes a family of codes. We show that Shannon's channel capacity may be saturated in equilibrium for many of the regular codes while slightly lower performance is obtained for others which may be of higher practical relevance. Decoding aspects are considered by employing the TAP approach which is identical to the commonly used belief-propagation-based decoding. We show that irregular codes may saturate Shannon's capacity but with improved dynamical properties. 1 Introduction The ever increasing information transmission in the modern world is based on reliably communicating messages through noisy transmission channels; these can be telephone lines, deep space, magnetic storing media etc. Error-correcting codes play a significant role in correcting errors incurred during transmission; this is carried out by encoding the message prior to transmission and decoding the corrupted received code-word for retrieving the original message. In his ground breaking papers, Shannon[l] analyzed the capacity of communication channels, setting an upper bound to the achievable noise-correction capability of codes, given their code (or symbol) rate, constituted by the ratio between the number of bits in the original message and the transmitted code-word. Shannon's bound is non-constructive and does not provide a recipe for devising optimal codes. The quest for more efficient codes, in the hope of saturating the bound set by Shannon, has been going on ever since, providing many useful but sub-optimal codes. One family of codes, presented originally by Gallager[2]' attracted significant interest recently as it has been shown to outperform most currently used techniques[3]. Gallager-type codes are characterized by several parameters, the choice of which defines a particular member of this family of codes. Current theoretical results[3] 273 Regular and Irregular Gallager-type Error-Correcting Codes offer only bounds on the error probability of various architectures, proving the existence of very good codes under some restrictions; decoding issues are examined via numerical simulations. In this paper we analyze the typical performance of Gallager-type codes for several parameter choices via methods of statistical mechanics. We then validate the analytical solution by comparing the results to those obtained by the TAP approach and via numerical methods. 2 The general framework In a general scenario, a message represented by an N dimensional Boolean vector JO which is transmitted through a noisy channel with some flipping probability p per bit (other noise types may also be studied). The received message J is then decoded to retrieve the original message. eis encoded to the M dimensional vector In this paper we analyze a slightly different version of Gallager-type codes termed the MN code[3] that is based on choosing two randomly-selected sparse matrices A and B of dimensionality M x N and M x M respectively; these are characterized by K and L non-zero unit elements per row and C and L per column respectively. The finite numbers K, C and L define a particular code; both matrices are known to both sender and receiver. Encoding is carried out by constructing the modulo 2 inverse of B and the matrix B- 1 A (mod 2); the vector JO = B- 1 A (mod 2, Boolean vector) constitutes the codeword. Decoding is carried out by taking the product of the matrix B and the received message J = JO +( (mod 2), corrupted by the Boolean noise vector (, resulting in Ae + B (. The equation e Ae + B( = AS + B'T (mod 2) e (1) is solved via the iterative methods of Belief Propagation (BP)[3] to obtain the most probable Boolean vectors Sand 'T; BP methods in the context of error-correcting codes have recently been shown to be identical to a TAP[4] based solution of a similar physical system[5]. The similarity between error-correcting codes of this type and Ising spin systems was first pointed out by Sourlas[6], who formulated the mapping of a simpler code, somewhat similar to the one presented here, onto an Ising spin system Hamiltonian. We recently extended the work of Sourlas, that focused on extensively connected systems, to the finite connectivity case[5] as well as to the case of MN codes [7]. To facilitate the current investigation we first map the problem to that of an Ising model with finite connectivity. We employ the binary representation (?1) of the dynamical variables Sand 'T and of the vectors J and JO rather than the Boolean (0,1) one; the vector JO is generated by taking products of the relevant binary = TIiE/.' ~i' where the indices J.L = (h, ... iK) correspond to the message bits non-zero elements of B-1 A, producing a binary version of JO. As we use statistical mechanics techniques, we consider the message and codeword dimensionality (N and M respectively) to be infinite, keeping the ratio between them R = N 1M, which constitutes the code rate, finite. Using the thermodynamic limit is quite natural as Gallager-type codes are usually used for transmitting long (10 4 - 105 ) messages, where finite size corrections are likely to be negligible. To explore the system's capabilities we examine the Hamiltonian J2 Y. Kabashima, T. Murayama, D. Saad and R. Vicente 274 The tensor product DlJ.uJ,.J.{Tl where JlJ.u = TIiEIJ. ~i TI jEu (j and u = (jl,'" iL), is the binary equivalent of Ae + B(, treating both signal (8 and index i) and noise (7" and index j) simultaneously. Elements of the sparse connectivity tensor D IJ.U take the value 1 if the corresponding indices of both signal and noise are chosen (Le., if all corresponding indices of the matrices A and Bare 1) and 0 otherwise; it has C unit elements per i-index and L per j-index representing the system's degree of connectivity. The f> function provides 1 if the selected sites' product TIiEIJ. Si TIjEu Tj is in disagreement with the corresponding element JIJ.U, recording an error, and 0 otherwise. Notice that this term is not frustrated, as there are M +N degrees of freedom and only M constraints from Eq.(l), and can therefore vanish at sufficiently low temperatures. The last two terms on the right represent our prior knowledge in the case of sparse or biased messages Fs and of the noise level Fr and require assigning certain values to these additive fields. The choice of f3 -+ 00 imposes the restriction of Eq.(l), limiting the solutions to those for which the first term of Eq.(2) vanishes, while the last two terms, scaled with f3, survive. Note that the noise dynamical variables 7" are irrelevant to measuring the retrieval success m = Jr (~~1 ~i sign (Si)!3 ) ~ . The latter monitors the normalized mean overlap between the Bayes-optimal retrieved message, shown to correspond to the alignment of (Si)!3 to the nearest binary value[6], and the original message; the subscript f3 denotes thermal averaging. Since the first part of Eq.(2) is invariant under the map Si -+ Si~i, Tj -+ Tj(j and JIJ.U -+ JIJ.U TIiEIJ. ~i TI jEu (j = 1, it is useful to decouple the correlation between the vectors 8, 7" and Rewriting Eq.(2) one obtains a similar expression apart from the last terms on the right which become Fs / f3 L:k Sk ~k and Fr / f3 ~k Tk (k. e,(. The random selection of elements in D introduces disorder to the system which is treated via methods of statistical physics. More specifically, we calculate the partition function Z(D , J) = Tr{8 ,7"} exp[-f31i] averaged over the disorder and the statistical properties of the message and noise, using the replica method[5, 8, 9]. Taking f3 -+ 00 gives rise to a set of order parameters (~ tZi Sf Sf ,.. ,S7) q"",(3 ?..? "Y = .=1 T"".(3, ..,"Y = (~ ty; rj rf, .. ,r?) .=1 (3400 (3400 (2) where a, f3, .. represent replica indices, and the variables Zi and 1j come from enforcing the restriction of C and L connections per index respectively[5]: f> ( "D . . (. L .) . <t,t2 ,?? ,JL> '2 ,?? ,'tK - c) i = 0 21T dZ ZL: h .... i K f<i.i 2 ? .. ?h 2 >-(C+l) 7r ' (3) and similarly for the restriction on the j indices. To proceed with the calculation one has to make an assumption about the order parameters symmetry. The assumption made here, and validated later on, is that of replica symmetry in the following representation of the order parameters and the related conjugate variables Qa ,!3 ..-y aq / dx 7r(X) xl , Qa,!3 .. -y = aq- / dx 1?(x) Xl r a,!3 .. -y ar / dy p(y) yl , r a,!3 .. -y = a; / dy p(Y) yl , (4) where l is the number of replica indices, a. are normalization coefficients, and 7r(x) , 1?(x) , p(y) and p(Y) represent probability distributions. Unspecified integrals Regular and Irregular Gallager-type Error-Correcting Codes 275 are over the range [-1, + 1]. One then obtains an expression for the free energy per spin expressed in terms of these probability distributions liN (In Z)~,(,'D The free energy can then be calculated via the saddle point method. Solving the equations obtained by varying the free energy w.r.t the probability distributions 1T(X), 1?(x), p(y) and p(y), is difficult as they generally comprise both delta peaks and regular[9] solutions for the ferromagnetic and paramagnetic phases (there is no spin-glass solution here as the system is not frustrated). The solutions obtained in the case of unbiased messages (the most interesting case as most messages are compressed prior to transmission) are for the ferromagnetic phase: 1T(X) = 8(x - 1) , 1?(x) = 8(x - 1) , p(y) = 8(y - 1) , p(Y) = 8(Y - 1), (5) and for the paramagnetic phase: = 8(x) = 8(Y) 1T(X) 8(x) , 1?(x) p(y) 1 + tanh Fr r( _ h F ) 1 - tanh Fr r( hF ) 2 u y tan r + 2 u Y + tan r? , p(Y) (6) These solutions obey the saddle point equations. However, it is unclear whether the contribution of other delta peaks or of an additional continuous solution will be significant and whether the solutions (5) and (6) are stable or not. In addition, it is also necessary to validate the replica symmetric ansatz itself. To address these questions we obtained solutions to the system described by the Hamiltonian (2) via TAP methods of finitely connected systems[5]; we solved the saddle point equations derived from the free energy numerically, representing all probability distributions by up to 104 bin models and by carrying out the integrations via Monte-Carlo methods; finally, to show the consistency between theory and practice we carried out large scale simulations for several cases, which will be presented elsewhere. 3 Structure of the solutions The various methods indicate that the solutions may be divided to two different categories: K =L =2 and either K ~ 3 or L ~ 3. We therefore treat them separately. For unbiased messages and either K ~ 3 or L ~ 3 we obtain the solutions (5) and (6) both by applying the TAP approach and by solving the saddle point equations numerically. The former was carried out at the value of Fr which corresponds to the true noise and input bias levels (for unbiased messages Fa = 0) and thus to Nishimori's condition[lO], where no replica symmetry breaking effects are expected. This is equivalent to having the correct prior within the Bayesian framework[6] and enables one to obtain analytic expressions for some observables as long as some gauge requirements are obeyed [10] . Numerical solutions show the emergence of stable dominant delta peaks, consistent with those of (5) and (6). The question of longitudinal mode stability (corresponding to the replica symmetric solution) was addressed by setting initial conditions for the numerical solutions close to the solutions (5) and (6), showing that they converge back to these solutions which are therefore stable. The most interesting quantity to examine is the maximal code rate, for a given corruption process, for which messages can be perfectly retrieved. This is defined in the case of K,L~3 by the value of R=KIC=NjM for which the free energy of the ferromagnetic solution becomes smaller than that of the paramagnetic solution, constituting a first order phase transition. A schematic description of the solutions obtained is shown in the inset of Fig.1a. The paramagnetic solution (m = 0) has a lower free energy than the ferromagnetic one (low Ihigh free energies are denoted 276 Y. Kabashima, T. Murayama, D. Saad and R. Vicente by the thick and thin lines respectively, there are no axis lines at m = 0,1) for noise levels P > Pc and vice versa for P ~ Pc; both solutions are stable. The critical code rate is derived by equating the ferromagnetic and paramagnetic free energies to obtain Rc = 1-H2(p) = 1+(plog2P+(1- p)log2(1- p)) . This coincides with Shannon's capacity. To validate these results we obtained TAP solutions for the unbiased message case (K = L = 3, C = 6) as shown in Fig.1a (as +) in comparison to Shannon's capacity (solid line). Analytical solutions for the saddle point equations cannot be obtained for biased patterns and we therefore resort to numerical methods ana the TAP approach. The maximal information rate (Le., code-rate xH2 (fs = (1 + tanh Fs)/2) - the source redundancy) obtained by the TAP method (0) and numerical solutions of the saddle point equations (0), for each noise level, are shown in Fig.1a. Numerical results have been obtained using 103 _10 4 bin models for each probability distribution and had been run for 105 steps per noise level point. The various results are highly consistent and practically saturate Shannon's bound for the same noise level. The MN code for K , L ~ 3 seems to offer optimal performance. However, the main drawback is rooted in the co-existence of the stable m = 1 and m = 0 solutions, shown in Fig.1a (inset), which implies that from some initial conditions the system will converge to the undesired paramagnetic solution. Moreover, studying the ferromagnetic solution numerically shows a highly limited basin of attraction, which becomes smaller as K and L increase, while the paramagnetic solution at m = 0 always enjoys a wide basin of attraction. Computer simulations (see also [3]) show that as initial conditions for the decoding process are typically of close-to-zero magnetization (almost no prior information about the original message is assumed) it is likely that the decoding process will converge to the paramagnetic solution. While all codes with K, L ~ 3 saturate Shannon's bound in their equilibrium properties and are characterized by a first order, paramagnetic to ferromagnetic, phase transition, codes with K = L = 2 show lower performance and different physical characteristics. The analytical solutions (5) and (6) are unstable at some flip rate levels and one resorts to solving the saddle point equations numerically and to TAP based solutions. The picture that emerges is sketched in the inset of Fig.1b: The paramagnetic solution dominates the high flip rate regime up to the point PI (denoted as 1 in the inset) in which a stable, ferromagnetic solution, of higher free energy, appears (thin lines at m = ?1). At a lower flip rate value P2 the paramagnetic solution becomes unstable (dashed line) and is replaced by two stable sub-optimal ferromagnetic (broken symmetry) solutions which appear as a couple of peaks in the various probability distributions; typically, these have a lower free energy than the ferromagnetic solution until P3, after which the ferromagnetic solution becomes dominant. Still, only once the sub-optimal ferromagnetic solutions disappear, at the spinodal point Ps, a unique ferromagnetic solution emerges as a single delta peak in the numerical results (plus a mirror solution). The point in which the sub-optimal ferromagnetic solutions disappear constitutes the maximal practical flip rate for the current code-rate and was defined numerically (0) and via TAP solutions (+) as shown in Fig.1b. Notice that initial conditions for TAP and the numerical solutions were chosen almost randomly, with a slight bias of 0(10- 12 ), in the initial magnetization. The TAP dynamical equations are identical to those used for practical BP decoding[5], and therefore provide equivalent results to computer simulations with the same parameterization, supporting the analytical results. The excellent convergence results obtained point out the existence of a unique pair of global solutions to which the system converges (below Ps) from practically all initial conditions. This observation and the practical implications of using K = L = 2 code have not been obtained by Regular and Irregular Gallager-type Error-Correcting Codes 277 information theory methods (e.g.[3]}j these prove the existence of very good codes for C = L ~ 3, and examine decoding properties only via numerical simulations. 4 Irregular Constructions Irregular codes with non-uniform number of non-zero elements per column and uniform number of elements per row were recently introduced [11, 12] and were found to outperform regular codes. It is relatively straightforward to adapt our methods to study these particular constructions. The restriction of the number of connections per index can be replaced by a set of N restrictions of the form (1), enforcing Cj non-zero elements in the j-th column of the matrix A, and other M restrictions enforcing Ll non-zero elements in the l-th column of the matrix B. By construction these restrictions must obey the relations E.7=l C j = MK and E~l Ll = ML. One can assume that a particular set of restrictions is generated independently by the probability distributions P(C) and P(L). With that we can compute average properties of irregularly constructed codes generated by arbitrary distributions. Proceeding along the same lines to those of the regular case one can find a very similar expression for the free energy which can be interpreted as a mixture of regular codes with column weights sampled with probabilities P(C) and P(L). As long as we choose probability distributions which vanish for C, L = 0 (avoiding trivial non-invertible matrices) and C, L = 1 (avoiding single checked bits), the solutions to the saddle point equations are the same as those obtained in the regular case (Eqs.5, 6) leading to exactly the same predictions for the maximum performance. The differences between regular and irregular codes show up in their dynamical behavior. In the irregular case with K > 2 and for biased messages the basin of attraction is larger for higher noise levels [13]. 5 Conclusion In this paper we examined the typical performance of Gallager-type codes. We discovered that for a certain choice of parameters, either K ~ 3 or L ~ 3, one potentially obtains optimal performance, saturating Shannon's bound. This comes at the expense of a decreasing basin of attraction making the decoding process increasingly impractical. Another code, K = L = 2, shows "close to optimal performance with a very large basin of attraction, making it highly attractive for practical purposes. The decoding performance of both code types was examined by employing the TAP approach, an iterative method identical to the commonly used BP. Both numerical and TAP solutions agree with the theoretical results. The equilibrium properties of regular and irregular constructions is shown to be the same. The improved performance of irregular codes reported in the literature can be explained as consequence of dynamical properties. This study examines the typical performance of these increasingly important error-correcting codes, from which optimal parameter choices can be derived, complementing the bounds and empirical results provided in the information theory literature. Important aspects that are yet to be investigated include other noise types, finite size effects and the decoding dynamics itself. Acknowledgement Support by the JSPS RFTF program (YK), The Royal Society and EPSRC grant GR/N00562 (DS) is acknowledged. Y. Kabashima. T. Murayama. D. Saad and R. Vicente 278 1 1 Ferro 0.8 0.8 ~ 0.6 0.6 a: I a: 0.4 0.4 p 0.2 0.2 0 0.1 0.2 0.3 P 0.4 0.5 0 0.1 0.2 0.3 0.4 0.5 P Figure 1: Critical code rate as a function of the flip rate p, obtained from numerical solutions and the TAP approach (N = 104 ), and averaged over 10 different initial conditions with error bars much smaller than the symbols size. (a) Numerical solutions for K = L = 3, C = 6 and varying input bias fs (0) and TAP solutions for both unbiased (+) and biased (0) messages; initial conditions were chosen close to the analytical ones. The critical rate is multiplied by the source information content to obtain the maximal information transmission rate, which clearly does not go beyond R = 3/6 in the case of biased messages; for unbiased patterns H 2 (fs) = 1. Inset: The ferromagnetic and paramagnetic solutions as functions of p; thick and thin lines denote stable solutions of lower and higher free energies respectively. (b) For the unbiased case of K = L = 2; initial conditions for the TAP (+) and the numerical solutions (0) are of almost zero magnetization. Inset: The ferromagnetic (optimal/sub-optimal) and paramagnetic solutions as functions of p; thick and thin lines are as in (a), dashed lines correspond to unstable solutions. References [1] C.E. Shannon, Bell Sys. Tech.J., 27, 379 (1948); 27, 623 (1948). [2] R.G. Gallager, IRE Trans.Info . Theory, IT-8, 21 (1962). [3] D.J.C. MacKay, IEEE Trans.IT, 45, 399 (1999) . [4] D. Thouless, P.W. Anderson and R.G. Palmer, Phil. Mag., 35 , 593 (1977). [5] Y. Kabashima and D. Saad, Europhys.Lett., 44 668 (1998) and 45 97 (1999). [6] N. Sourlas, Nature, 339, 693 (1989) and Euro.Phys.Lett., 25 , 159 (1994). [7] Y. Kabashima, T. Murayama and D. Saad, Phys.Rev.Lett., (1999) in press. [8] K.Y.M. Wong and D. Sherrington, J.Phys.A, 20, L793 (1987). [9] C. De Dominicis and P.Mottishaw, J.Phys.A, 20, L1267 (1987). [10] H. Nishimori, Prog. Theo.Phys., 66, 1169 (1981). [11] M. Luby et. ai, IEEE proceedings of ISIT98 (1998) and Analysis of Low Density Codes and Improved Designs Using Irregular Graphs, preprint. [12] D.J.C. MacKay et. al, IEEE Trans.Comm., 47, 1449 (1999). [13] R. Vicente et. ai, http://xxx.lanl.gov/abs/cond-mat/9908358 (1999).
1700 |@word version:2 achievable:1 seems:1 simulation:5 tr:1 solid:1 initial:9 mag:1 longitudinal:1 current:3 comparing:1 paramagnetic:13 si:5 assigning:1 dx:2 attracted:1 must:1 yet:1 additive:1 numerical:14 partition:1 enables:1 analytic:1 treating:1 selected:3 devising:1 parameterization:1 complementing:1 sys:1 hamiltonian:3 provides:1 ire:1 simpler:1 rc:1 along:1 constructed:2 become:1 ik:1 retrieving:1 prove:1 expected:1 behavior:1 examine:3 mechanic:2 decreasing:1 gov:1 increasing:1 becomes:4 provided:1 moreover:1 medium:1 unspecified:1 interpreted:1 impractical:1 ti:2 njm:1 exactly:1 scaled:1 uk:1 zl:1 unit:2 grant:1 appear:1 producing:1 negligible:1 treat:1 limit:1 consequence:1 encoding:2 subscript:1 plus:1 studied:1 examined:3 equating:1 co:1 limited:1 palmer:1 range:1 averaged:2 practical:5 unique:2 practice:1 empirical:1 bell:1 word:2 regular:13 onto:1 close:4 selection:1 cannot:1 context:1 tiie:1 applying:1 wong:1 restriction:9 equivalent:3 map:2 dz:1 phil:1 straightforward:1 go:1 independently:1 focused:1 ferro:1 disorder:2 correcting:9 communicating:1 examines:1 attraction:5 his:1 retrieve:1 proving:1 stability:1 compt:1 limiting:1 construction:4 play:1 tan:2 modulo:1 element:11 jlj:1 ising:3 role:1 epsrc:1 preprint:1 solved:2 eis:1 calculate:1 ferromagnetic:16 connected:2 yk:1 vanishes:1 broken:1 comm:1 dynamic:1 carrying:1 solving:3 observables:1 various:4 represented:1 monte:1 choosing:1 europhys:1 quite:1 encoded:1 larger:1 otherwise:2 compressed:1 emergence:1 noisy:2 itself:2 analytical:5 product:5 jij:3 fr:5 maximal:4 j2:1 sourlas:3 relevant:1 murayama:4 description:1 validate:3 recipe:1 convergence:1 transmission:6 requirement:1 p:2 converges:1 tk:2 finitely:1 nearest:1 ij:1 received:3 eq:6 p2:1 come:2 indicate:1 implies:1 thick:3 drawback:1 tokyo:1 correct:1 ana:1 bin:2 sand:2 require:1 investigation:1 probable:1 correction:2 practically:2 sufficiently:1 considered:1 ground:1 exp:1 equilibrium:3 mapping:1 purpose:1 birmingham:1 currently:1 tanh:3 vice:1 gauge:1 hope:1 clearly:1 always:1 rather:1 varying:2 validated:1 derived:3 tech:1 glass:1 typically:2 relation:1 going:1 sketched:1 issue:1 spinodal:1 denoted:2 integration:1 mackay:2 field:1 comprise:1 f3:7 having:1 once:1 identical:4 survive:1 constitutes:4 thin:4 others:1 t2:1 employ:1 modern:1 randomly:3 simultaneously:1 thouless:1 replaced:2 phase:5 ab:1 freedom:1 interest:1 message:26 highly:3 saturated:1 alignment:1 introduces:1 analyzed:1 xh2:1 mixture:1 pc:2 tj:3 implication:1 integral:1 necessary:1 theoretical:2 mk:1 column:6 boolean:5 ar:1 measuring:1 uniform:2 jsps:1 gr:1 reported:1 obeyed:1 corrupted:2 density:1 peak:5 physic:2 yl:2 decoding:12 ansatz:1 invertible:1 transmitting:1 jo:6 connectivity:4 choose:1 resort:2 leading:1 japan:1 syst:1 de:1 coefficient:1 later:1 analyze:2 bayes:1 hf:1 capability:2 contribution:1 il:1 spin:4 who:1 characteristic:1 correspond:3 bayesian:1 carlo:1 corruption:1 kabashima:5 phys:5 checked:1 ty:1 jeu:2 energy:12 couple:1 sampled:1 kic:1 knowledge:1 emerges:2 dimensionality:2 cj:1 back:1 appears:1 higher:4 originally:1 xxx:1 improved:3 anderson:1 correlation:1 until:1 d:1 propagation:2 defines:1 mode:1 facilitate:1 effect:2 normalized:1 unbiased:7 true:1 former:1 symmetric:2 undesired:1 attractive:1 ll:2 during:1 rooted:1 coincides:1 sherrington:1 magnetization:3 temperature:1 recently:4 physical:2 b4:1 jl:2 slight:1 numerically:5 significant:3 rftf:1 versa:1 ai:2 consistency:1 similarly:1 pointed:1 aq:2 had:1 stable:8 similarity:1 etc:1 dominant:2 retrieved:2 irrelevant:1 apart:1 scenario:1 codeword:3 termed:1 certain:2 binary:5 success:1 transmitted:3 additional:1 somewhat:1 converge:3 signal:2 dashed:2 thermodynamic:1 rj:1 characterized:3 calculation:1 offer:2 long:3 retrieval:1 lin:1 divided:1 adapt:1 schematic:1 prediction:1 ae:3 represent:3 tzi:1 normalization:1 irregular:13 addition:1 separately:1 addressed:1 source:2 saad:6 biased:5 recording:1 member:1 mod:4 zi:1 architecture:1 perfectly:1 whether:2 expression:4 s7:1 f:6 proceed:1 deep:1 useful:2 generally:1 extensively:1 category:1 http:1 outperform:2 notice:2 sign:1 delta:4 per:11 mat:1 group:1 redundancy:1 monitor:1 acknowledged:1 rewriting:1 l793:1 replica:7 graph:1 run:1 inverse:1 prog:1 family:3 almost:3 p3:1 dy:2 bit:5 bound:8 constraint:1 bp:4 aspect:2 relatively:1 conjugate:1 jr:1 smaller:3 slightly:2 increasingly:2 rev:1 making:2 explained:1 invariant:1 inti:1 errorcorrecting:1 equation:10 agree:1 flip:5 irregularly:1 studying:1 multiplied:1 obey:2 disagreement:1 magnetic:1 luby:1 yokohama:1 existence:4 original:6 denotes:1 include:1 log2:1 uj:1 disappear:2 society:1 tensor:2 question:2 quantity:1 flipping:1 fa:1 unclear:1 sci:1 capacity:5 unstable:3 trivial:1 enforcing:3 code:51 index:12 ratio:2 providing:1 difficult:1 potentially:1 expense:1 info:1 rise:1 design:1 reliably:1 upper:1 observation:1 finite:6 thermal:1 supporting:1 extended:1 ever:2 communication:1 discovered:1 arbitrary:1 introduced:1 pair:1 lanl:1 connection:2 tap:17 qa:2 address:1 beyond:1 bar:1 trans:3 dynamical:6 usually:1 pattern:2 below:1 regime:1 program:1 rf:1 royal:1 belief:2 overlap:1 critical:3 natural:1 treated:1 mn:3 representing:2 technology:1 aston:1 picture:1 axis:1 carried:5 bare:1 prior:5 literature:2 acknowledgement:1 nishimori:2 interesting:2 h2:1 incurred:1 degree:2 basin:5 consistent:2 imposes:1 storing:1 pi:1 row:3 lo:1 elsewhere:1 dominicis:1 last:3 keeping:1 free:12 theo:1 enjoys:1 bias:3 institute:1 wide:1 taking:3 sparse:4 calculated:1 lett:3 world:1 transition:2 commonly:2 made:1 employing:2 constituting:1 obtains:3 ml:1 global:1 receiver:1 assumed:1 continuous:1 iterative:2 sk:1 channel:4 nature:1 symmetry:4 investigated:2 excellent:1 constructing:1 main:1 constituted:1 noise:15 site:1 fig:6 euro:1 tl:1 dlj:1 sub:5 comprises:1 decoded:1 sf:2 xl:2 breaking:2 vanish:2 saturate:3 inset:6 showing:1 symbol:2 dominates:1 mirror:1 ihigh:1 likely:2 sender:1 gallager:12 explore:1 saddle:8 expressed:1 saturating:2 corresponds:1 frustrated:2 formulated:1 content:1 vicente:5 telephone:1 typical:3 infinite:1 specifically:1 averaging:1 decouple:1 shannon:11 cond:1 quest:1 support:1 latter:1 relevance:1 constructive:1 dept:1 avoiding:2
767
1,702
Differentiating Functions of the Jacobian with Respect to the Weights Gary William Flake NEC Research Institute 4 Independence Way Princeton, NJ 08540 jiake@research.nj.nec.com Barak A. Pearlmutter Dept of Computer Science, FEC 313 University of New Mexico Albuquerque, NM 87131 bap@cs.unm.edu Abstract For many problems, the correct behavior of a model depends not only on its input-output mapping but also on properties of its Jacobian matrix, the matrix of partial derivatives of the model's outputs with respect to its inputs. We introduce the J-prop algorithm, an efficient general method for computing the exact partial derivatives of a variety of simple functions of the Jacobian of a model with respect to its free parameters. The algorithm applies to any parametrized feedforward model, including nonlinear regression, multilayer perceptrons, and radial basis function networks. 1 Introduction Let f (x, w) be an n input, m output, twice differentiable feedforward model parameterized by an input vector, x, and a weight vector w. Its Jacobian matrix is defined as J= [ ~ ()xl : aim a~" ~l aim = df(x, w) . dx ax" The algorithm we introduce can be used to optimize functions of the form aXI (1) or Ev(w) 2 1 = 211Jv - bll (2) where u, v, a, and b are user-defined constants. Our algorithm, which we call J-prop, can be used to calculate the exact value of both Eu / or Ev / in 0 (1) times the time required to calculate the normal gradient. Thus, I-prop is suitable for training models to have specific first derivatives, or for implementing several other well-known algorithms such as Double Backpropagation [1] and Tangent Prop [2]. a aw a aw Clearly, being able to optimize Equations 1 and 2 is useful; however, we suspect that the formalism which we use to derive our algorithm is actually more interesting because it allows us to modify J-prop to easily be applicable to a wide-variety of model types and G. W. Flake and B. A. Pear/mutter 436 objective functions. As such, we spend a fair portion of this paper describing the mathematical framework from which we later build J-prop. This paper is divided into four more sections. Section 2 contains background information and motivation for why optimizing the properties of the Jacobian is an important problem. Section 3 introduces our formalism and contains the derivation of the J-prop algorithm. Section 4 contains a brief numerical example of J-prop. And, finally, Section 5 describes further work and gives our conclusions. 2 Background and motivation Previous work concerning the modeling of an unknown function and its derivatives can be divided into works that are descriptive or prescriptive. Perhaps the best known descriptive result is due to White et al. [3,4], who show that given noise-free data, a multilayer perceptron (MLP) can approximate the higher derivatives of an unknown function in the limit as the number of training points goes to infinity. The difficulty with applying this result is the strong requirements on the amount and integrity of the training data; requirements which are rarely met in practice. This problem was specifically demonstrated by Principe, Rathie and Kuo [5] and Deco and Schiirmann [6], who showed that using noisy training data from chaotic systems can lead to models that are accurate in the input-output sense, but inaccurate in their estimates of quantities related to the Jacobian of the unknown system, such as the largest Lyapunov exponent and the correlation dimension. MLPs are particularly problematic because large weights can lead to saturation at a particular sigmoidal neuron which, in tum, results in extremely large first derivatives at the neuron when evaluated near the center of the sigmoid transition. Several methods to combat this type of over-fitting have been proposed. One of the earliest methods, weight decay [7], uses a penalty term on the magnitude of the weights. Weight decay is arguably optimal for models in which the output is linear in the weights because minimizing the magnitude of the weights is equivalent to minimizing the magnitude of the model's first derivatives. However, in the nonlinear case, weight decay can have suboptimal performance [1] because large (or small) weights do not always correspond to having large (or small) first derivatives. The Double Backpropagation algorithm [1] adds an additional penalty term to the error function equal to II E / 112. Training on this function results in a form of regularization that is in many ways an elegant combination of weight decay and training with noise: it is strictly analytic (unlike training with noise) but it explicitly penalizes large first derivatives ofthe model (unlike weight decay). Double Backpropagation can be seen as a special case of J-prop, the algorithm derived in this paper. a ax As to the general problem of coercing the first derivatives of a model to specific values, Simard, et at., [2] introduced the Tangent Prop algorithm, which was used to train MLPs for optical character recognition to be insensitive to small affine transformations in the character space. Tangent Prop can also be considered a special case of J-prop. 3 Derivation We now define a formalism under which J-prop can be easily derived. The method is very similar to a technique introduced by Pearlmutter [8] for calculating the product of the Hessian of an MLP and an arbitrary vector. However, where Pearlmutter used differential operators applied to a model's weight space, we use differential operators defined with respect to a model's input space. Our entire derivation is presented in five steps. First, we will define an auxiliary error Differentiating Functions of the Jacobian 437 function that has a few useful mathematical properties that simplify the derivation. Next, we will define a special differential operator that can be applied to both the auxiliary error function, and its gradient with respect to the weights. We will then see that the result of applying the differential operator to the gradient of the auxiliary error function is equivalent to analytically calculating the derivatives required to optimize Equations 1 and 2. We then show an example of the technique applied to an MLP. Finally, in the last step, the complete algorithm is presented. To avoid confusion, when referring to generic data-driven models, the model will always be expressed as a vector function y = f (x, w), where x refers to the model input and w refers to a vector of all of the tunable parameters of the model. In this way, we can talk about models while ignoring the mechanics of how the models work internally. Complementary to the generic vector notation, the notation for an MLP uses only scalar symbols; however, these symbols must refer to internal variable of the model (e.g., neuron thresholds, net inputs, weights, etc.), which can lead to some ambiguity. To be clear, when using vector notation, the input and output of an MLP will always be denoted by x and y, respectively, and the collection of all of the weights (including biases) map to the vector w. However, when using scalar arithmetic, the scalar notation for MLPs will apply. 3.1 Auxiliary error function Our auxiliary error function, E, is defined as E(x, w) = u T f(x, w). (3) Note that we never actually optimize with respect E; we define it only because it has the property that aE/ax = u T J, which will be useful to the derivation shortly. Note that a E / ax appears in the Taylor expansion of E about a point in input space: -T E(x + Ax, w) = E(x, w) + ~! Ax + 0 (1IAXI12) . (4) Thus, while holding the weights, w, fixed and letting Ax be a perturbation of the input, x, Equation 4 characterizes how small changes in the input of the model change the value of the auxiliary error function. Be setting Ax = rv, with v being an arbitrary vector and r being a small value, we can rearrange Equation 4 into the form: ~ [E(x+rv,w) -E(x,w)] +O(r) lim -1 [E(x r~O r a- + rv, w) aE(x + rv,w) r I -] - E(x, w) . (5) r=O This final expression will allow us to define the differential operator in the next subsection. 3.2 Differential operator Let h(x, w) be an arbitrary twice differentiable function. We define the differentiable operator (6) G. W Flake and B. A. Pearlmutter 438 which has the property that Rv{E(x, w)} = u T Jv. Being a differential operator, Rv{-} obeys all of the standard rules for differentiation: Rv{c} Rv{ c? h(x, w)} Rv{h(x,w) + g(x,w)} Rv{h(x, w) . g(x, w)} Rv{h(g(x, w), w)} o = Rv{!h(X,W) } c? Rv{h(x, w)} Rv{ h(x, w)} + Rv{g(x, w)} Rv{h(x,w)}? g(x,w) + h(x,w)? Rv{g(x,w)} h'(g(x,w))? Rv{g(x,w)} d dt Rv{h(x, w)} The operator also yields the identity Rv{ x} = v. 3.3 Equivalence We will now see that the result of calculating Rv{ a E / aw} can be used to calculate both aEu/aw and aEv/aw. Note that Equations 3-5 all assume that both u and v are independent of x and w. To calculate aEu/aw and aEv/aw, we will actually set u or v to a value that depends on both x and w; however, the derivation still works because our choices are explicitly made in such a way that the chain rule of differentiation is not supposed to be applied to these terms. Hence, the correct analytical solution is obtained despite the dependence. To optimize with respect to Equation 1, we use: II a 1 J T u - a 112 = (au T J) T (J T u - a) aw'2 ~ = Rv { aE } ' aw (7) with v = (JT U - a). To optimize with respect to Equation 2, we use: ~! IIJv aw2 bll 2 = (Jv _ b)T (aJv) aw = Rv{ aE} , aw (8) with u = (Jv - b). 3.4 Method applied to MLPs We are now ready to see how this technique can be applied to a specific type of model. Consider an MLP with L + 1 layers of nodes defined by the equations: y~ = g(x~) N/ Xl t L 1-1 I Yj Wij - (9) rii' (10) j In these equations, superscripts denote the layer number (starting at 0), subscripts index over terms in a particular layer, and NI is the number of input nodes in layer l . Thus, y~ is the output of neuron i at node layer l, and xi is the net input coming into the same neuron. Moreover, yf is an output of the entire MLP while y? is an input going into the MLP. The feedback equations calculated with respect to E are: aE ayf (11) 439 Differentiating Functions of the Jacobian 8E (12) 8y~ 8E (13) 8x't 8E (14) 8w!j 8E 8E 8()! 8x lt J (15) ' where the Ui term is a component in the vector u from Equation 1. Applying the R v {?} operator to the feedforward equations yields: (16) Rv{Y?} Rv{yD g'(x~)Rv{ xD (for 1 > 0) (17) N/ L Rv{y~-l } W~j' Rv{ x~} (18) j where the Vi term is a component in the vector v from Equation 2. As the final step, we apply the R v {?} operator to the feedback equations, which yields: Rv{:~ } (19) Rv{ :~} (20) Rv{ :~} (21) :!,} (22) Rv{ ::;} (23) Rv{ 3.5 o Complete algorithm Implementing this algorithm is nearly as simple as implementing normal gradient descent. For each type of variable that is used in an MLP (net input, neuron output, weights, thresholds, partial derivatives, etc.), we require that an extra variable be allocated to hold the result of applying the R v {?} operator to the original variable. With this change in place, the complete algorithm to compute 8Eu /8w is as follows : ? Set u and a to the user specified vectors from Equation 1. ? Set the MLP inputs to the value of x that J is to be evaluated at. ? Perform a normal feedforward pass using Equations 9 and 10. ? Set 8E/8yf to Ui. G. W. Flake and B. A. Pearlmutter 440 (a) (b) Figure 1: Learning only the derivative: showing (a) poor approximation of the function with (b) excellent approximation of the derivative. ? Perform the feedback pass with Equations 11-15. Note that values in the aEjay? terms are now equal to JT U. ? Set v to (JT u - a) ? Perform a Rv{ .} forward pass with Equations 16-18. ? Set the Rv{ 8Ej8yf} terms to O. ? Perform a Rv{?} backward pass with Equations 19-23. After the last step, the values in the Rv{ 8E j 8w!j} and Rv{ 8 E j aeD terms contain the required result. It is important to note that the time complexity of the "J?forward" and "J. backward" calculations are nearly identical to the typical output and gradient evaluations (i.e., the "forward" and "backward" passes) of the models used. A similar technique can be used for calculating 8Evj8w. The main difference is that the Rv{ .} forward pass is performed between the normal forward and backward passes because u can only be determined after the Rv{ f (z, w)} has been calculated. 4 Experimental results To demonstrate the effectiveness and generality of the J-prop algorithm, we have implemented it on top of an existing neural network library [9] in such a way that the algorithm can be used on a large number of architectures, including MLPs, radial basis function net? works, and higher order networks. We trained an MLP with ten hidden tanh nodes on 100 points with conjugate gradient. The training exemplars consisted of inputs in [-1, 1] and a target derivative from 3 cos( 3x) + 5cos(lOx). Our unknown function (which the MLP never sees data from) is sin(3x) + sin(lOx). The model quickly converges to a solution in approximately 100 iterations. l Figure 1 shows the performance of the MLP. Having never seen data from the unknown function, the MLP yields a poor approximation of the function, but a very accurate approximation of the function's derivative. We could have trained on both outputs and derivatives, but our goal was to illustrate that J?prop can target derivatives alone. Differentiating Functions of the Jacobian 5 441 Conclusions We have introduced a general method for calculating the weight gradient of functions of the Jacobian matrix of feedforward nonlinear systems. The method can be easily applied to most nonlinear models in common use today. The resulting algorithm, J-prop, can be easily modified to minimize functionals from several application domains [10]. Some possible uses include: targeting known first derivatives, implementing Tangent Prop and Double Backpropagation, enforcing identical VO sensitivities in auto-encoders, deflating the largest eigenvalue and minimizing all eigenvalue bounds, optimizing the determinant for blind source separation, and building nonlinear controllers. While some special cases of the J-prop algorithm have already been studied, a great deal is unknown about how optimization of the Jacobian changes the overall optimization problem. Some anecdotal evidence seems to imply that optimization of the Jacobian can lead to better generalization and faster training. It remains to be seen if J-prop used on a nonlinear extension of linear methods will lead to superior solutions. Acknowledgements We thank Frans Coetzee, Yannis Kevrekidis, Joe O'Ruanaidh, Lucas Parra, Scott Rickard, Justinian Rosca, and Patrice Simard for helpful discussions. GWF would also like to thank Eric Baum and the NEC Research Institute for funding the time to write up these results. References [1] H. Drucker and Y. Le Cun. Improving generalization performance using double backpropagation. IEEE Transactions on Neural Networks, 3(6), November 1992. [2] P. Simard, B. Victorri, Y. Le Cun, and J. Denker. Tangent prop-A formalism for specifying selected invariances in an adaptive network. In John E. Moody, Steve J. Hanson, and Richard P. Lippmann, editors, Advances in Neural Information Processing Systems, volume 4, pages 895-903. Morgan Kaufmann Publishers, Inc., 1992. [3] H. White and A. R. Gallant. On learning the derivatives of an unknown mapping with multilayer feedforward networks. In Halbert White, editor, Artificial Neural Networks, chapter 12, pages 206-223. Blackwell, Cambridge, Mass., 1992. [4] H. White, K. Hornik, and M. Stinchcombe. Universal approximation of an unknown mapping and its derivative. In Halbert White, editor, Artificial Neural Networks, chapter 6, pages 55-77. Blackwell, Cambridge, Mass., 1992. [5] J. Principe, A. Rathie, and J. Kuo. Prediction of chaotic time series with neural networks and the issues of dynamic modeling. Bifurcations and Chaos, 2(4), 1992. [6] G. Deco and B. Schiirmann. Dynamic modeling of chaotic time series. In Russell Greiner, Thomas Petsche, and Stephen Jose Hanson, editors, Computational Learning Theory and Natural Learning Systems, volume IV of Making Learning Systems Practical, chapter 9, pages 137-153. The MIT Press, Cambridge, Mass., 1997. [7] G. E. Hinton. Learning distributed representations of concepts. In Proc. Eigth Annual Con! Cognitive Science Society, pages 1-12, Hillsdale, NJ, 1986. Erlbaum. [8] Barak A. Pearlmutter. Fast exact multiplication by the Hessian. Neural Computation, 6(1):147-160,1994. [9] G. W. Flake. Industrial strength modeling tools. Submitted to NIPS 99, 1999. [10] G. W. Flake and B. A. Pearl mutter. Optimizing properties of the Jacobian of nonlinear feedforward systems. In preperation, 1999.
1702 |@word ruanaidh:1 determinant:1 seems:1 series:2 contains:3 prescriptive:1 existing:1 com:1 dx:1 must:1 john:1 numerical:1 analytic:1 alone:1 selected:1 node:4 sigmoidal:1 five:1 mathematical:2 differential:7 fitting:1 frans:1 introduce:2 behavior:1 mechanic:1 notation:4 moreover:1 kevrekidis:1 mass:3 transformation:1 differentiation:2 nj:3 combat:1 xd:1 internally:1 arguably:1 modify:1 limit:1 despite:1 subscript:1 yd:1 approximately:1 twice:2 au:1 studied:1 equivalence:1 specifying:1 co:2 obeys:1 practical:1 yj:1 practice:1 backpropagation:5 chaotic:3 universal:1 radial:2 refers:2 targeting:1 operator:12 applying:4 optimize:6 equivalent:2 map:1 demonstrated:1 center:1 baum:1 go:1 starting:1 rule:2 target:2 today:1 user:2 exact:3 us:3 recognition:1 particularly:1 calculate:4 eu:2 russell:1 ui:2 complexity:1 dynamic:2 trained:2 eric:1 basis:2 easily:4 chapter:3 talk:1 derivation:6 train:1 fast:1 artificial:2 spend:1 noisy:1 final:2 superscript:1 patrice:1 descriptive:2 differentiable:3 eigenvalue:2 net:4 analytical:1 product:1 coming:1 deflating:1 supposed:1 double:5 requirement:2 converges:1 derive:1 illustrate:1 exemplar:1 strong:1 auxiliary:6 c:1 implemented:1 met:1 lyapunov:1 correct:2 implementing:4 hillsdale:1 require:1 coercing:1 generalization:2 parra:1 strictly:1 extension:1 hold:1 considered:1 normal:4 great:1 mapping:3 proc:1 applicable:1 tanh:1 largest:2 tool:1 anecdotal:1 mit:1 clearly:1 always:3 aim:2 modified:1 avoid:1 earliest:1 ax:8 derived:2 pear:1 industrial:1 sense:1 helpful:1 inaccurate:1 entire:2 hidden:1 wij:1 going:1 overall:1 issue:1 denoted:1 exponent:1 lucas:1 special:4 bifurcation:1 equal:2 never:3 having:2 identical:2 nearly:2 simplify:1 richard:1 few:1 william:1 mlp:14 evaluation:1 introduces:1 rearrange:1 aeu:2 chain:1 accurate:2 partial:3 iv:1 taylor:1 penalizes:1 halbert:2 formalism:4 modeling:4 erlbaum:1 encoders:1 aw:11 referring:1 sensitivity:1 quickly:1 moody:1 deco:2 nm:1 ambiguity:1 aed:1 cognitive:1 derivative:21 simard:3 inc:1 explicitly:2 depends:2 vi:1 blind:1 later:1 performed:1 characterizes:1 portion:1 mlps:5 minimize:1 ni:1 kaufmann:1 who:2 correspond:1 ofthe:1 yield:4 albuquerque:1 submitted:1 coetzee:1 con:1 tunable:1 lim:1 subsection:1 actually:3 appears:1 tum:1 higher:2 dt:1 steve:1 mutter:2 evaluated:2 generality:1 correlation:1 nonlinear:7 yf:2 perhaps:1 building:1 contain:1 consisted:1 concept:1 regularization:1 analytically:1 hence:1 white:5 deal:1 sin:2 complete:3 demonstrate:1 confusion:1 pearlmutter:6 vo:1 chaos:1 funding:1 sigmoid:1 common:1 superior:1 insensitive:1 volume:2 fec:1 aw2:1 refer:1 cambridge:3 etc:2 add:1 integrity:1 showed:1 optimizing:3 driven:1 seen:3 morgan:1 additional:1 arithmetic:1 stephen:1 ii:2 rv:40 faster:1 calculation:1 divided:2 concerning:1 prediction:1 regression:1 multilayer:3 ae:5 controller:1 df:1 iteration:1 background:2 aev:2 victorri:1 source:1 allocated:1 publisher:1 extra:1 unlike:2 pass:2 suspect:1 elegant:1 effectiveness:1 call:1 near:1 feedforward:7 variety:2 independence:1 architecture:1 suboptimal:1 drucker:1 expression:1 penalty:2 hessian:2 useful:3 clear:1 amount:1 ten:1 problematic:1 write:1 four:1 threshold:2 rosca:1 jv:4 backward:4 jose:1 parameterized:1 place:1 separation:1 layer:5 bound:1 annual:1 preperation:1 strength:1 infinity:1 extremely:1 optical:1 combination:1 poor:2 flake:6 conjugate:1 describes:1 bap:1 character:2 cun:2 ayf:1 making:1 equation:19 remains:1 describing:1 letting:1 rathie:2 apply:2 denker:1 generic:2 petsche:1 shortly:1 original:1 thomas:1 top:1 include:1 calculating:5 build:1 society:1 objective:1 already:1 quantity:1 dependence:1 gradient:7 thank:2 parametrized:1 enforcing:1 index:1 minimizing:3 mexico:1 holding:1 rii:1 unknown:8 perform:4 gallant:1 neuron:6 descent:1 november:1 hinton:1 bll:2 perturbation:1 arbitrary:3 introduced:3 required:3 specified:1 blackwell:2 hanson:2 pearl:1 nip:1 able:1 ev:2 scott:1 saturation:1 including:3 stinchcombe:1 suitable:1 difficulty:1 natural:1 brief:1 library:1 imply:1 ready:1 eigth:1 lox:2 auto:1 acknowledgement:1 tangent:5 multiplication:1 interesting:1 affine:1 editor:4 last:2 free:2 bias:1 allow:1 barak:2 institute:2 wide:1 perceptron:1 differentiating:4 distributed:1 feedback:3 axi:1 dimension:1 transition:1 calculated:2 forward:5 collection:1 made:1 adaptive:1 transaction:1 unm:1 functionals:1 approximate:1 lippmann:1 xi:1 why:1 ignoring:1 hornik:1 improving:1 expansion:1 excellent:1 domain:1 main:1 motivation:2 noise:3 fair:1 complementary:1 xl:2 jacobian:13 yannis:1 specific:3 jt:3 showing:1 symbol:2 decay:5 evidence:1 joe:1 rickard:1 nec:3 magnitude:3 lt:1 greiner:1 expressed:1 scalar:3 applies:1 gary:1 prop:20 identity:1 goal:1 change:4 specifically:1 typical:1 determined:1 kuo:2 pas:5 experimental:1 invariance:1 perceptrons:1 rarely:1 principe:2 internal:1 dept:1 princeton:1
768
1,703
Distributed Synchrony of Spiking Neurons in a Hebbian Cell Assembly David Horn Nir Levy School of Physics and Astronomy, Raymond and Beverly Sackler Faculty of Exact Sciences, Tel Aviv University, Tel Aviv 69978, Israel horn~neuron.tau.ac.il nirlevy~post.tau.ac.il Isaac Meilijson Eytan Ruppin School of Mathematical Sciences, Raymond and Beverly Sackler Faculty of Exact Sciences, Tel Aviv University, Tel Aviv 69978, Israel isaco~math.tau.ac.il ruppin~math.tau.ac.il Abstract We investigate the behavior of a Hebbian cell assembly of spiking neurons formed via a temporal synaptic learning curve. This learning function is based on recent experimental findings . It includes potentiation for short time delays between pre- and post-synaptic neuronal spiking, and depression for spiking events occuring in the reverse order. The coupling between the dynamics of the synaptic learning and of the neuronal activation leads to interesting results. We find that the cell assembly can fire asynchronously, but may also function in complete synchrony, or in distributed synchrony. The latter implies spontaneous division of the Hebbian cell assembly into groups of cells that fire in a cyclic manner. We invetigate the behavior of distributed synchrony both by simulations and by analytic calculations of the resulting synaptic distributions. 1 Introduction The Hebbian paradigm that serves as the basis for models of associative memory is often conceived as the statement that a group of excitatory neurons (the Hebbian cell assembly) that are coupled synaptically to one another fire together when a subset of the group is being excited by an external input. Yet the details of the temporal spiking patterns of neurons in such an assembly are still ill understood. Theoretically it seems quite obvious that there are two general types of behavior: synchronous neuronal firing, and asynchrony where no temporal order exists in the assembly and the different neurons fire randomly but with the same overall rate. Further subclassifications were recently suggested by [BruneI, 1999]. Experimentally this question is far from being settled because evidence for the associative 130 D. Hom, N. Levy, 1. Meilijson and E. Ruppin memory paradigm is quite scarce. On one hand, one possible realization of associative memories in the brain was demonstrated by [Miyashita, 1988] in the inferotemporal cortex. This area was recently reinvestigated by [Yakovlev et al., 1998] who compared their experimental results with a model of asynchronized spiking neurons. On the other hand there exists experimental evidence [Abeles, 1982] for temporal activity patterns in the frontal cortex that Abeles called synfire-chains. Could they correspond to an alternative type of synchronous realization of a memory attractor? To answer these questions and study the possible realizations of attractors in cortical-like networks we investigate the temporal structure of an attractor assuming the existence of a synaptic learning curve that is continuously applied to the memory system. This learning curve is motivated by the experimental observations of [Markram et al., 1997, Zhang et al., 1998] that synaptic potentiation or depression occurs within a critical time window in which both pre- and post-synaptic neurons have to fire. If the pre-synaptic neuron fires first within 30ms or so, potentiation will take place. Depression is the rule for the reverse order. The regulatory effects of such a synaptic learning curve on the synapses of a single neuron that is subjected to external inputs were investigated by [Abbott and Song, 1999] and by [Kempter et al., 1999]. We investigate here the effect of such a rule within an assembly of neurons that are all excited by the same external input throughout a training period, and are allowed to influence one another through their resulting sustained activity. 2 The Model We study a network composed of N E excitatory and NJ inhibitory integrate-and-fire neurons. Each neuron in the network is described by its subthreshold membrane potential Vi{t) obeying . Vi{t) 1 =- - Vi{t) + R1i(t) (1) Tn where Tn is the neuronal integration time constant. A spike is generated when Vi{t) reaches the threshold Vrest + fJ, upon which a refractory period of TRP is set on and the membrane potential is reset to Vreset where Vrest < Vreset < Vrest + fJ. Ii{t) is the sum of recurrent and external synaptic current inputs. The net synaptic input charging the membrane of excitatory neuron i at time t is R1i(t) = L J~E{t) L 0 (t - t~ - I j Td) - L j Ji~J L 0 (t - tj - Td) +r xt (2) m summing over the different synapses of j = 1, ... , NE excitatory neurons and of j = 1, ... ,NJ inhibitory neurons, with postsynaptic efficacies J~E{t) and Ji~J respectively. The sum over 1 (m) represents a sum on different spikes arriving at synapse j, at times t t; + Td (t tj + Td), where t~ (tj) is the emission time of the l-th (m-th) spike from the excitatory (inhibitory) neuron j and Td is the synap- = = tic delay. Iext, the external current, is assumed to be random and independent at each neuron and each time step, drawn from a Poisson distribution with mean Aext. Analogously, the synaptic input to the inhibitory neuron i at time tis j j m We assume full connectivity among the excitatory neurons, but only partial connectivity between all other three types of possible connnections, with connection Distributed Synchrony of Spiking Neurons in a Hebbian Cell Assembly 131 probabilities denoted by eEl, e l E and C l I. In the following we will report simulation results in which the synaptic delays Td were assigned to each synapse, or pair of neurons, randomly, chosen from some finite set of values. Our analytic calculation will be done for one fixed value of this delay parameter. The synaptic efficacies between excitatory neurons are assumed to be potentiated or depressed according to the firing patterns of the pre- and post-synaptic neurons. In addition we allow for a uniform synaptic decay. Thus each excitatory synapse obeys (4) where the synaptic decay constant Ts is assumed to be very large compared to the membrane time constant Tn. J/JE(t) are constrained to vary in the range [0, Jma:~ ]. The change in synaptic efficacy is defined by Fij (t), as Fij(t) = L [6(t - t:)Kp(t; - t:) + 6(t - t;)KD(t; - t:)] (5) k ,l where Kp and KD are the potentiation and depression branches of the kernel function K(6) = -cO exp [- (a6 + b)2] (6) plotted in Figure 1. Following [Zhang et al., 1998] we distinguish between the situation where the postsynaptic spike, at t~, appears after or before the presynaptic spike, at t~, using the asymmetric kernel that captures the essence of their experimental observations. .. -o !.'--~-~~--' o-~-~~-----' k /I =t'_t , I Figure 1: The kernel function whose left part, Kp, leads to potentiation of the synapse, and whose right branch, KD, causes synaptic depression. 3 Distributed Synchrony of a Hebbian Assembly We have run our system with synaptic delays chosen randomly to be either 1, 2, or 3ms, and temporal parameters Tn chosen as 40ms for excitatory neurons and 20ms for inhibitory ones. Turning external input currents off after a while we obtained sustained firing activities in the range of 100-150 Hz. We have found, in addition to synchronous and asynchronous realizations of this attractor, a mode of distributed synchrony. A characteristic example of a long cycle is shown in Figure 2: The 100 excitatory neurons split into groups such that each group fires at the same frequency and at a fixed phase difference from any other group. The J/JE synaptic efficacies D. Horn, N Levy, 1. Meilijson and E. Ruppin 132 ':~ :tI ':f ': f j j r I : I: : I : r : I : r r :I :1 I :I I :I : I ] :1 I :1 : 1 1 :1 : I :1 , I : " : I : "I 1 : 1: I ": " r n ~ JO Figure 2: Distributed synchronized firing mode. The firing patterns of six cell assemblies of excitatory neurons are displayed vs time (in ms). These six groups of neurons formed in a self-organized manner for a kernel function with equal potentiation and depression. The delays were chosen randomly from three values, 1 2 or 3ms, and the system is monitored every 0.5ms . are initiated as small random values. The learning process leads to the self-organized synaptic matrix displayed in Figure 3(a). The block form of this matrix represents the ordered couplings that are responsible for the fact that each coherent group of neurons feeds the activity of groups that follow it. The self-organized groups form spontaneously. When the synapses are affected by some external noise, as can come about from Hebbian learning in which these neurons are being coupled with other pools of neurons, the groups will change and regroup, as seen in Figure 3(b) and 3(c). (a) (b) (c) Figure 3: A synaptic matrix for n = 6 distributed synchrony. The synaptic matrix between the 100 excitatory neurons of our system is displayed in a grey-level code with black meaning zero efficacy and white standing for the synaptic upper-bound. (a) The matrix that exists during the distributed synchronous mode of Figure 2. Its basis is ordered such that neurons that fire together are grouped together. (b) Using the same basis as in (a) a new synaptic matrix is shown, one that is formed after stopping the sustained activity of Figure 2, introducing noise in the synaptic matrix, and reinstituting the original memory training. (c) The same matrix as (b) is shown in a new basis that exhibits connections that lead to a new and different realization of distributed synchrony. A stable distributed synchrony cycle can be simply understood for the case of a single synaptic delay setting the basic step, or phase difference , of the cycle. When several delay parameters exist, a situation that probably more accurately represents the a-function character of synaptic transmission in cortical networks, distributed Distributed Synchrony of Spiking Neurons in a Hebbian Cell Assembly 133 synchrony may still be obtained, as is evident from Figure 2. After some time the cycle may destabilize and regrouping may occur by itself, without external interference. The likelihood of this scenario is increased because different synaptic connections that have different delays can interfere with one another. Nonetheless, over time scales of the type shown in Figure 2, grouping is stable. 4 Analysis of a Cycle In this section we analyze the dynamics of the network when it is in a stable state of distributed synchrony. We assume that n groups of neurons are formed and calculate the stationary distribution of JffE(t) . In this state the firing pattern of every two neurons in the network can be characterized by their frequency l/(t) and by their relative phase 8. We assume that 8 is a random normal variable with mean J.Lo and standard deviation 0'0 . Thus, Eq. 4 can be rewritten as the following stochastic differential equation dJi~E(t) = [J.LFij(t) - :s J!jE(t)] dt+O'Fij(t)dW(t) (7) where Fij (t) (Eq. 5) is represented here by a drift term J.LFij (t) and a diffusion term O'Fij (t) which are its mean and standard deviation. W(t) describes a Wiener process. Note that both J.LFij (t) and O'Fij (t) are calculated for a specific distribution of 8 and are functions of J.Lo and 0'0. The stochastic process that satisfies Eq. 7 will satisfy the Fokker-Plank equation for the probability distribution f of JIfE, 2 8 [(J.LF.] (t) _ _1) JPlE f(JPlE t) ] 0'2Fij (t) 8 f(JPlE + 8JEE2 8f(JPlE t) = ___ .J' 8t 8JPlE 1J .. T'J tJ' 2 'J' t) (8) ij S with reflecting boundary conditions imposed by the synaptic bounds, 0 and Jmax . Since we are interested in the stable state of the process we solve the stationary equation. The resulting density function is [1 (2J.LFij J EE 1J EE2) 1 EE N f(Jij ,J.Lt5, 0'15) = O'}ij (t) exp O'}ij (t) ij - Ts ij (9) where (10) Eq. 9 enables us to calculate the stationary distribution of the synaptic efficacies between the presynaptic neuron i and the post-synaptic neuron j given their frequency l/ and the parameters J.Lo and 0'15. An example of a solution for a 3-cycle is shown in Figure 4. In this case all neurons fire with frequency l/ = (3Td)-1 and J.Lt5 takes one of the values -Td, 0, Td. Simulation results of a 3-cycle in a network of excitatory and inhibitory integrateand-fire neurons described in Section 2 are given in Figure 5. As can be seen the results obtained from the analysis match those observed in the simulation. 5 Discussion The interesting experimental observations of synaptic learning curves [Markram et al., 1997, Zhang et al., 1998] have led us to study their implications for the firing patterns of a Hebbian cell assembly. We find that, in addition D. Horn, N. Levy, 1. Meilijson and E. Ruppin 134 (a) (b) 70 60 so 40 30 20 ) ,0 01 0 .2 0.3 0 04 0 0.' 0.3 0.2 0.4 0. 5 J EE " Figure 4: Results of the analysis for n = 3, a6 = 2ms and Td = 2.5ms. (a) The synaptic matrix. Each of the nine blocks symbolizes a group of connections between neurons that have a common phase-lag J..l6 . The mean of Ji~E was calculated for each cell by Eq. 9 and its value is given by the gray scale tone. (b) The distribution of synaptic values between all excitatory neurons. (a) (b) 5o0 0 , - - - - - - - - - - - - - , 4500 4000 3500 3000 2500 2000 1500 500 0. ' 0 .2 0.3 0.4 0. 5 o 0" - - 0 ..'..- - '... 0.2:--~0.3::----::"" 0.4c--'"-:" 0. 5 , JEE Figure 5: Simulation results for a network of N E = 100 and NJ = 50 integrateand-fire neurons, when the network is in a stable n = 3 state. Tn = 10ms for both excitatory and inhibitory neurons. The average frequency of the neurons is 130 Hz. (a) The excitatory synaptic matrix. (b) Histogram of the synaptic efficacies. to the expected synchronous and asynchronous modes, an interesting behavior of distributed synchrony can emerge. This is the phenomenon that we have investigated both by simulations and by analytic evaluation. Distributed synchrony is a mode in which the Hebbian cell assembly breaks into an n-cycle. This cycle is formed by instantaneous symmetry breaking, hence specific classification of neurons into one of the n groups depends on initial conditions , noise, etc. Thus the different groups of a single cycle do not have a semantic invariant meaning of their own. It seems perhaps premature to try and identify these cycles with synfire chains [Abeles, 1982] that show recurrence of firing patterns of groups of neurons with periods of hundreds of ms. Note however , that if we make such an identification , it is a different explanation from the model of [Herrmann et al., 1995J , which realizes the synfire chain by combining sets of preexisting patterns into a cycle. The simulations in Figures 2 and 3 were carried out with a learning curve that possessed equal potentiation and depression branches , i.e. was completely antisymmetric in its argument. In that case no synaptic decay was allowed . Figure 5, on the other hand, had stronger potentiation than depression, and a finite synaptic Distributed Synchrony ofSpiking Neurons in a Hebbian Cell Assembly 135 decay time was assumed. Other conditions in these nets were different too, yet both had a window of parameters where distributed synchrony showed up . Using the analytic approach of section 4 we can derive the probability distribution of synaptic values once a definite cyclic pattern of distributed synchrony is formed. An analytic solution of the combined dynamics of both the synapses and the spiking neurons is still an open challenge. Hence we have to rely on the simulations to prove that distributed synchrony is a natural spatiotemporal behavior that follows from combined neuronal dynamics and synaptic learning as outlined in section 2. To the extent that both types of dynamics reflect correctly the dynamics of cortical neural networks, we may expect distributed synchrony to be a mode in which neuronal attractors are being realized. The mode of distrbuted synchrony is of special significance to the field of neural computation since it forms a bridge between the feedback and feed-forward paradigms. Note that whereas the attractor that is formed by the Hebbian cell assembly is of global feedback nature, i.e. one may regard all neurons of the assembly as being connected to other neurons within the same assembly, the emerging structure of distributed synchrony shows that it breaks down into groups. These groups are connected to one another in a self-organized feed-forward manner, thus forming the cyclic behavior we have observed. References [Abbott and Song, 1999] L. F. Abbott and S. Song. Temporally asymmetric hebbian learning, spike timing and neuronal response variability. In M. S. Kearns, S. A. Solla, and D. A. Cohn, editors, Advances in Neural Information Processing Systems 11: Proceedings of the 1998 Conference, pages 69 - 75. MIT Press, 1999. [Abeles, 1982] M. Abeles. Local Cortical Circuits. Springer, Berlin, 1982. [BruneI, 1999] N. BruneI. Dynamics of sparsely connected networks of excitatory and inhibitory spiking neurons. Journal of Computational Neuroscience, 1999. [Herrmann et al., 1995] M. Herrmann, J . Hertz, and A. Prugel-Bennet. Analysis of synfire chains. Network: Compo in Neural Systems, 6:403 - 414, 1995. [Kempter et al. , 1999] R. Kempter, W. Gerstner, and J . Leo van Hemmen. Spikebased compared to rate-based hebbian learning. In M. S. Kearns , S. A. Solla, and D. A. Cohn, editors, Advances in Neural Information Processing Systems 11: Proceedings of the 1998 Conference, pages 125 - 131. MIT Press, 1999. [Markram et al., 1997] H. Markram, J. Lubke, M. Frotscher, and B. Sakmann. Regulation of synaptic efficacy by coincidence of postsynaptic aps and epsps. Science, 275(5297):213 - 215 , 1997. [Miyashita, 1988] Y. Miyashita. Neuronal correlate of visual associative long-term memory in the primate temporal cortex. Nature, 335:817 - 820, 1988. [Yakovlev et al., 1998] V. Yakovlev, S. Fusi, E . Berman, and E . Zohary. Inter-trial neuronal activity in inferior temporal cortex: a putative vehicle to generate longterm visual associations. Nature Neurosc ., 1(4) :310 - 317, 1998. [Zhang et al., 1998] L. I. Zhang, H. W. Tao, C. E. Holt, W . A. Harris, and M. Poo. A critical window for cooperation and competition among developing retinotectal synapses. Nature, 395:37 - 44, 1998.
1703 |@word trial:1 longterm:1 faculty:2 seems:2 stronger:1 open:1 grey:1 simulation:8 excited:2 initial:1 cyclic:3 efficacy:8 current:3 activation:1 yet:2 analytic:5 enables:1 aps:1 v:1 stationary:3 tone:1 short:1 compo:1 math:2 zhang:5 mathematical:1 differential:1 prove:1 sustained:3 manner:3 theoretically:1 inter:1 expected:1 behavior:6 brain:1 td:10 window:3 zohary:1 circuit:1 israel:2 tic:1 emerging:1 astronomy:1 finding:1 nj:3 temporal:8 every:2 ti:2 before:1 understood:2 timing:1 local:1 initiated:1 firing:8 black:1 co:1 bennet:1 range:2 regroup:1 obeys:1 horn:4 responsible:1 spontaneously:1 block:2 definite:1 lf:1 area:1 pre:4 holt:1 spikebased:1 influence:1 imposed:1 demonstrated:1 poo:1 rule:2 dw:1 jmax:1 spontaneous:1 exact:2 asymmetric:2 sparsely:1 observed:2 coincidence:1 capture:1 calculate:2 cycle:12 connected:3 solla:2 dynamic:7 upon:1 division:1 basis:4 completely:1 represented:1 leo:1 preexisting:1 kp:3 quite:2 whose:2 lag:1 solve:1 itself:1 asynchronously:1 associative:4 net:2 jij:1 reset:1 combining:1 realization:5 competition:1 transmission:1 coupling:2 recurrent:1 ac:4 derive:1 ij:5 school:2 eq:5 epsps:1 implies:1 come:1 synchronized:1 berman:1 vrest:3 fij:7 stochastic:2 potentiation:8 integrateand:2 normal:1 exp:2 vary:1 realizes:1 bridge:1 grouped:1 mit:2 emission:1 likelihood:1 stopping:1 interested:1 tao:1 overall:1 among:2 ill:1 plank:1 denoted:1 classification:1 constrained:1 integration:1 special:1 frotscher:1 equal:2 once:1 field:1 represents:3 report:1 randomly:4 composed:1 phase:4 fire:12 attractor:6 brunei:3 investigate:3 evaluation:1 tj:4 chain:4 implication:1 partial:1 plotted:1 increased:1 a6:2 introducing:1 deviation:2 subset:1 uniform:1 hundred:1 delay:9 too:1 answer:1 spatiotemporal:1 abele:5 combined:2 density:1 standing:1 physic:1 eel:1 off:1 pool:1 together:3 continuously:1 analogously:1 connectivity:2 jo:1 reflect:1 settled:1 external:8 potential:2 includes:1 satisfy:1 vi:4 depends:1 vehicle:1 break:2 try:1 analyze:1 meilijson:4 synchrony:22 il:4 formed:7 lubke:1 wiener:1 beverly:2 who:1 characteristic:1 correspond:1 subthreshold:1 symbolizes:1 identify:1 identification:1 accurately:1 synapsis:5 reach:1 synaptic:43 nonetheless:1 frequency:5 isaac:1 obvious:1 monitored:1 organized:4 reflecting:1 appears:1 feed:3 dt:1 follow:1 response:1 synapse:4 done:1 hand:3 synfire:4 cohn:2 prugel:1 interfere:1 mode:7 perhaps:1 asynchrony:1 gray:1 aviv:4 effect:2 hence:2 assigned:1 semantic:1 white:1 during:1 self:4 recurrence:1 inferior:1 essence:1 m:11 evident:1 occuring:1 complete:1 tn:5 fj:2 meaning:2 ruppin:5 instantaneous:1 recently:2 common:1 spiking:10 ji:3 dji:1 refractory:1 association:1 r1i:2 outlined:1 depressed:1 had:2 stable:5 cortex:4 etc:1 inferotemporal:1 own:1 recent:1 showed:1 reverse:2 scenario:1 regrouping:1 seen:2 lfij:4 paradigm:3 period:3 ii:1 branch:3 full:1 hebbian:15 match:1 characterized:1 calculation:2 long:2 post:5 basic:1 poisson:1 histogram:1 kernel:4 synaptically:1 cell:14 isaco:1 addition:3 whereas:1 probably:1 hz:2 ee:3 split:1 synchronous:5 motivated:1 six:2 o0:1 song:3 cause:1 nine:1 depression:8 generate:1 exist:1 inhibitory:8 neuroscience:1 conceived:1 correctly:1 affected:1 group:18 threshold:1 destabilize:1 drawn:1 abbott:3 diffusion:1 sum:3 run:1 place:1 throughout:1 putative:1 fusi:1 bound:2 hom:1 sackler:2 distinguish:1 activity:6 occur:1 argument:1 developing:1 according:1 kd:3 membrane:4 describes:1 hertz:1 postsynaptic:3 character:1 primate:1 invariant:1 interference:1 equation:3 subjected:1 serf:1 rewritten:1 alternative:1 existence:1 original:1 assembly:18 l6:1 neurosc:1 question:2 realized:1 occurs:1 spike:6 exhibit:1 berlin:1 presynaptic:2 extent:1 assuming:1 code:1 regulation:1 statement:1 sakmann:1 potentiated:1 upper:1 neuron:52 observation:3 finite:2 possessed:1 t:2 displayed:3 situation:2 variability:1 drift:1 david:1 pair:1 synap:1 connection:4 coherent:1 yakovlev:3 miyashita:3 suggested:1 pattern:9 challenge:1 tau:4 memory:7 explanation:1 charging:1 event:1 critical:2 natural:1 rely:1 turning:1 scarce:1 ne:1 temporally:1 carried:1 coupled:2 nir:1 raymond:2 vreset:2 relative:1 kempter:3 asynchronized:1 expect:1 interesting:3 integrate:1 jee:1 editor:2 lo:3 excitatory:17 cooperation:1 asynchronous:2 arriving:1 allow:1 markram:4 emerge:1 distributed:22 regard:1 curve:6 calculated:2 cortical:4 boundary:1 feedback:2 van:1 forward:2 herrmann:3 iext:1 premature:1 far:1 correlate:1 global:1 summing:1 assumed:4 regulatory:1 retinotectal:1 nature:4 tel:4 symmetry:1 investigated:2 gerstner:1 antisymmetric:1 significance:1 noise:3 allowed:2 neuronal:9 je:3 hemmen:1 obeying:1 breaking:1 levy:4 down:1 xt:1 specific:2 decay:4 evidence:2 grouping:1 exists:3 led:1 simply:1 forming:1 visual:2 ordered:2 trp:1 springer:1 fokker:1 satisfies:1 harris:1 experimentally:1 change:2 kearns:2 called:1 eytan:1 experimental:6 latter:1 frontal:1 phenomenon:1
769
1,704
Constructing Heterogeneous Committees Using Input Feature Grouping: Application to Economic Forecasting Yuansong Liao and John Moody Department of Computer Science, Oregon Graduate Institute, P.O.Box 91000, Portland, OR 97291-1000 Abstract The committee approach has been proposed for reducing model uncertainty and improving generalization performance. The advantage of committees depends on (1) the performance of individual members and (2) the correlational structure of errors between members. This paper presents an input grouping technique for designing a heterogeneous committee. With this technique, all input variables are first grouped based on their mutual information. Statistically similar variables are assigned to the same group. Each member's input set is then formed by input variables extracted from different groups. Our designed committees have less error correlation between its members, since each member observes different input variable combinations. The individual member's feature sets contain less redundant information, because highly correlated variables will not be combined together. The member feature sets contain almost complete information, since each set contains a feature from each information group. An empirical study for a noisy and nonstationary economic forecasting problem shows that committees constructed by our proposed technique outperform committees formed using several existing techniques. 1 Introduction The committee approach has been widely used to reduce model uncertainty and improve generalization performance. Developing methods for generating candidate committee members is a very important direction of committee research. Good candidate members of a committee should have (1) good (not necessarily excellent) individual performance and (2) small residual error correlations with other members. Many techniques have been proposed to reduce residual correlations between members. These include resampling the training and validation data [3], adding randomness to data [7], and decorrelation training [8]. These approaches are only effective for certain models and problems. Genetic algorithms have also been used to generate good and diverse members [6]. Input feature selection is one of the most important stages of the model learning process. It has a crucial impact both on the learning complexity and the general- Y. Liao and J. Moody 922 ization performance. It is essential that a feature vector gives sufficient information for estimation. However, too many redundant input features not only burden the whole learning process, but also degrade the achievable generalization performance. Input feature selection for individual estimators has received a lot of attention because of its importance. However, there has not been much research on feature selection for estimators in the context of committees. Previous research found that giving committee members different input features is very useful for improving committee performance [4], but is difficult to implement [9]. The feature selection problem for committee members is conceptually different than for single estimators. When using committees for estimation, as we stated previously, committee members not only need to have reasonable performance themselves, but should also make decisions independently. When all committee members are trained to model the same underlying function, it is difficult for committee members to optimize both criteria at the same time. In order to generate members that provide a good balance between the two criteria, we propose a feature selection approach, called input feature grouping, for committee members. The idea is to give each member estimator of a committee a rich but distinct feature sets, in the hope that each member will generalize independently with reduced error correlations. The proposed method first groups input features using a hierarchical clustering algorithm based on their mutual information, such that features in different groups are less related to each other and features within a group are statistically similar to each other. Then the feature set for each committee member is formed by selecting a feature from each group. Our empirical results demonstrate that forming a heterogeneous committee using input feature grouping is a promising approach. 2 Committee Performance Analysis There are many ways to construct a committee. In this paper, we are mainly interested in heterogeneous committees whose members have different input feature sets. Committee members are given different subsets of the available feature set. They are trained independently, and the committee output is either a weighted or unweighted combination of individual members' outputs. In the following, we analyze the relationship between committee errors and average member errors from the regression point of view and discuss how the residual correlations between members affect the committee error. We define the training data V = {(X.B, y.B);;:3 = 1,2, . . . N} and the test data T = {(XI', YI'); JL = 1,2, ... oo}, f. '" where both are assumed to be generated by the model: Y = t(X) + f. , N(o, (72) . The data V and T are independent, and inputs are drawn from an unknown distribution. Assume that a committee has K members. Denote the available input features as X = [Xl, X2, . .. ,X m ], the feature sets for the ith and jth members as Xi = [Xiu Xi2' . .. , xm;] and Xj = [Xjl) Xj2' . .. ,x mj ] respectively, where Xi EX, Xj E X and Xi =I X j , and the mapping function of the ith and lh member models trained on data from V as fi(Xd and fi(Xj ). Define the model error ef = tl' - h(Xn , for all JL = 1,2,3, ... ,00 and i = 1,2, ... , K. 923 Constructing Heterogeneous Committees for Economic Forecasting The MSE of a committee is K = K ~2 L ?11 [(en2] + ~2 L ?11 [efej] i=l , (1) i#j and the average MSE made by the committee members acting individually is K Eave = ~ L ?11 [(en 2 1 , (2) i=l where ?[.] denotes the expectation over all test data T. Using Jensen's inequality, we get Ec ~ Eave, which indicates that the performance of a committee is always equal to or better than the average performance of its members. We define the average model error correlation as C = K(i -1) l:~j ?11 [efejl , and then have 1 K-1 1 K-1 (3) Ec = KEave + ~C = (K + ~q)Eave , where q = Be . We consider the following four cases of q: ave ? Case 1: - K~l ~ q < O. In this case, the model errors between members are anti-correlated, which might be achieved through decorrelation training. ? Case 2: q = O. In this case, the model errors between members are uncorrelated, and we have: Ec = Eave. That is to say, a committee can do much better than the average performance of its members. k <q< 1. If Eave is bounded above, when the committee size we have Ec = qEave . This gives the asymptotic limit of a committee's performance. As the size of a committee goes to infinity, the committee error is equal to the average model error correlation C. The difference between Ec and Eave is determined by the ratio q. ? Case 3: 0 K -t 00, ? Case 4: q = 1. In this case, Ec is equal to Eave. This happens only when ei = ej, for all i,j = 1, ... ,K. It is obvious that there is no advantage to combining a set of models that act identically. It is clear from the analyses above that a committee shows its advantage when the ratio q is less than one. The smaller the ratio q is, the better the committee performs compared to the average performance of its members. For the committee to achieve substantial improvement over a single model, committee members not only should have small errors individually, but also should have small residual correlations between each other. 3 Input Feature Grouping One way to construct a feature subset for a committee member is by randomly picking a certain number of features from the original feature set. The advantage of this method is that it is simple. However, we have no control on each member's performance or on the residual correlation between members by randomly selecting subsets. Y. Liao and J. Moody 924 Instead of randomly picking a subset of features for a member , we propose an input feature grouping method for forming committee member feature sets. The input grouping method first groups features based on a relevance measure in a way such that features between different groups are less related to one another and features within a group are more related to one another. After grouping, there are two ways to form member feature sets. One method is to construct the feature set for each member by selecting a feature from ? each group. Forming a member's feature set in this way, each member will have enough information to make decision , and its feature set has less redundancy. This is the method we use in this paper. Another way is to use each group as the feature set for a committee member. In this method each member will only have partial information. This is likely to hurt individual member's performance. However, because the input features for different members are less dependent, these members tend to make decisions more independently. There is always a trade-off between increasing members ' independence and hurting individual members' performance. If there is no redundancy among input feature representations, removing several features may hurt individual members' performance badly, and the overall committee performance will be hurt even though members make decisions independently. This method is currently under investigation. The mutual information [(Xi; X j) between two input variables Xi and X j is used as the relevance measure to group inputs. The mutual information [(Xi; Xj) , which is defined in equation 4, measures the dependence between the two random variables. (4) If features Xi and X j are highly dependent, [(Xi; X j) will be large. Because the mutual information measures arbitrary dependencies between random variables , it has been effectively used for feature selections in complex prediction tasks [1], where methods bases on linear relations like the correlation are likely to make mistakes. The fact that the mutual information is independent of the coordinates chosen permits a robust estimation. 4 Empirical Studies We apply the input grouping method to predict the one-month rate of change of the Index of Industrial Production (IP), one of the key measures of economic activity. It is computed and published monthly. Figure 4 plots monthly IP data from 1967 to 1993. Nine macroeconomic time series , whose names are given in Table 1, are used for forecasting IP. Macroeconomic forecasting is a difficult task because data are usually limited, and these series are intrinsically very noise and nonstationary. These series are preprocessed before they are applied to the forecasting models. The representation used for input series is the first difference on one month time scales of the logged series. For example, the notation IP.L.Dl represents IP.L.Dl == In(IP(t)) -In(IP(t-l)). The target series is IP.L.FDl , which is defined as IP.L.FDI == In(IP(t+l)) - In(IP(t)). The data set has been one of our benchmarks for various studies [5, 10]. Constructing Heterogeneous Committees for Economic Forecasting 925 Index of Industrial Production: 1967 ? 1993 Figure 1: U.S. Index of Industrial Production (IP) for the period 1967 to 1993. Shaded regions denote official recessions, while unshaded regions denote official expansions. The boundaries for recessions and expansions are determined by the National Bureau of Economic Research based on several macroeconomic series. As is evident for IP, business cycles are irregular in magnitude, duration, and structure, making prediction of IP an interesting challenge. Series IP SP DL M2 CP CB HS TB3 Tr Description Index of Industrial Production Standard & Poor's 500 Index of Leading Indicators Money Supply Consumer Price Index Moody's Aaa Bond Yield HOUSing Starts 3-month Treasury Bill Yield Yield Curve Slope: (10-Year Bond Composite)-(3-Month Treasury Bill) Table 1: Input data series. Data are taken from the Citibase database. During the grouping procedure, measures of mutual information between all pairs of input variables are computed first. A simple histogram method is used to calculate these estimates. Then a hierarchical clustering algorithm [2] is applied to these values to group inputs. Hierarchical clustering proceeds by a series of successive fusions of the nine input variables into groups. At any particular stage, the process fuses variables or groups of variables which are closest, base on their mutual information estimates. The distance between two groups is defined as the average of the distances between all pairs of individuals in the two groups. The result is presented by a tree which illustrates the fusions made at each successive level (see Figure 2). From the clustering tree, it is clear that we can break the input variables into four groups: (IP.L.Dl, DL.L.Dl) measure recent economic changes, (SP.L .Dl) reflects recent stock market momentum, (CB.D1, TB3 .D1, Tr.D1) give interest rate information, and (M2.L.D1, CP.L.D1, HS.L.D1) provide inflation information. The grouping algorithm meaningfully clusters the nine input series. Y. Liao and J. Moody 926 ~ :; ~ ::l ::I ~ ? ~ ~ :;; ~ S ~ es ~ Figure 2: Variable grouping based on mutual information. Y label is the distance. Eighteen differept subsets of features can be generated from the four groups by selecting a feature from each group. Each subset is given to a committee member. For example, the subsets (IP.L.Dl, SP.L.Dl, CB.Dl, M2.L.Dl) and (DL.L.Dl, SP.L.Dl, TB3.Dl, M2.L.Dl) are used as feature sets for different committee members. A committee has totally eighteen members. Data from Jan. 1950 to Dec. 1979 is used for training and validation, and from Jan. 1980 to Dec. 1989 is used for testing. Each member is a linear model that is trained using neural net techniques. We compare the input grouping method with three other committee member generating methods: baseline, random selection, and bootstrapping. The baseline method is to train a committee member using all the input variables. Members are only different in their initial weights. The bootstrapping method also trains a member using all the input features, but each member has different bootstrap replicates of the original training data as its training and validation sets. The random selection method constructs a feature set for a member by randomly picking a subset from the available features. For comparison with the grouping method, each committee generated by these three methods has 18 members. Twenty runs are performed for each of the four methods in order to get reliable performance measures. Figure 3 shows the boxplots of normalized MSE for the four methods. The grouping method gives the best result, and the performance improvement is significant compared to other methods. The grouping method outperforms the random selection method by meaningfully grouping of input features. It is interesting to note that the heterogeneous committee methods, grouping and random selection, perform better than homogeneous methods for this data set. One of the reasons for this is that giving different members different input sets increases their model independence. Another reason could be that the problem becomes easier to model because of smaller feature sets. 5 Conclusions The performance of a committee depends on both the performance of individual members and the correlational structure of errors between members. An empirical study for a noisy and nonstationary economic forecasting problem has demonstrated that committees constructed by input variable grouping outperform committees formed by randomly selecting member input variables. They also outperform committees without any input variable manipulation. 927 Constructing Heterogeneous Committees Jor Economic Forecasting Commltt.. Performanc. Comp.rteon (20 rurw) 0.84 8; I I 0.82 UJ ~ 0.8 ) J 0.78 ! 9 0.75 I -L 1 0.74 8 I -I1 1 : Gr~lng. 2 3 2:Random ??lectlon. 3.BaM"n..... Bootatrtpplng Figure 3: Comparison between four different committee member generating methods. The proposed grouping method gives the best result, and the performance improvement is significant compared to the other three methods. References [1] R . Battiti. Using mutual information for selecting features in supervised neural net learning. IEEE TI-ans. on Neural Networks, 5(4), July 1994. [2] B.Everitt. Cluster Analysis. Heinemann Educational Books, 1974. [3] L. Breiman. Bagging predictors. Machine Learning, 24(2):123-40, 1996. [4] K.J. Cherkauer. Human expert-level performance on a scientific image analysis task by a system using combined artificai neural networks. In P. Chan, editor, Working Notes of the AAAI Workshop on Integrating Multiple Learned Models, pages 15-2l. 1996. [5] J. Moody, U. Levin, and S. Rehfuss. Predicting the U.S. index of industrial production. In proceedings of the 1993 Parallel Applications in Statistics and Economics Conference, Zeist, The Netherlands . Special issue of Neural Network World, 3(6):791794, 1993. [6] D. Opitz and J. Shavlik. Generating accurate and diverse members of a neuralnetwork ensemble. In D . Touretzky, M. Mozer, and M. Hasselmo, editors, Advances in Neural Information Processing Systems 8. MIT Press, Cambridge, MA, 1996. [7] Y. Raviv and N. Intrator. Bootstrapping with noise: An effective regularization technique. Connection Science, 8(3-4):355-72, 1996. [8] B. E. Rosen. Ensemble learning using decorrelated neural networks. Science, 8(3-4):373-83, 1996. Connection [9] K. Tumer and J. Ghosh. Error correlation and error reduction in ensemble classifiers. Connection Science, 8(3-4):385-404, December 1996. [10] L. Wu and J . Moody. A smoothing regularizer for feedforward and recurrent neural networks. Neural Computation, 8.3:463- 491, 1996.
1704 |@word h:2 achievable:1 tr:2 reduction:1 initial:1 contains:1 series:11 selecting:6 genetic:1 outperforms:1 existing:1 john:1 designed:1 plot:1 resampling:1 ith:2 tumer:1 successive:2 constructed:2 supply:1 market:1 themselves:1 increasing:1 totally:1 becomes:1 underlying:1 bounded:1 notation:1 ghosh:1 bootstrapping:3 act:1 ti:1 xd:1 classifier:1 control:1 before:1 limit:1 mistake:1 might:1 shaded:1 limited:1 graduate:1 statistically:2 testing:1 implement:1 bootstrap:1 procedure:1 jan:2 empirical:4 composite:1 integrating:1 get:2 selection:10 context:1 optimize:1 bill:2 unshaded:1 demonstrated:1 go:1 attention:1 educational:1 independently:5 duration:1 economics:1 m2:4 estimator:4 coordinate:1 hurt:3 target:1 xjl:1 homogeneous:1 designing:1 database:1 calculate:1 region:2 cycle:1 trade:1 observes:1 substantial:1 mozer:1 complexity:1 trained:4 stock:1 various:1 regularizer:1 train:2 distinct:1 effective:2 whose:2 widely:1 say:1 cherkauer:1 statistic:1 noisy:2 ip:17 housing:1 advantage:4 net:2 propose:2 combining:1 achieve:1 description:1 xj2:1 cluster:2 generating:4 raviv:1 oo:1 fdi:1 recurrent:1 ex:1 received:1 direction:1 human:1 generalization:3 investigation:1 inflation:1 cb:3 mapping:1 predict:1 estimation:3 bond:2 currently:1 label:1 individually:2 grouped:1 hasselmo:1 weighted:1 reflects:1 hope:1 mit:1 always:2 ej:1 breiman:1 improvement:3 portland:1 indicates:1 mainly:1 rehfuss:1 industrial:5 ave:1 baseline:2 dependent:2 relation:1 interested:1 i1:1 overall:1 among:1 issue:1 smoothing:1 special:1 mutual:10 equal:3 construct:4 represents:1 rosen:1 randomly:5 national:1 individual:10 interest:1 highly:2 replicates:1 accurate:1 partial:1 lh:1 tree:2 subset:8 predictor:1 levin:1 gr:1 too:1 dependency:1 combined:2 off:1 picking:3 together:1 moody:7 aaai:1 book:1 expert:1 leading:1 oregon:1 depends:2 performed:1 view:1 lot:1 break:1 analyze:1 start:1 parallel:1 slope:1 formed:4 ensemble:3 yield:3 conceptually:1 generalize:1 comp:1 published:1 randomness:1 touretzky:1 decorrelated:1 obvious:1 intrinsically:1 supervised:1 box:1 though:1 stage:2 correlation:11 working:1 ei:1 scientific:1 name:1 contain:2 normalized:1 ization:1 regularization:1 assigned:1 during:1 criterion:2 evident:1 complete:1 demonstrate:1 performs:1 cp:2 image:1 ef:1 fi:2 jl:2 significant:2 monthly:2 cambridge:1 hurting:1 everitt:1 money:1 recession:2 base:2 closest:1 recent:2 chan:1 manipulation:1 certain:2 inequality:1 battiti:1 yi:1 neuralnetwork:1 period:1 redundant:2 july:1 multiple:1 impact:1 prediction:2 regression:1 heterogeneous:8 liao:4 expectation:1 histogram:1 achieved:1 dec:2 irregular:1 crucial:1 tend:1 member:72 meaningfully:2 december:1 nonstationary:3 feedforward:1 identically:1 enough:1 affect:1 xj:4 independence:2 economic:9 reduce:2 idea:1 forecasting:9 nine:3 useful:1 clear:2 netherlands:1 reduced:1 generate:2 outperform:3 diverse:2 group:21 redundancy:2 four:6 key:1 drawn:1 preprocessed:1 boxplots:1 fuse:1 year:1 run:1 uncertainty:2 logged:1 lng:1 almost:1 reasonable:1 wu:1 decision:4 badly:1 activity:1 infinity:1 x2:1 department:1 developing:1 combination:2 poor:1 smaller:2 making:1 happens:1 en2:1 aaa:1 taken:1 equation:1 previously:1 discus:1 committee:62 xi2:1 available:3 permit:1 apply:1 hierarchical:3 intrator:1 original:2 bagging:1 bureau:1 denotes:1 clustering:4 include:1 giving:2 uj:1 opitz:1 dependence:1 distance:3 degrade:1 reason:2 consumer:1 index:7 relationship:1 ratio:3 balance:1 difficult:3 stated:1 unknown:1 twenty:1 perform:1 benchmark:1 eighteen:2 anti:1 arbitrary:1 pair:2 connection:3 learned:1 proceeds:1 usually:1 bam:1 xm:1 challenge:1 reliable:1 decorrelation:2 business:1 predicting:1 indicator:1 residual:5 improve:1 asymptotic:1 interesting:2 validation:3 sufficient:1 editor:2 uncorrelated:1 production:5 jth:1 institute:1 shavlik:1 boundary:1 curve:1 xn:1 world:1 jor:1 rich:1 unweighted:1 made:2 ec:6 assumed:1 xi:9 table:2 promising:1 mj:1 robust:1 improving:2 expansion:2 mse:3 excellent:1 necessarily:1 complex:1 constructing:4 official:2 sp:4 whole:1 noise:2 tl:1 en:1 momentum:1 xl:1 candidate:2 removing:1 treasury:2 jensen:1 grouping:20 essential:1 burden:1 dl:16 fusion:2 adding:1 effectively:1 importance:1 workshop:1 magnitude:1 illustrates:1 easier:1 likely:2 forming:3 xiu:1 extracted:1 ma:1 month:4 price:1 change:2 heinemann:1 determined:2 reducing:1 acting:1 correlational:2 called:1 e:1 macroeconomic:3 relevance:2 d1:6 correlated:2
770
1,705
A recurrent model of the interaction between Prefrontal and Inferotemporal cortex in delay tasks ALFONSO RENART, NESTOR PARGA Departamento de F{sica Te6rica Universidad Aut6noma de Madrid Canto Blanco, 28049 Madrid, Spain http://www.ft.uam.es/neurociencialGRUPO/grup0.1!nglish.html and EDMUND T. ROLLS Oxford University Department of Experimental Psychology South Parks Road, Oxford OX] 3UD, England Abstract A very simple model of two reciprocally connected attractor neural networks is studied analytically in situations similar to those encountered in delay match-to-sample tasks with intervening stimuli and in tasks of memory guided attention. The model qualitatively reproduces many of the experimental data on these types of tasks and provides a framework for the understanding of the experimental observations in the context of the attractor neural network scenario. 1 Introduction Working memory is usually defined as the capability to actively hold information in memory for short periods of time. In primates, visual working memory is usually studied in experiments in which, after the presentation of a given visual stimulus, the monkey has to withhold its response during a certain delay period in which no specific visual stimulus is shown. After the delay, another stimulus is presented and the monkey has to make a response which depends on the interaction between the two stimuli. In order to bridge the temporal gap between the stimuli, the first one has to be held in memory during the delay. Electrophysiological recordings in primates during the performance of this type of tasks has revealed that some populations of neurons in different brain areas such as prefrontal (PF), inferotemporal (IT) or posterior parietal (PP) cortex, maintain approximately constant firing rates during the delay periods (for a review see [1]) and this delay activity states have been postulated as the internal representations of the stimuli provoking them [2]. Although up to now most of the modeling effort regarding the operation of networks able to support stable delay activity states has been put in the study of un i-modular (homogeneous) networks, there is evidence that in order for the monkey to solve the tasks satisfactorily, the interaction of several different neural structures is needed. A number of studies of delay match-to-sample tasks with intervening stimuli in primates performed by Desimone and 172 A. Renart, N Parga and E. TRolls colleagues has revealed that although IT cortex supports delay activity states and shows memory related effects (differential responses to the same, fixed stimulus depending on its status on the trial, e.g. whether it matches or not the sample), it cannot, by itself, provide the information necessary to solve the task, as the delay activity states elicited by each of the stimuli in a sequence are disrupted by the input information associated with each new stimulus presented [3, 4, 5]. Another structure is therefore needed to store the information for the whole duration of the trial. PF cortex is a candidate, since it shows selective delay activity maintained through entire trials even with intervening stimuli [6]. A series of parallel experiments by the same group on memory guided attention [7, 8] have also shown differential firing of IT neurons in response to the same visual stimulus shown after a delay (an array of figures), depending on previous information shown before the delay (one of the figures in the array working as a target stimulus). This evidence suggests a distributed memory system as the proper scenario to study working memory tasks as those described above. Taking into account that both IT and PF cortex are known to be able to support delay activity states, and that they are bi-directionally connected, in this paper we propose a simple model consisting of two reciprocally connected attractor neural networks to be identified with IT and PF cortex. Despite its simplicity, the model is able to qualitatively reproduce the behavior of IT and PF cortex during delay match-to-sample tasks with intervening stimuli, the behavior of IT cells during memory guided attention tasks, and to provide an unified picture of these experimental data in the context of associative memory and attractor neural networks. 2 Model and dynamics The model network consists of a large number of (excitatory) neurons arranged in two modules. Following [9, 10], each neuron is assumed to be a dynamical element which transforms an incoming afferent current into an output spike rate according to a given transduction function. A given afferent current Iai to neuron i (i = 1, ... ,N) in module a (a IT, PF) decays with a characteristic time constant T but increases proportionally to the spike rates Vbj of the rest of the neurons in the network (both from inside and outside its module) connected to it, the contribution of each presynaptic neuron, e.g. neuron j from module b, being proportional to the synaptic efficacy between the two. This can be expressed through the following equation = Jt/ d1ai(t) = _ Iai(t) dt T + '" J~~,b) ~ ~J . + h(~xt) VbJ bj a~ (1) . An external current h~~xt) from outside the network, representing the stimuli, can also be imposed on every neuron. Selective stimuli are modeled as proportional to the stored patterns, i.e. h~~ezt) = haTJ~i' where ha is the intensity of the external current to module a. The transduction function of the neurons transforming currents into rates has been chosen as a threshold hyperbolic tangent of gain G and threshold O. The synaptic efficacies between the neurons of each module and between the neurons in different modules are respectively [11, 12] p o "'(TJai I-' - I) (I-' I) 1(1 _J J)Nt ~ TJaj - J(a,a) ij - i.../- J' r a = IT,PF (2) p j (a,b) - ij - "'( I-' I) (I-' I) 1(1 _9 J)Nt ~ TJai TJbj - \.J" v ~,J a .../r b . (3) 173 Recurrent Model of IT-PF Interactions in Delay Tasks The intra-modular connections express the learning of P binary patterns {17~i = 0,1, f.L = 1, ... , P} by each module, each of them signaling which neurons are active in each of the sustained activity configurations. Each variable Tl~i is supposed to take the values 1 and 0 with probabilities f and (1 - f) respectively, independently across neurons and across patterns. The inter-modular connections reflect the temporal associations between the sustained activity states of each module. In this way, every stored pattern f.L in the IT module has an associated pattern in the PF module which is labelled by the same index. The normalization constant Nt = N(Jo + g) has been chosen so that the sum of the magnitudes of the inter- and the intra-modular connections remains constant and equal to 1 while their relative values are varied. When this constraint is imposed the strength of the connections can be expressed in terms of a single independent parameter 9 measuring the relative intensity of the inter- vs. the intra-modular connections (Jo can be set equal to 1 everywhere). We will limit our study to the case where the number of stored patterns per module P does not increase proportionally to the size of the modules N since a large number of stored patterns does not seem necessary to describe the phenomenology of the delay match-to-sample experiments. Since the number of neurons in a typical network one may be interested in is very large, e.g. '" 10 5 - 106 , the analytical treatment of the set of coupled differential equations (1) becomes intractable. On the other hand, when the number of neurons is large, a reliable description of the asymptotic solutions of these equations can be found using the techniques of statistical mechanics [13, 9]. In this framework, instead of characterizing the states of the system by the state of every neuron, this characterization is performed in terms of macroscopic quantities called order parameters which measure and quantify some global properties of the network as a whole. The relevant order parameters appearing in the description of our system are the overlaps of the state of each module with each of the stored patterns m~, defined as: 1 m~ = N? 2)17~i - f)Vai X i where the symbol ? ?1/ , (4) ... ?1/ stands for an average over the stored patterns. Using the free energy per neuron of the system at zero temperature :F (which we do not write explicitly to reduce the technicalities to a minimum) we have modeled the experiments by giving the order parameters the following dynamics: (5) This dynamics ensures that the stationary solutions, corresponding to the values of the order parameters at the attractors, correspond also to minima of the free energy, and that, as the system evolves, the free energy is always minimized through its gradient. The time constant of the macroscopic dynamics is a free parameter which has been chosen equal to the time constant of the individual neurons, reflecting the assumption that neurons operate in parallel. Its value has been set to T = 10 ms. Equations (5) have been solved by a simple discretizing procedure (first order Runge-Kutta method). Since not all neurons in the network receive the same inputs, not all of them behave in the same way, i.e. have the same firing rates. In fact, the neurons in each of the module can be split into different sub-populations according to their state of activity in each of the stored patterns. The mean firing rate of the neurons in each SUb-population depends on the particular state realized by the network (characterized by the values of the order parameters). Associated to each pattern there are two larger sub-populations, to be denoted as foreground (all active neurons) and background (all inactive neurons) of that pattern. 174 A. Renart, N. Parga and E. T. Rolls The overlap with a given pattern can be expressed as the difference between the mean firing rate of the neurons in its foreground and its background. The average is performed over all other sub-populations to which each neuron in the foreground (background) may belong to, where the probability of a given sub-population is equal to the fraction of neurons in the module belonging to it (determined by the probability distribution of the stored patterns as given above). This partition of the neurons into sub-populations is appealing since, in experiments, cells are usually classified in terms of their response properties to a set of fixed stimuli, i.e. whether each stimulus is effective or ineffective in driving their response. The modeling of the different experiments proceeded according to the macroscopic dynamics (5), where each stimulus was implemented as an extra current for a desired period of time. 3 Sequence with intervening stimuli In order to study delay match-to-sample tasks with intervening stimuli [5, 6], the module to be identified with IT was sequentially stimulated with external currents proportional to some of the stored patterns with a delay between them. To take into account the large fraction of PF neurons with non-selective responses to the visual stimuli (which may be involved in other aspects of the task different from the identification of the stimuli), and since the neurons in our modules are, by definition, stimulus selective (although they are probably connected to the non-selective neurons) a constant, non-selective current of the same intensity as the selective input to the IT module was applied (during the same time) equally to all sub-populations of the PF module. The external current to the IT module was stimulus selective because the fraction of IT neurons with non-selective responses to the visual stimuli is very small [6]. The results can be seen in Figure 1 where the sequence ABA with A as the sample stimulus and B as a non-matching stimulus has been studied. The values of the model parameters are listed in the caption. In Figure 1a, the mean firing rates of the foreground populations of patterns AIT and BIT of the IT module have been plotted as a function of time. The main result is that, as observed in the experiments, the delay activity in the IT module is determined by the last stimulus presented. The delay activity provoked by a given stimulus is disrupted by the next, unless it corresponds to the same stimulus, in which case the effect of the stimulus is to increase the firing rate of the neurons in its foreground. We have checked that no noticeable effects occur if more nonmatching stimuli are presented (they are all equivalent with respect to the sample) or if a non-match stimulus is repeated. If the coupling g between the modules is weak enough [12] the behavior in the PF module is different. This can be seen in Figure 1b, where the time evolution of the mean firing rates of the foreground of the two associated patterns ApF and BpF stored in the PF module are shown. In agreement with the findings of Desimone and colleagues, the neurons in the PF module remain correlated with the sample for the whole trial, despite the non-selective signal received by all PF neurons (not only those in the foreground of the sample) and the fact that the selective current from the IT module tends to activate the pattern associated with the current stimulus. Desimone and colleagues [5, 6] report that the response of some neurons (not necessarily those with sample selective delay activity or with stimulus selective responses) in both IT and PF cortex to some stimuli, is larger if those stimuli are matches in their trials than if the same stimuli are non-matches. This has been denoted as match enhancement. In the present scenario the explanation is straightforward: when a stimulus is a non-match, IT and PF are in different states and therefore send inconsistent signals to each other. The firing rate of the neurons of each module is maintained in that case solely by the contribution to the total current coming from the recurrent collaterals. On the other hand, when the stimulus is the match, both modules find themselves in states associated in the synapses 175 Recurrent Model of IT-PF Interactions in Delay Tasks (a) (b) 0.8 0.7 Ir- 0.5 C> 0.4 a: .'"Ii:c: i 0.7 !i '\ 0.6 ., a; r- ' ''I ! 0.6 ! , i I 0.3 \ 0.2 0.1 0.5 C> 0.4 a: I \ "'---._.._-- ., a; c: ?c Ii: \. 0.3 0.2 r.~ .... . . _ .._ .-J~_ _ _--,?L.__ O'---'--~~-~~~-"""""'-'---' o 2 3 4 56789 Time (s) . \ fL... i\..... ?? "l. . ____J \. ___.___._.J l.__._.___. 0.1 ...........-,---, OL--'--~~-~~~- o 234 5 6 7 8 9 Time (s) Figure 1: (a) Mean rates in the foreground of patterns AIT (solid line) and BIT (dashed line) in the IT module as a function of time. (b) Same but for patterns A pF and BpF of the PF module. Model parameters are G = 1.3, () = 10- 3 , f = 0.2, 9 = 10- 2 , h = 0.13. Stimuli are presented during 500 ms at seconds 0, 3, and 6 following the sequence ABA. between the neurons connecting them, PF because it has remained that way the whole trial, and IT because it is driven by the current stimulus. When this happens, the contribution to the total current from the recurrent collaterals and from the long range afferents add up consistently, and the firing rate increases. In order for this explanation to hold there should be a correlation between the top-down input from PF and the sensory bottom-up signal to IT. Indeed, experimental evidence for such a correlation has very recently been found [14]. This is an important experimental finding which supports our theory. Looking at Figure 1, one sees that the effect is not evident in the model during the time of stimulus presentation, which is the period where it has been reported. The effect is, in fact, present, although its magnitude is too small to be noticeable in the figure. We would argue, however, that this quantitative difference is an artifact of the model. This is because the enhancement effect is very noticeable on the delay periods, where essentially the same neurons are active as during the stimulus presentations (i.e., where the same correlations between the top-down and bottom-up signals exist) but with lower firing rates. During stimulus presentations the firing rates are closer to the saturation regime, and therefore the dynamical response range of the neurons is largely reduced. 4 Memory guided attention To test the differential response of cells as a function of the contents of memory, we have followed [7, 8] and studied a sub-population of IT cells which are simultaneously in the foreground of one of the patterns (AIT) and in the background of another (BIT) in the same conditions as the previous section (same model parameters). In Figure 2a the response of this sub-population as a function of time has been plotted in two different situations. In the first one, the effective stimulus AIT was shown first (throughout this section non selective stimulation of PF proceeded as in the last section) and after a delay, a stimulus array equal to the sum of AIT and BIT was presented. The second situation is exactly equal, except for the fact that the cue stimulus shown first was the ineffective stimulus BIT. The response of the same sub-population to the same stimulus array is totally different and determined by the cue stimulus: If the sub-population is in the background of the cue, its response is null during the trial except for the initial period of the presentation of the array. In accordance with the experimental observations [7, 8], its response grows initially (as one would expect, since during the array presentation time, stimulation is symmetric with A. Renart, N. Parga and E. T. Rolls 176 respect of A and B) but is later suppressed by the top-down signal being sent by the PF module. This suppression provides a clear example of a situation in which the contents of memory (in the form of an active PF activity state) are explicitly gating the access of sensory information to IT, implementing a non-spatial attentional mechanism. (a) 0.8 (b) .---~~~-~-~-~---, r 0.7 r- 0.6 0.5 Q) iii cr: 0.4 cr: 0.4 c ."'" ?c u: *'" c u: 0.2 r\ 0.1 _.. _--_..-----_...._....! \ .._----_ .._"- 2 3 Time (s) 4 5 I ! ? I i i l.-\ 0.5 0.3 0 r-?1 0.7 0.6 l_. _ _. _ ._._. 0.3 0.2 0.1 r--\ 0 -0.1 6 0 2 3 4 5 6 Time (s) Figure 2: (a) Mean rates as a function of time in IT neurons which are both in the foreground AIT and in the background of BIT when the cue stimulus is AIT (solid line) or BIT (dashed line). (b) Mean rates of the same neurons when CIT is the cue stimulus and the array is AIT alone (long dashed line), BIT alone (short dashed line) or the sum of AIT and BIT (solid line). Cue present until 500 ms. Array present from 3000 ms to 3500 ms. Model parameters as in Figure 1 In the model, the PF module remains in a state correlated with the cue during the whole trial (to our knowledge there are no measurements of PF activity during memory guided attention tasks) and therefore provides a persistent signal 'in the direction' ofthe cue which biases the competition between AIT and BIT established at the onset of the array. This is how the gating mechanism is implemented. The competitive interactions between the stimuli in the array are studied in Figure 2b, which is an emulation of the target-absent trials of [8]. In this figure, the same sub-population is studied under situations in which the cue stimulus is not present in the array (another one of the stored patterns, i.e. CIT) The three curves correspond to different arrays: The effective stimulus alone, the ineffective stimulus alone, and a sum of the two as in the previous experiment. In all three, the PF module remains in a sustained activity state correlated with CIT the whole trial and therefore, since the patterns are independent, the signal it sends to IT is symmetric with respect of A and B. Thus, the response of the sub-population during the array is in this case unbiased, and the effect of the competitive interactions can be isolated. The result is that, as observed experimentally, the response to the complex array is intermediate between the one to the effective stimulus alone and the one to the ineffective stimulus alone. The nature of the competition in an attractor network like the one under study here is based on the fact that complex stimulus combinations are not stored in the recurrent collaterals of each module. These connections tend to stabilize the individual patterns which, being independent, tend to cancel each other when presented together. After the array is presented, the state of the IT module, which is correlated with CIT in the initial delay, becomes correlated with AIT or BIT if they are presented alone. When the array contains both of them in a symmetric fashion, since the sum of the patterns is not a stored pattern itself, the IT module remains correlated with pattern CIT due to the signal from the PF module. 177 Recurrent Model of IT-PF Interactions in Delay Tasks 5 Discussion We have proposed a toy model consisting of two reciprocally connected attractor modules which reproduces nicely experimental observations regarding intra-trial data in delay match-to-sample and memory guided attention experiments in which the interaction between IT and PF cortex is relevant. Several important issues are taken into account in the model: a complex interaction between the PF and IT modules resultant from the association of frequent patterns of activity in both modules, delay activity states in each module which exert mutually modulatory influences on each other, and a common substrate (we emphasize that the results on Sections 3 and 4 where obtained with exactly the same model parameters, just by changing the type of task) for the explanation of apparently diverse phenomena. Perception is clearly an active process which results from the complex interactions between past experience and incoming sensory information. The main goal of this model was to show that a very simple associational (Hebbian) pattern of connectivity between a perceptual module and a 'working memory' module can provide the basic ingredients needed to explain coherently different experimentally found neural mechanisms related to this process. The model has clear limitations in terms of 'biological realism' which will have to be improved in order to use it to make quantitative predictions and comparisons, and does not provide a complete an exhaustive account of the very complex and diverse phenomena in which temporo-frontal interactions are relevant (there is, for example, the issue of how to reset PF activity in between trials [15]). However, it is precisely the simplicity of the mechanism it provides and the fact that it captures the essential features of the experiments, despite being so simple, what makes it likely that it will remain relevant after being refined. Acknowledgements This work was funded by a Spanish grant PB96-0047. We acknowledge the Max Planck Institute for Physics of Complex Systems in Dresden, Germany, for the hospitality received by A.R. and N.P. during the meeting held there from March 1 to 26, 1999. References [1] J. M. Fuster. Memory in the cerebral cortex. Cambridge, MA: MIT Press (1995) [2] D. J. Amit. Belulvioral and Brain Sciences 18, 617-657 (1995) [3] G. C. Baylis & E. T. Rolls. Exp. Brain Res. 65,614-622 (1987) [4] E. K. Miller, L. Li & R. Desimone. 1. Neurosci. 13, 1460-1478 (1993) [5] E. K. Miller & R. Desimone. Science 263,520-522 (1994) [6] E. K. Miller, C. A. Erickson & R. Desimone. 1. Neurosci. 16,5154-5167 (1996) [7] L. Chelazzi, E. K. Miller, J. Duncan & R. Desimone. Nature 363,345-347 (1993) [8] L. Chelazzi, J. Duncan, E. K. Miller & R. Desimone. 1. Neurophysiol. 80,2918-2940 (1998) [9] R. Kuhn. In Statistical Meclulnics of Neural Networks. (ed. L. Garrido), 19-32. Berlin: Springer-Verlag (1990) [10] D. J. Amit & M. V. Tsodyks. Network 2,259-273 (1991) [11] A. Renart, N. Parga & E. T. Rolls. Neural Computation 11, 1349-1388 (1999). [12] A. Renart, N. Parga & E. T. Rolls. Network 10,237-255 (1999). [13] M. Mezard, G. Parisi & M. Virasoro. Spin glass theory and beyond. Singapore: World Scientific (1987) [14] H. Tomita, M Ohbayashi, K. Nakahara, I. Hasegawa & Y. Miyashita. Nature 401, 699-703 (1999) [15] D. Durstewitz, M. Kelc & o. Giintiirkiin. 1. Neurosci. 19,2807-2822 (1999)
1705 |@word trial:12 proceeded:2 solid:3 initial:2 configuration:1 series:1 efficacy:2 contains:1 past:1 current:14 nt:3 partition:1 v:1 stationary:1 cue:9 alone:7 realism:1 short:2 provides:4 characterization:1 troll:1 differential:4 persistent:1 consists:1 sustained:3 inside:1 inter:3 indeed:1 behavior:3 themselves:1 mechanic:1 brain:3 ol:1 pf:33 totally:1 becomes:2 spain:1 null:1 what:1 monkey:3 unified:1 finding:2 temporal:2 quantitative:2 every:3 exactly:2 grant:1 planck:1 before:1 accordance:1 tends:1 limit:1 despite:3 oxford:2 firing:12 solely:1 approximately:1 exert:1 studied:6 dresden:1 suggests:1 bi:1 range:2 satisfactorily:1 signaling:1 procedure:1 area:1 hyperbolic:1 matching:1 road:1 cannot:1 put:1 context:2 influence:1 www:1 equivalent:1 imposed:2 send:1 straightforward:1 attention:6 duration:1 independently:1 simplicity:2 array:16 population:15 nglish:1 target:2 caption:1 substrate:1 homogeneous:1 agreement:1 element:1 observed:2 ft:1 module:45 bottom:2 solved:1 capture:1 tsodyks:1 ensures:1 connected:6 transforming:1 dynamic:5 neurophysiol:1 describe:1 effective:4 activate:1 outside:2 refined:1 exhaustive:1 modular:5 larger:2 solve:2 itself:2 directionally:1 associative:1 runge:1 sequence:4 parisi:1 analytical:1 propose:1 interaction:12 coming:1 reset:1 frequent:1 relevant:4 supposed:1 intervening:6 canto:1 description:2 competition:2 blanco:1 enhancement:2 depending:2 recurrent:7 coupling:1 ij:2 received:2 noticeable:3 implemented:2 quantify:1 direction:1 guided:6 kuhn:1 emulation:1 implementing:1 biological:1 hold:2 exp:1 baylis:1 bj:1 garrido:1 driving:1 bridge:1 mit:1 clearly:1 hospitality:1 always:1 cr:2 consistently:1 suppression:1 glass:1 entire:1 initially:1 selective:14 reproduce:1 interested:1 germany:1 issue:2 html:1 denoted:2 spatial:1 equal:6 nicely:1 park:1 cancel:1 aba:2 foreground:10 minimized:1 report:1 stimulus:62 simultaneously:1 nestor:1 individual:2 consisting:2 attractor:7 maintain:1 intra:4 held:2 desimone:8 closer:1 necessary:2 experience:1 collateral:3 unless:1 desired:1 plotted:2 re:1 isolated:1 virasoro:1 modeling:2 measuring:1 delay:30 too:1 stored:13 reported:1 disrupted:2 universidad:1 physic:1 connecting:1 together:1 jo:2 connectivity:1 reflect:1 provoked:1 prefrontal:2 external:4 actively:1 toy:1 account:4 li:1 de:2 stabilize:1 postulated:1 explicitly:2 afferent:3 depends:2 onset:1 performed:3 later:1 apparently:1 competitive:2 capability:1 elicited:1 parallel:2 contribution:3 ir:1 spin:1 roll:6 characteristic:1 largely:1 miller:5 correspond:2 ofthe:1 weak:1 identification:1 parga:6 classified:1 explain:1 synapsis:1 synaptic:2 checked:1 definition:1 ed:1 energy:3 colleague:3 pp:1 involved:1 resultant:1 associated:6 gain:1 treatment:1 knowledge:1 electrophysiological:1 reflecting:1 dt:1 response:18 improved:1 iai:2 arranged:1 ox:1 just:1 correlation:3 until:1 working:5 hand:2 artifact:1 scientific:1 grows:1 effect:7 unbiased:1 evolution:1 analytically:1 symmetric:3 during:17 spanish:1 maintained:2 m:5 evident:1 complete:1 temperature:1 recently:1 common:1 stimulation:2 cerebral:1 association:2 belong:1 measurement:1 cambridge:1 funded:1 stable:1 access:1 cortex:10 add:1 inferotemporal:2 posterior:1 driven:1 scenario:3 store:1 certain:1 verlag:1 binary:1 discretizing:1 meeting:1 seen:2 minimum:2 ud:1 period:7 dashed:4 ii:2 signal:8 hebbian:1 match:13 england:1 characterized:1 long:2 equally:1 prediction:1 basic:1 renart:6 essentially:1 normalization:1 cell:4 receive:1 background:6 sends:1 macroscopic:3 extra:1 rest:1 operate:1 ineffective:4 south:1 recording:1 probably:1 tend:2 alfonso:1 sent:1 inconsistent:1 seem:1 revealed:2 split:1 enough:1 iii:1 intermediate:1 psychology:1 identified:2 reduce:1 regarding:2 absent:1 inactive:1 whether:2 effort:1 modulatory:1 proportionally:2 listed:1 clear:2 transforms:1 cit:5 reduced:1 http:1 exist:1 singapore:1 per:2 diverse:2 write:1 express:1 group:1 threshold:2 changing:1 fraction:3 sum:5 everywhere:1 throughout:1 duncan:2 bit:11 fl:1 followed:1 encountered:1 chelazzi:2 activity:18 strength:1 occur:1 constraint:1 precisely:1 aspect:1 department:1 according:3 combination:1 march:1 belonging:1 across:2 remain:2 suppressed:1 appealing:1 evolves:1 primate:3 happens:1 taken:1 equation:4 mutually:1 remains:4 mechanism:4 needed:3 operation:1 uam:1 phenomenology:1 appearing:1 top:3 tomita:1 giving:1 amit:2 quantity:1 spike:2 realized:1 coherently:1 erickson:1 gradient:1 kutta:1 attentional:1 berlin:1 presynaptic:1 argue:1 modeled:2 index:1 hasegawa:1 vai:1 proper:1 observation:3 neuron:43 acknowledge:1 behave:1 parietal:1 situation:5 looking:1 varied:1 intensity:3 nonmatching:1 connection:6 established:1 miyashita:1 able:3 beyond:1 usually:3 dynamical:2 pattern:29 perception:1 regime:1 provoking:1 saturation:1 reliable:1 memory:18 reciprocally:3 apf:1 explanation:3 max:1 overlap:2 representing:1 picture:1 coupled:1 review:1 understanding:1 acknowledgement:1 tangent:1 relative:2 asymptotic:1 expect:1 limitation:1 proportional:3 ingredient:1 excitatory:1 last:2 free:4 sica:1 l_:1 bias:1 institute:1 taking:1 characterizing:1 distributed:1 curve:1 withhold:1 stand:1 world:1 sensory:3 qualitatively:2 emphasize:1 status:1 technicality:1 reproduces:2 active:5 incoming:2 global:1 sequentially:1 assumed:1 un:1 fuster:1 stimulated:1 nature:3 necessarily:1 complex:6 main:2 neurosci:3 whole:6 vbj:2 ait:11 repeated:1 madrid:2 tl:1 transduction:2 fashion:1 sub:13 mezard:1 candidate:1 perceptual:1 down:3 remained:1 specific:1 departamento:1 jt:1 xt:2 gating:2 symbol:1 decay:1 evidence:3 intractable:1 essential:1 edmund:1 magnitude:2 associational:1 gap:1 likely:1 visual:6 expressed:3 durstewitz:1 springer:1 corresponds:1 ma:1 ezt:1 goal:1 presentation:6 nakahara:1 labelled:1 content:2 experimentally:2 typical:1 determined:3 except:2 called:1 total:2 e:1 experimental:8 internal:1 support:4 frontal:1 phenomenon:2 correlated:6
771
1,706
From Coexpression to Coregulation: An Approach to Inferring Transcriptional Regulation among Gene Classes from Large-Scale Expression Data Eric Mjolsness Jet Propulsion Laboratory California Institute of Technology Pasadena CA 91109-8099 mjolsness@jpl.nasa.gov Tobias Mann Jet Propulsion Laboratory California Institute of Technology Pasadena CA 91109-8099 mann@aigjpl.nasa.gov Rebecca Castaiio Jet Propulsion Laboratory California Institute of Technology Pasadena CA 91109-8099 becky@aigjpl.nasa.gov Barbara Wold Division of Biology California Institute of Technology Pasadena CA 91125 woldb@its.caltech.edu Abstract We provide preliminary evidence that eXlstmg algorithms for inferring small-scale gene regulation networks from gene expression data can be adapted to large-scale gene expression data coming from hybridization microarrays. The essential steps are (1) clustering many genes by their expression time-course data into a minimal set of clusters of co-expressed genes, (2) theoretically modeling the various conditions under which the time-courses are measured using a continious-time analog recurrent neural network for the cluster mean time-courses, (3) fitting such a regulatory model to the cluster mean time courses by simulated annealing with weight decay, and (4) analysing several such fits for commonalities in the circuit parameter sets including the connection matrices. This procedure can be used to assess the adequacy of existing and future gene expression time-course data sets for determ ining transcriptional regulatory relationships such as coregulation . 1 Introduction In a cell, genes can be turned "on" or "off' to varying degrees by the protein products of other genes. When a gene is "on" it is transcribed to produce messenger RNA (mRNA) which can subsequently be translated into protein molecules. Some of these proteins are transcription factors which bind to DNA at specific sites and thereby affect which genes are transcribed and how often. This trancriptional Inferring Transcriptional Regulation among Gene Classes 929 regulation feedback circuitry provides a fundamental mechanism for information processing in the cell. It governs differentiation into diverse cell types and many other basic biological processes. Recently, several new technologies have been developed for measuring the "expression" of genes as mRNA or protein product. Improvements in conventional f1uorescently labeled antibodies against proteins have been coupled with confocal microscopy and image processing to partially automate the simultaneous measurement of small numbers of proteins in large numbers of individual nuclei in the fruit fly Drosophila melanogaster [1]. In a complementary way, the mRNA levels of thousands of genes, each averaged over many cells, have been measured by hybridization arrays for various species including the budding yeast Saccharomyces cerevisiae [2] . The high-spatial-resolution protein antibody data has been quantitatively modeled by "gene regulation network" circuit models [3] which use continuous-time, analog, recurrent neural networks (Hopfield networks without an objective function) to model transcriptional regulation [4][5] . This approach requires some machine learning technique to infer the circuit parameters from the data, and a particular variant of simulated annealing has proven effective [6][7]. Methods in current biological use for analysing mRNA hybridization data do not infer regulatory relationships, but rather simply cluster together genes with similar patterns of expression across time and experimental conditions [8][9] . In this paper, we explore the extension of the gene circuit method to the mRNA hybridization data which has much lower spatial resolution but can currently assay a thousand times more genes than immunofluorescent image analysis. The essential problem with using the gene circuit method, as employed for immunoflourescence data, on hybridization data is that the number of connection strength parameters grows between linearly and quadratically in the number of genes (depending on sparsity assumptions) . This requires more data on each gene, and even if that much data is available, simulated annealing for circuit inference does not seem to scale well with the number of unknown parameters. Some form of dimensionality reduction is called for. Fortunately dimensionality reduction is available in the present practice of clustering the large-scale time course expression data by genes, into gene clusters. In this way one can derive a small number of cluster-mean time courses for "aggregated genes", and then fit a gene regulation circuit to these cluster mean time courses. We will discuss details of how this analysis can be performed and then interpreted. A similar approach using somewhat different algorithms for clustering and circuit inference has been taken by Hertz [10]. In the following , we will first summarize the data models and algorithms used, and then report on preliminary experiments in applying those algorithms to gene expression data for 2467 yeast genes [9][11]. Finally we will discuss prospects for and limitations of the approach. 2 Data Models and Algorithms The data model is as follows . We imagine that there is a small, hidden regulatory network of " aggregate genes" which regulate one another by the analog neural network dynamics [3] T . dv; I dt = g(~ T.v + h) ~ 1/ J / I Xv . I I E. Mjolsness, T Mann, R. Castano and B. Wold 930 In which Vi is the continuous-valued state variable for gene product i, ~j is the matrix of positive, zero, or negative connections by which one transcription factor can enhance or repress another, and gO is a nonlinear monotonic sigmoidal activation function. When a particular matrix entry ~j is nonzero, there is a regulatory "connection" from gene product} to gene i. The regulation is enhancing if T is positive and repressing if it is negative. If ~j is zero there is no connection . This network is run forwards from some initial condition and time-sampled to generate a wild-type time course for the aggregate genes. In addition, various other time courses can be generated under alternative experimental conditions by manipulating the parameters. For example an entire aggregate gene (corresponding to a cluster of real genes) could be removed from the circuit or otherwise modified to represent mutants. External input conditions could be modeled as modifications to h . Thus we get one or several time courses (trajectories) for the aggregate genes. From such aggregate time courses, actual gene data is generated by addition of Gaussian-distributed noise to the logarithms of the concentration variables. Each time point in each cluster has its own scalar standard deviation parameter (and a mean arising from the circuit dynamics). Optionally, each gene's expression data may also be multiplied by a time-independent proportionality constant. Regulatory aggregate genes (large circles) and cluster member genes (small circles). T Given this data generation model and suitable gene expression data, the problem is to infer gene cluster memberships and the circuit parameters for the aggregate genes' regulatory relationships. Then, we would like to use the inferred cluster memberships and regulatory circuitry to make testable biological predictions. This data model departs from biological reality in many ways that could prove to be important, both for inference and for prediction. Except for the Gaussian noise model, each gene in a cluster is models as fully coregulated with every other one they are influenced in the same ways by the same regulatory connection strengths. Also, the nonlinear circuit model must not only reflect transcriptional regulation, but all other regulatory circuitry affecting measured gene expression such as kinasephosphatase networks. Under this data model, one could formulate a joint Bayesian inference problem for the clustering and circuit inference aspects of fitting the data. But given the highly provisional nature of the model, we simply apply in sequence an existing mixtureof-Gaussians clustering algorithm to preprocess the data and reduce its dimensionality , and then an existing gene circuit inference algorithm . Presumably a joint optimization algorithm could be obtained by iterating these steps. 2.1 Clustering A widely used clustering algorithm for mixure model estimation is ExpectationMaximization (EM)[12]. We use EM with a diagonal covariance in the Gaussian, so that for each feature vector component a (a combination of experimental condition Inferring Transcriptional Regulation among Gene Classes 931 and time point in a time course) and cluster a there is a standard deviation parameter G aa . In preprocessing, each concentration data point is divided by its value at time zero and then a logarithm taken. The log ratios are clustered using EM. Optionally, each gene's entire feature vector may be normalized to unit length and the cluster centers likewise normalized during the iterative EM algorithm. In order to choose the number of clusters, k, we use the cross-validation algorithm described by Smyth [13]. This involves computing the likelihood of each optimized fit on a test set and averaging over runs and over divisions of the data into training and test sets. Then, we can examine the likelihood as a function of k in order to choose k. Normally one would pick k so as to maximize cross-validated likelihood. However, in the present application we also want to reward small values of k which lead to smalIer circuits for the circuit inference phase of the algorithm. The choice of k will be discussed in the next section. 2.2 Circuit Inference We use the Lam-Delosme variant of simulated annealing (SA) to derive connection strengths T, time constants t, and decay rates f..., as in previous work using this gene circuit method [4][5]. We set h to zero. The score function which SA optimizes is S(T,r,A) = AI(v;(t;T,r,A)-vi (t?)2 + WI7;j2 h ij +exp[B(I7;i2 + I A7 + Ir;)]-l ij The first term represents the fit to data Vi. The second term is a standard weight decay term. The third term forces solutions to stay within a bounded region in weight space. We vary the weight decay coefficient W in order to encourage relatively sparse connection matrix solutions. 3 3.1 Results Data We used the Saccharomyces cerevisiae data set of [9]. It includes three longer time courses representing different ways to synchronize the normal cell cycle [II], and five shorter time courses representing altered conditions. We used all eight time courses for clustering, but just 8 time points of one of the longer time courses (alpha factor synchronized cell cycle) for the circuit inference. It is likely that multiple long time courses under altered conditions will be required before strong biological predictions can be made from inferred regulatory circuit models. 3.2 Clustering We found that the most likely number of classes as determined by cross validation was about 27, but that there is a broad plateau of high-likelihood cluster numbers from 15 to 35 (Figure I). This is similar to our results with another gene expression data set for the nematode worm Caenorhabditis e/egans supplied by Stuart Kim; these more extensive clustering experiments are summarized in Figure 2. Clustering experiments with synthetic data is used to understand these results. These experiments show that the cross-validated log likelihood curve can indicate the number of clusters present in the data, justifying the use of the curve for that E. Mjolsness, T. Mann, R. Castano and B. Wold 932 purpose. In more detail, synthetic data generated from 14 20-dimensional spherical Gaussian clusters were clustered using the EMlCV algorithm. The likelihoods showed a sharp peak at k=14 unlike Figures 1 or 2. In another experiment, 14 20dimensional spherical Gaussian superclusters were used to generate second-level clusters (3 subclusters per supercluster), which in turn generated synthetic data points. This two-level hierarchical model was then clustered with the EMlCV method. The likelihood curves (not shown) were quite similar to Figures 1 and 2, with a higher-likelihood plateau from roughly 14 to 40. x 10" ~, :- ~ :1; ,i ./ I l ~'.~ ; --~--~,~ ~ ---7.1~--~~~ ' ---=a--~~7---~~--~~ ero.V.....d~ Figure 1. Cross-validated log-likelihood scores, displayed and averaged over 5 runs, for EM clustering of S. cerevisiae gene expression data [9]. Horizontal axis: k, the "requested" or maximal number of cluster centers in the fit. Some cluster centers go unmatched to data. Vertical axis: log likelihood score for the fit, scatterplotted and' averaged. Likelihoods have not been integrated over any range of parameters for hypothesis testing. k ranges from 2 to 40 in increments of 1. Solid line shows average likelihood value for each k. ~, + _u + 7---!" ~10o-----:':-'.-----::20o----,. c!=----~",-----;.. Number of Clusters ++ + +++ ~~---~~--~,~~--~,~~--~~ ~--~* Number of Clusters Figure 2. Cross-validated log-likelihood scores, averaged over 13 runs, for EM clustering of C. elegans gene expression data from S. Kim's lab. Horizontal axis: k, the "requested" or maximal number of cluster centers in the fit. Some cluster centers go unmatched to data. Vertical axis: log likelihood score for the fit, as an average over 13 runs plus or minus one standard deviation. (Left) Fine-scale plot, k =2 to 60 in increments of 2. (Right) Coarsescale plot, k=2 to 202 in increments of 10. Both plots show an extended plateau of relatively likely fits between roughly k =14 and k =40. From Figures 1 and 2 and the synthetic data experiments mentioned above, we can guess at appropriate values for k which take into account both the measured likelihood of clustering and the requirements for few parameters in circuit-fitting. For example choosing k=15 clusters would put us at the beginning of the plateau, losing very little cluster likelihood in return for reducing the aggregate genes circuit size from 27 to 15 players. The interpretation would be that there are about 15 superclusters in hierarchically clustered data, to which we will fit a 15-player Inferring Transcriptional Regulation among Gene Classes 933 regulatory circuit. Much more aggressive would be to pick k=7 or 8 clusters, for a relatively significant drop in log-likelihood in return for a further substantial decrease in circuit size. An acceptable range of cluster numbers (and circuit sizes) would seem to be k=8 to 15. 3.3 Gene Circuit Inference It proved possible to fit the k= 15 time course using weight decay W=1 but without using hidden units. W=O and W=3 gave less satisfactory results. Four of the 15 clusters are shown in Figure 3 for one good run (W= 1). Scores for our first few (unselected) runs at the current parameter settings are shown in Table 1. Each run took between 24 and 48 hours on one processor of an Sun UItrasparc 60 computer. Even with weight decay, it is possible that successful fits are really overfits with this particular data since there are about twice as many parameters as data points. Weight Decay W <Score> 0 3 <Simulated Annealing Moves>/1 0/\6 Number of runs 1.536 +/- 0.134 2.803 +/- 0.437 3 0.787 +/- 0.394 2.782 +/- 0.200 10 1.438 +/- 0.037 2.880 +/- 0.090 4 Table 1. Score function parameters were A= 1.0. B=O.O 1. Annealing runs statistics are reported when the temperature dropped below 0.0001. All the best scores and visually acceptable fits occurred in W=I runs. The average values of the data fit, weight decay, and penalty terms in the score function for W=1 were {0.378, 0.332, 0.0667} after slightly more annealing. There were a few significant similarities between the connection matrices computed in the two lowest-scoring runs. The most salient feature in the lowest-scoring network was a set of direct feedback loops among its strongest connections: cluster 8 both excited and was inhibited by cluster 10, and cluster 10 excited and was inhibited by cluster 15. This feature was preserved in the second-best run. A systematic search for "concensus circuitry" shows convergence towards a unique connection matrix for the 8-point time series data used here, but more complete 16time-point data gives mUltiple "clusters" of connection matrices. From parametercounting one might expect that making robust and unique regulatory predictions will require the use of more trajectory data taken under substantially different conditions. Such data is expected to be forthcoming. 4 Discussion We have illustrated a procedure for deriving regulatory models from large-scale gene expression data. As the data becomes more comprehensive in the number and nature of conditions under which comparable time courses are measured, this procedure can be used to determine when biological hypotheses about gene regulation can be robustly derived from the data. Acknowledgments This work was supported in part by the Whittier Foundation, the Office of Naval Research under contract NOOO 14-97-1-0422, and the NASA Advanced Concepts Program. Stuart Kim (Stanford University) provided the C. elegans gene expression array data. The GRN simulation and inference code is due in part to Charles Garrett and George Marnellos. The EM clustering code is due in part to Roberto Manduchi. 934 E. Mjolsness, T. Mann, R. Castano and B. Wold 't ~ : : : : ~i ::v:: ===:=;; J J ; : ; ; ; :~ 0.5,'-- ---'2-- , --'-3- -- .' - ----'-5- - -6'-------'-7 - - - ' 8 - 2 3 4 5 6 7 8 :~ , 2 3 4 5 6 7 8 ongIIlIII doll JIIII,1<od,.11h an ? . gm ftl JIIIIrkeel WIth an 0 Figure 3. Four clusters (numbers 9-12) of a 15-c1uster mixture of Gaussians model of 2467 genes each assayed over an eight-point time course; cluster means (shown as x) are fit to a gene regulation network model (shown as 0). References [I] D. Kosman, J. Reinitz, and D. H. Sharp, "Automated Assay of Gene Expression at Cellular Resolution " Pacific Symposium on Biocomputing '98. Eds. R. Altman, A. K. Dunker, L. Hunter, and T. E. Klein" World Scientific 1998. [2] J. L. DeRisi, V. R. IyeT, and P. O. Brown, "Exploring the Metabolic and Genetic Control of Gene Expreession on a Genomic Scale". Science 278, 680-686. [3] A Connectionist Model of Development, E. Mjolsness, D. H. Sharp, and J. Reinitz, Journal of Theoretical Biology 152:429-453, 1991. [4] J. Reinitz, E. Mjolsness, and D. H. Sharp, "Model for Cooperative Control of Positional Information in Drosophila by Bicoid and Maternal Hunchback". 1. Experimental Zoology 271:47-56, 1995. Los Alamos National Laboratory Technical Report LAUR-92-2942 1992. [5] 1. Reinitz and D. H. Sharp, "Mechanism of eve Stripe Formation". Mechanisms of Development 49: 133-158, 1995. [6] [7] J. Lam and 1. M. Delosme. "An Efficient Simulated Annealing Schedule: Derivation" and" ... Implementation and Evaluation". Technical Reports 8816 and 8817, Yale University Electrical Engineering Department, New Haven CT 1988. [8] x. Wen, S. Fuhrman, G. S. Michaels, D. B. Carr, S. Smith, J. L. Barker, and R. Somogyi, "Large-Scale Temporal Gene Expression Mapping of Central Nervous System Development", Proc. Natl. Acal. Sci. USA 95:334-339, January 1998. [9] M. B. Eisen, P. T. Spellman, P. O. Brown, and D. Botstein, "Cluster Analysis and Display of Genome-Wide Expression Patterns", Proc. Natl. Acad. Scie. USA 95:1486314868, December 1998. [10] 1. Hertz, lecture at Krogerrup Denmark computational biology summer school, July 1998. [11] Spellman PT, Sherlock G, Zhang MQ, et aI., "Comprehensive identification of cell cycle-regulated genes of the yeast Saccharomyces cerevisiae by microarray hybridization", Mol. Bio. Cell. 9: (12) 3273-3297 Dec 1998. [12] Dempster, A. P., Laird, N. M. and Rubin, D. B. "Maximum likelihood from incomplete data via the EM algorithm," 1. Royal Statistical Society, Series B, 39:1-38,1977. [13] P. Smyth, "Clustering using Monte Carlo Cross-Validation", Proceedings of the 2 nd International Conference on Knowledge Discovery and Data Mining, AAAI Press, 1996.
1706 |@word nd:1 proportionality:1 simulation:1 covariance:1 excited:2 pick:2 thereby:1 minus:1 solid:1 reduction:2 initial:1 series:2 score:10 genetic:1 existing:3 current:2 od:1 activation:1 must:1 plot:3 drop:1 guess:1 nervous:1 beginning:1 smith:1 provides:1 sigmoidal:1 zhang:1 five:1 provisional:1 scie:1 direct:1 symposium:1 assayed:1 prove:1 fitting:3 wild:1 theoretically:1 expected:1 roughly:2 examine:1 spherical:2 gov:3 actual:1 little:1 becomes:1 provided:1 bounded:1 circuit:26 lowest:2 interpreted:1 substantially:1 developed:1 differentiation:1 temporal:1 every:1 control:2 unit:2 normally:1 bio:1 positive:2 before:1 dropped:1 bind:1 engineering:1 xv:1 acad:1 might:1 plus:1 twice:1 co:1 range:3 averaged:4 unique:2 acknowledgment:1 testing:1 practice:1 procedure:3 protein:7 get:1 put:1 applying:1 conventional:1 center:5 mrna:5 go:3 barker:1 resolution:3 formulate:1 array:2 deriving:1 mq:1 increment:3 altman:1 imagine:1 gm:1 pt:1 smyth:2 losing:1 hypothesis:2 stripe:1 labeled:1 cooperative:1 fly:1 electrical:1 thousand:2 region:1 cycle:3 sun:1 mjolsness:7 decrease:1 prospect:1 removed:1 mentioned:1 substantial:1 dempster:1 reward:1 tobias:1 dynamic:2 division:2 eric:1 translated:1 joint:2 hopfield:1 various:3 derivation:1 effective:1 monte:1 aggregate:8 formation:1 choosing:1 nematode:1 quite:1 widely:1 valued:1 stanford:1 otherwise:1 statistic:1 laird:1 sequence:1 ero:1 took:1 lam:2 coming:1 product:4 maximal:2 caenorhabditis:1 j2:1 turned:1 loop:1 los:1 convergence:1 cluster:39 requirement:1 produce:1 depending:1 recurrent:2 derive:2 measured:5 ij:2 school:1 expectationmaximization:1 sa:2 strong:1 involves:1 indicate:1 synchronized:1 concensus:1 subsequently:1 mann:5 require:1 clustered:4 really:1 preliminary:2 drosophila:2 biological:6 extension:1 exploring:1 normal:1 exp:1 presumably:1 visually:1 mapping:1 circuitry:4 automate:1 vary:1 commonality:1 purpose:1 estimation:1 proc:2 currently:1 genomic:1 rna:1 cerevisiae:4 gaussian:5 modified:1 rather:1 varying:1 office:1 validated:4 derived:1 naval:1 improvement:1 saccharomyces:3 mutant:1 likelihood:18 manduchi:1 kim:3 inference:11 membership:2 entire:2 integrated:1 pasadena:4 hidden:2 manipulating:1 among:5 development:3 spatial:2 biology:3 represents:1 broad:1 stuart:2 future:1 report:3 connectionist:1 quantitatively:1 inhibited:2 few:3 haven:1 wen:1 national:1 comprehensive:2 individual:1 phase:1 highly:1 mining:1 evaluation:1 mixture:1 zoology:1 natl:2 encourage:1 shorter:1 incomplete:1 logarithm:2 circle:2 theoretical:1 minimal:1 modeling:1 measuring:1 deviation:3 entry:1 alamo:1 successful:1 reported:1 synthetic:4 fundamental:1 peak:1 international:1 stay:1 systematic:1 off:1 contract:1 enhance:1 together:1 michael:1 reflect:1 central:1 aaai:1 choose:2 transcribed:2 unmatched:2 external:1 return:2 account:1 aggressive:1 dunker:1 summarized:1 includes:1 coefficient:1 vi:3 performed:1 lab:1 overfits:1 ass:1 ir:1 likewise:1 preprocess:1 bayesian:1 grn:1 identification:1 hunter:1 carlo:1 trajectory:2 processor:1 simultaneous:1 plateau:4 messenger:1 influenced:1 strongest:1 ed:1 against:1 sampled:1 proved:1 knowledge:1 dimensionality:3 schedule:1 garrett:1 nasa:4 higher:1 dt:1 subclusters:1 botstein:1 wold:4 just:1 horizontal:2 nonlinear:2 a7:1 yeast:3 scientific:1 grows:1 usa:2 normalized:2 concept:1 brown:2 laboratory:4 nonzero:1 assay:2 i2:1 satisfactory:1 illustrated:1 during:1 complete:1 carr:1 temperature:1 image:2 recently:1 charles:1 becky:1 analog:3 discussed:1 interpretation:1 occurred:1 measurement:1 significant:2 ai:2 longer:2 similarity:1 own:1 showed:1 optimizes:1 barbara:1 caltech:1 scoring:2 fortunately:1 somewhat:1 repress:1 george:1 employed:1 aggregated:1 maximize:1 determine:1 july:1 ii:1 multiple:2 infer:3 technical:2 jet:3 cross:7 long:1 divided:1 justifying:1 coexpression:1 prediction:4 variant:2 basic:1 enhancing:1 represent:1 microscopy:1 cell:8 dec:1 preserved:1 addition:2 affecting:1 want:1 fine:1 annealing:8 ftl:1 derisi:1 microarray:1 unlike:1 hybridization:6 member:1 elegans:2 december:1 seem:2 adequacy:1 eve:1 automated:1 affect:1 fit:15 gave:1 forthcoming:1 laur:1 reduce:1 microarrays:1 i7:1 expression:20 penalty:1 reinitz:4 iterating:1 governs:1 dna:1 acal:1 generate:2 supplied:1 arising:1 per:1 klein:1 diverse:1 four:2 salient:1 run:13 antibody:2 acceptable:2 comparable:1 ct:1 summer:1 display:1 yale:1 adapted:1 strength:3 aspect:1 ining:1 relatively:3 pacific:1 department:1 combination:1 hertz:2 across:1 slightly:1 em:8 modification:1 making:1 dv:1 taken:3 discus:2 turn:1 mechanism:3 available:2 gaussians:2 doll:1 multiplied:1 apply:1 eight:2 hierarchical:1 regulate:1 appropriate:1 robustly:1 alternative:1 clustering:16 testable:1 society:1 objective:1 move:1 concentration:2 diagonal:1 transcriptional:7 regulated:1 simulated:6 sci:1 propulsion:3 cellular:1 denmark:1 length:1 code:2 modeled:2 relationship:3 ratio:1 optionally:2 regulation:13 negative:2 implementation:1 unknown:1 budding:1 vertical:2 displayed:1 january:1 extended:1 sharp:5 inferred:2 rebecca:1 coregulated:1 required:1 nooo:1 extensive:1 connection:12 optimized:1 california:4 quadratically:1 hour:1 below:1 pattern:2 sparsity:1 summarize:1 program:1 sherlock:1 including:2 royal:1 suitable:1 force:1 synchronize:1 advanced:1 representing:2 spellman:2 altered:2 technology:5 unselected:1 confocal:1 axis:4 coupled:1 roberto:1 discovery:1 fully:1 expect:1 lecture:1 generation:1 limitation:1 proven:1 validation:3 foundation:1 nucleus:1 degree:1 fruit:1 rubin:1 metabolic:1 course:21 supported:1 understand:1 institute:4 wide:1 sparse:1 distributed:1 feedback:2 curve:3 world:1 genome:1 eisen:1 forward:1 made:1 preprocessing:1 melanogaster:1 alpha:1 transcription:2 gene:63 continuous:2 regulatory:14 iterative:1 supercluster:1 search:1 reality:1 table:2 nature:2 molecule:1 ca:4 robust:1 mol:1 requested:2 hierarchically:1 linearly:1 noise:2 complementary:1 site:1 inferring:5 third:1 departs:1 specific:1 decay:8 jpl:1 evidence:1 essential:2 simply:2 explore:1 likely:3 maternal:1 positional:1 expressed:1 partially:1 scalar:1 monotonic:1 aa:1 towards:1 analysing:2 determined:1 except:1 reducing:1 averaging:1 called:1 specie:1 worm:1 experimental:4 player:2 biocomputing:1
772
1,707
An MEG Study of Response Latency and Variability in the Human Visual System During a Visual-Motor Integration Task Akaysha C. Tang Dept. of Psychology University of New Mexico Albuquerque, NM 87131 akaysha@unm.edu Barak A. Pearlmutter Dept. of Computer Science University of New Mexico Albuquerque, NM 87131 bap@cs. unm. edu Tim A. Hely Santa Fe Institute 1399 Hyde Park Road Santa Fe, NM 87501 Michael Zibulevsky Dept. of Computer Science University of New Mexico Albuquerque, NM 87131 Michael P. Weisend VA Medical Center 1501 San Pedro SE Albuquerque, NM 87108 timhely@santafe. edu michael@cs.unm.edu mweisend@unm.edu Abstract Human reaction times during sensory-motor tasks vary considerably. To begin to understand how this variability arises, we examined neuronal populational response time variability at early versus late visual processing stages. The conventional view is that precise temporal information is gradually lost as information is passed through a layered network of mean-rate "units." We tested in humans whether neuronal populations at different processing stages behave like mean-rate "units". A blind source separation algorithm was applied to MEG signals from sensory-motor integration tasks. Response time latency and variability for multiple visual sources were estimated by detecting single-trial stimulus-locked events for each source. In two subjects tested on four visual reaction time tasks, we reliably identified sources belonging to early and late visual processing stages. The standard deviation of response latency was smaller for early rather than late processing stages. This supports the hypothesis that human populational response time variability increases from early to late visual processing stages. 1 Introduction In many situations, precise timing of a motor output is essential for successful task completion. Somehow the reliability in the output timing is related to the reliability of the underlying neural systems associated with different stages of processing. Recent literature from animal studies suggests that individual neurons from different brain regions and different species can be surprising reliable [1, 2, 5, 7-9, 14, 17, 18], A. C. Tang, B. A. Pearlmutter. T. A. Hely, M. Zibulevsky and M P. Weisend 186 on the order of a few milliseconds. Due to the low spatial resolution of electroencephalography (EEG) and the requirement of signal averaging due to noisiness of magnetoencephalography (MEG), in vivo measurement of human populational response time variability from different processing stages has not been available. In four visual reaction time (RT) tasks, we estimated neuronal response time variability at different visual processing stages using MEG. One major obstacle that has prevented the analysis of response timing variability using MEG before is the relative weakness of the brain's magnetic signals (lOOf!') compared to noise in a shielded environment (magnetized lung contaminants: 106 f!'j abdominal currents lO5f!'j cardiogram and oculogram: 104 f!'j epileptic and spontaneous activity: lO3f!') and in the sensors (10fT) [13]. Consequently, neuronal responses evoked during cognitive tasks often require signal averaging across many trials, making analysis of singletrial response times unfeasible. Recently, Bell-Sejnowski Infomax [1995] and Fast ICA [10] algorithms have been used successfully to isolate and remove major artifacts from EEG and MEG data [11, 15, 20] . These methods greatly increase the effective signal-to-noise ratio and make single-trial analysis of EEG data feasible [12]. Here, we applied a SecondOrder Blind Identification algorithm (SOBI) [4] (another blind source separation, or BSS, algorithm) to MEG data to find out whether populational response variability changes from early to late visual processing stages. 2 2.1 Methods Experimental Design Two volunteer normal subjects (females, right handed) with normal or correctedto-normal visual acuity and binocular vision participated in four different visual RT tasks. Subjects gave informed consent prior to the experimental procedure. During each task we recorded continuous MEG signals at a 300Hz sampling rate with a band-pass filter of I-100Hz using a 122 channel Neuromag-122. In all four tasks, the subject was presented with a pair of abstract color patterns, one in the left and the other in the right visual field. One of the two patterns was a target pattern. The subject pressed either a left or right mouse button to indicate on which side the target pattern was presented. When a correct response was given, a low or high frequency tone was presented binaurally following respectively a correct or wrong response. The definition of the target pattern varied in the four tasks and was used to control task difficulty which ranged from easy (task 1) to more difficult (task 4) with increasing RTs. (The specific differences among the four tasks are not important for the analysis which follows and are not discussed further.) In this study we focus on the one element that all tasks have in common, Le. activation of multiple visual areas along the visual pathways. Our goal is to identify visual neuronal sources activated in all four visual RT tasks and to measure and compare response time variability between neuronal sources associated with early and later visual processing stages. Specifically, we test the hypothesis that populational neuronal response times increase from early to later visual processing stages. 2.2 Source Separation Using SOBI In MEG, magnetic activity from different neuronal populations is observed by many senSOrs arranged around the subject's head. Each sensor responds to a mixture of the signals emitted by multiple sources. We used the Second-Order Blind Identi- MEG Study ofResponse Latency and Variability 187 fication algorithm (SOBI) [4] (a BSS algorithm) to simultaneously separate neuromagnetic responses from different neuronal populations associated with different stages of visual processing. Responses from different neuronal populations will be referred to as source responses and the neuronal populations that give rise to these responses will be referred to as neuronal sources or simply sources. These neuronal sources often, but not always, consist of a spatially contiguous population of neurons. BSS separates the measured sensor signals into maximally independent components, each having its own spatial map. Previously we have shown that some of these BSS separated components correspond to noise sources, and many others correspond to neuronal sources [19]. To establish the identity of the components, we analyzed both temporal and spatial properties of the BSS separated components. Their temporal properties are displayed using MEG images, similar to the ERP images described by [12] but without smoothing across trials. These MEG images show stimulus or response locked responses across many trials in a map, from which response latencies across all displayed trials can be observed with a glance. The spatial properties of the separated components are displayed using a field map that shows the sensor projection of a given component. The intensity at each point on the field map indicates how strongly this component influences the sensor at this location. The correspondence between the separated components and neuronal populational responses at different visual processing stages were established by considering both spatial and temporal properties of the separated components [19]. For example, a component was identified as an early visual neuronal source if and only if (1) the field pattern, or the sensor projection, of the separated component showed a focal response over the occipital lobe, and (2) the ERP image showed visual stimulus locked responses with latencies shorter than all other visual components and falling within the range of early visual responses reported in studies using other methods. Only those components consistent both spatially and temporally with known neurophysiology and neuroanatomy were identified as neuronal sources. 2.3 Single Event Detection and Response Latency Estimation For all established visual components we calculated the single-trial response latency as follows. First, a detection window was defined using the stimulus-triggered average (STA). The beginning of the detection window was defined by the time at which the STA first exceeded the range of baseline fluctuation. Baseline fluctuation was estimated from the time of stimulus onset for approximately 50ms (the visual response occurred no earlier than 60ms after stimulus onset.) The detection window ended when the STA first returned to the same level as when the detection window began. The detection threshold was determined using a control window with the same width as the detection window, but immediately preceding the detection window. The threshold was adjusted until no more than five false detections occurred within the control window for each ninety trials. We estimated RTs using the leading edge of the response, rather than the time of the peak as this is more robust against noise. 3 Results In both subjects across all four visual RT tasks, SOBI generated components that corresponded to neuronal populational responses associated with early and late stages of visual processing. In both subjects, we identified a single component with a sensor projection at the occipital lobe whose latency was the shortest among all 188 1 2 3 4 1 2 3 4 A. C. Tang, B. A. Pearlmutter, T. A. Hely, M Zibulevsky and M P. Weisend e ??? e ??? e ??? e ??? e ??? e ??? e ??? e ??? late source ? ??? e ??? -- . a? ? ??? e ??? e ??? e ??? e ??? Figure 1: MEG images and field maps for an early and a late source from each task, for subject 1 (top) and subject 2 (bottom). MEG image pixels are brightnesscoded source strength. Each row of a bitmap is one trial, running 1170ms from left to right. Vertical bars mark stimulus onset, and 333ms of pre-stimulus activity is shown. Each panel contains 90 trials. Field map brightness indicates the strength with which a source activates each of the 61 sensor pairs. visual stimulus locked components within task and subject (Fig. 1 left). We identified multiple components that had sensor projections either at occipital-parietal, occipital-temporal, or temporal lobes, and whose response latencies are longer than early-stage components within task and subject (Fig. 1 right). Fig. 2a shows examples of detected single-trial responses for one early and one late visual component (left: early; right: late) from one task. To minimize false positives, the detection threshold was set high (allowing 5 false detections out of 90 trials) at the expense of a low detection rate (15%- 67%.) When Gaussian filters were applied to the raw separated data, the detection rates were increased to 22- 91% (similar results hold but not shown) . Fig. 2b shows such detected response time histograms superimposed on the stimulus triggered average using raw separated data. One early (top row) and two late visual components (middle and bottom rows) are plotted for each of the four experiments in subject one. The histogram width is smallest for early visual components (short mean response latency) and larger for late visual components (longer latency.) We computed the standard deviation of component response times as a measure of response variability. Fig. 2c shows the response variability as a function of mean response latency for subject one. Early components (solid boxes, shorter mean latency) have smaller variability (height of the boxes) while late components (dashed boxes, longer mean latency) have larger variability (height of the boxes) . Multiple 189 MEG Study of Response Latency and Variability :bd:Qlld~~ -'~[d llJ l~J l~J -- -- -- -l~J ~~ I!] ' : ~ -- -- -. -. a 100 200 300 40Q 0 100 200 300 40Q a 100 ~- 200 300 400 0 100 ~~ 200 300 400 -~ o 100 200 300 40CI C 100 200 300 40CI 0 100 200 300 400 0 100 200 300 .wo o 100 200 300 400 0 100 200 300 400 0 100 200 300 400 0 100 200 300 400 40 \ i : : ! : .-----------: ,------, i 1 1. ____ 1 100 150 200 latency (ms) Figure 2: (a, left) Response onset was estimated for each trial via threshold crossing within a window of eligibility. (b, top right) The stimulus-locked averages for a number of sources overlaid on histograms of response onset times. (c, bottom right) Scatter plot of visual components from all experiments on subject 1 showing the standard deviation of the latency (y axis) versus the mean latency (x axis), with the error bars in each direction indicating one standard error in the respective measurement. Lines connect sources from each task. visual components from each task are connected by a line. Four tasks were shown here. There is a general trend of increasing standard deviation of response times as a function of early-to-late processing stages (increasing mean latency from left to right). For the early visual components the standard deviation ranges from 6.6?0.63ms to 13.4?1.23ms, and for the late visual components , from 9.9?0.86ms to 38.8?3.73ms (t = 3.565, p = 0.005.) 4 Discussion By applying SOBl to MEG data from four visual RT tasks, we separated components corresponding to neuronal populational responses associated with early and 190 A. C. Tang, B. A. Pearlmutter; T. A. Hely, M Zibulevsky and M P. Weisend later stage visual processing in both subjects across all tasks. We performed singletrial RT detection on these early- and late-stage components and estimated both the mean and stdev of their response latency. We found that variability of the populational response latency increased from early to late processing stages. These results contrast with single neuron recordings obtained previously. In early and late visual processing stages, the rise time of mean firing rate in single units remained constant, suggesting an invariance in RT variability [16]. Characterizing the precise relationship between single neuron and populational response reliability is difficult without careful simulations or simultaneous single unit and MEG recording. However, some major differences exist between the two types of studies. While MEG is more likely to sample a larger neuronal population, single unit studies are more likely to be selective to those neurons that are already highly reliable in their responses to stimulus presentation. It is possible that the most reliable neurons at both the early and late processing stages are equally reliable while large differences exist between the early and late stages for the low reliability neurons. Previously, ICA algorithms have been used successfully to separate out various noise and neuronal sources in MEG data [19, 20J. Here we show that SOBI can also be used to separate different neuronal sources, particularly those associated with different processing stages. The SOBI algorithm assumes that the components are independent across multiple time scales and attempts to minimize the temporal correlation at these time scales. Although neuronal sources at different stages of processing are not completely independent as assumed in SOBl's derivation, BSS algorithms of this sort are quite robust even when the underlying assumptions are not fully met [6J, i.e. the goodness of the separation is not significantly affected. The ultimate reality check should come from satisfying physiological and anatomical constraints derived from prior knowledge of the neural system under study. This was carried out for our analysis. Firstly, the average response latencies of the separated components fell within the range of latencies reported in MEG studies using conventional source modeling methods. Secondly, the spatial patterns of sensor responses to these separated components are consistent with the known functional anatomy of the visual system. We have attempted to rule out many confounding factors. Our observed results cannot be accounted for by a higher signal to noise ratio in the early visual responses. The increase in measured onset response time variability from early to late visual processing stages was actually accompanied by an slightly lower signalto-noise ratio among the early components. The number of events detected for the later components were also slightly greater than the earlier components. The higher signal-to-noise ratio at later components should reduce noise-induced variability in the later components, which would bias against the hypothesis that later visual responses have greater response time variability. We also found that response duration and detection window size cannot account for the observed differential variabilities. Later visual responses also had gentler onset slopes (as measured by the stimulustriggered average). Sensor noise unavoidably introduces noise into the response onset detection process. We cannot rule out the possibility that the interaction of the noise with the response onset profiles might give rise to the observed differential variabilities. Similarly, we cannot rule out the possibility that even greater control of the experimental situation, such as better fixation and more effective head restraints, would differentially reduce the observed variabilities. In general, all measured variabilities can only be upper bounds, subject to downward revision as improved instrumentation and experiments become available. It is with this caution in mind that we conclude that response time variability of neuronal populations increases from early to late processing stages in the human visual system. MEG Study of Response Latency and Variability 191 Acknowledgments This research was supported by NSF CAREER award 97-02-311, and by the National Foundation for Functional Brain Imaging. References [1] M. Abeles, H. Bergman, E. Margalit, and E Vaadia. Spatiotemporal firing patterns in the frontal cortex of behaving monkeys. J. Neurophys., 70:1629-1638, 1993. [2] W. Bair and C. Koch. Temporal precision of spike trains in extrastriate cortex of the behaving macaque monkey. Neural Computation, 8(6):1184-1202, 1996. [3] A. J. Bell and T. J . Sejnowski. An information-maximization approach to blind separation and blind deconvolution. Neural Computation, 7(6):1129-1159, 1995. [4] A. Belouchrani, K. A. Meraim, J.-F. Cardoso, and E. Moulines. Second-order blind separation of correlated sources. In Proc. Int. ConJ. on Digital Sig. Proc., pages 346-351, Cyprus, 1993. [5) M. J. Berry, W. K. Warland, and M. Meister. The structure and precision of retinal spike trains. Proc. Natl. Acad. Sci. USA, 94:5411-5416, 1997. [6] J .-F. Cardoso. Blind signal separation: statistical principles. Proceedings of the IEEE, 9(10):2009-2025, October 1998. [7] R. R . de Ruyter van Steveninck, G. D. Lewen, S. P. Strong, R. Koberle, and W. Bialek. Reproducibility and variability in neural spike trains. Science, 275:1805-1808, 1997. [8] R. C. deCharms and M. M. Merzenich. Primary cortical representation of sounds by the coordination of action-potential timing. Nature, 381:610-3, 1996. [9] M. Gur, A. Beylin, and D. M. Snodderly. Response variability of neurons in primary visual cortex (VI) of alert monkeys. J. Neurosci., 17(8):2914-2920, 1997. [10] A. Hyvarinen and E. Oja. A fast fixed-point algorithm for independent component analysis. Neural Computation, 9(7), October 1997. [11] T.-P. Jung, C. Humphries, T.-W. Lee, M. J. McKeown, V. Iragui, S. Makeig, and T. J. Sejnowski. Removing electroencephalographic artifacts by blind source separation. Psychophysiology, 1999. In Press. [12] T.-P. Jung, S. Makeig, M. Westerfield, J. Townsend, E. Courchesne, and T. J. Sejnowski. Analyzing and visualizing single-trial event-related potentials. In Advances in Neural Infonnation Processing Systems 11, pages 118-124. MIT Press, 1999. [13] J. D. Lewine and W. W. Orrison, II. Magnetoencephalography and magnetic source imaging. In Functional Brain Imaging, pages 369-417. Mosby, St. Louis, 1995. [14] Z. F. Mainen and T. J . Sejnowski. Reliability of spike timing in neocortical neurons. Science, 268:1503-1506, 1995. [15] S. Makeig, T.-P. Jung, A. J. Bell, D. Ghahremani, and T . J . Sejnowski. Blind separation of auditory event-related brain responses into independent components. Proc. Nat. A cad. Sci., 94:10979-84, 1997. [16] P. Marsalek, C. Koch, and J. Maunsell. On the relationship between synaptic input and spike output jitter in individual neurons. Proc. Natl. Acad. Sci., 94:735-40, 1997. [17] D. S. Reich, J. D. Victor, B. W. Knight, and T . Ozaki. Response variability and timing precision of neuronal spike trains in vivo. J. Neurophys., 77:2836-2841, 1997. [18] A. C. Tang, A. M. Bartels, and T. J . Sejnowksi. Effects of cholinergic modulation on responses of neocortical neurons to fluctuating inputs. Cereb. Cortex, 7:502-9, 1997. [19] A. C. Tang, B. A. Pearlmutter, M. Zibulevsky, and R. Loring. Response time variability in the human sensory and motor systems. In Computational Neuroscience, 1999. To appear as a special issue of Neurocomputing. [20] R. Vigario, V. Jousmaki, M. Hamruainen, R. Hari, and E. Oja. Independent component analysis for identification of artifacts in magnetoencephalographic recordings. In Advances in Neural Infonnation Processing Systems 10. MIT Press, 1998.
1707 |@word neurophysiology:1 trial:14 middle:1 cyprus:1 simulation:1 lobe:3 brightness:1 pressed:1 solid:1 extrastriate:1 contains:1 mainen:1 reaction:3 bitmap:1 current:1 neurophys:2 surprising:1 cad:1 activation:1 scatter:1 bd:1 motor:5 remove:1 plot:1 tone:1 rts:2 beginning:1 short:1 detecting:1 location:1 firstly:1 five:1 height:2 along:1 alert:1 differential:2 become:1 ozaki:1 magnetoencephalographic:1 fixation:1 pathway:1 westerfield:1 ica:2 brain:5 moulines:1 window:10 electroencephalography:1 increasing:3 considering:1 begin:1 revision:1 underlying:2 panel:1 monkey:3 informed:1 caution:1 ended:1 temporal:8 makeig:3 wrong:1 control:4 unit:5 medical:1 maunsell:1 appear:1 louis:1 before:1 positive:1 timing:6 acad:2 analyzing:1 fluctuation:2 firing:2 approximately:1 modulation:1 might:1 examined:1 evoked:1 suggests:1 locked:5 range:4 steveninck:1 acknowledgment:1 lost:1 procedure:1 area:1 bell:3 significantly:1 projection:4 pre:1 road:1 shielded:1 unfeasible:1 cannot:4 layered:1 influence:1 applying:1 conventional:2 map:6 humphries:1 center:1 occipital:4 duration:1 resolution:1 immediately:1 rule:3 fication:1 courchesne:1 gur:1 population:8 spontaneous:1 target:3 snodderly:1 hypothesis:3 secondorder:1 bergman:1 sig:1 element:1 crossing:1 trend:1 particularly:1 satisfying:1 observed:6 ft:1 bottom:3 region:1 cardiogram:1 connected:1 zibulevsky:5 binaurally:1 knight:1 environment:1 neuromagnetic:1 santafe:1 completely:1 various:1 stdev:1 derivation:1 train:4 separated:11 fast:2 effective:2 sejnowski:6 detected:3 corresponded:1 whose:2 quite:1 larger:3 triggered:2 vaadia:1 interaction:1 unavoidably:1 singletrial:2 consent:1 reproducibility:1 differentially:1 requirement:1 mckeown:1 tim:1 completion:1 measured:4 strong:1 c:2 indicate:1 come:1 met:1 direction:1 anatomy:1 correct:2 filter:2 human:7 require:1 hyde:1 secondly:1 adjusted:1 hold:1 around:1 koch:2 normal:3 overlaid:1 major:3 vary:1 early:29 smallest:1 estimation:1 proc:5 infonnation:2 coordination:1 successfully:2 neuromag:1 mit:2 sensor:12 always:1 activates:1 gaussian:1 rather:2 derived:1 acuity:1 focus:1 noisiness:1 indicates:2 superimposed:1 check:1 greatly:1 contrast:1 electroencephalographic:1 baseline:2 akaysha:2 margalit:1 bartels:1 selective:1 sobi:6 pixel:1 issue:1 among:3 animal:1 spatial:6 integration:2 smoothing:1 special:1 field:6 having:1 sampling:1 park:1 others:1 stimulus:12 few:1 sta:3 oja:2 simultaneously:1 national:1 neurocomputing:1 individual:2 attempt:1 detection:16 abdominal:1 restraint:1 highly:1 possibility:2 cholinergic:1 weakness:1 introduces:1 mixture:1 analyzed:1 activated:1 natl:2 edge:1 conj:1 shorter:2 respective:1 plotted:1 handed:1 increased:2 earlier:2 obstacle:1 modeling:1 contiguous:1 goodness:1 maximization:1 deviation:5 successful:1 reported:2 connect:1 spatiotemporal:1 abele:1 considerably:1 st:1 peak:1 lee:1 lewine:1 contaminant:1 infomax:1 michael:3 mouse:1 nm:5 recorded:1 cognitive:1 leading:1 suggesting:1 account:1 potential:2 de:1 accompanied:1 retinal:1 int:1 blind:10 onset:9 vi:1 later:8 view:1 performed:1 lung:1 sort:1 slope:1 vivo:2 minimize:2 correspond:2 identify:1 identification:2 raw:2 albuquerque:4 meraim:1 weisend:4 simultaneous:1 synaptic:1 definition:1 against:2 frequency:1 associated:6 auditory:1 color:1 knowledge:1 actually:1 exceeded:1 higher:2 psychophysiology:1 response:65 maximally:1 improved:1 arranged:1 box:4 strongly:1 stage:26 binocular:1 until:1 correlation:1 glance:1 somehow:1 artifact:3 usa:1 effect:1 ranged:1 merzenich:1 spatially:2 visualizing:1 during:4 width:2 eligibility:1 gentler:1 m:9 neocortical:2 pearlmutter:5 cereb:1 image:6 recently:1 began:1 common:1 functional:3 discussed:1 occurred:2 measurement:2 focal:1 similarly:1 had:2 reliability:5 reich:1 longer:3 cortex:4 behaving:2 own:1 recent:1 female:1 showed:2 confounding:1 instrumentation:1 victor:1 greater:3 preceding:1 neuroanatomy:1 shortest:1 signal:11 dashed:1 ii:1 multiple:6 sound:1 prevented:1 equally:1 award:1 va:1 vision:1 vigario:1 volunteer:1 histogram:3 participated:1 source:30 fell:1 subject:17 isolate:1 hz:2 recording:3 induced:1 emitted:1 easy:1 psychology:1 gave:1 identified:5 reduce:2 whether:2 bair:1 epileptic:1 ultimate:1 passed:1 wo:1 returned:1 action:1 latency:25 santa:2 se:1 cardoso:2 band:1 exist:2 nsf:1 millisecond:1 estimated:6 neuroscience:1 anatomical:1 affected:1 four:11 threshold:4 falling:1 erp:2 imaging:3 button:1 jitter:1 separation:9 bound:1 correspondence:1 activity:3 strength:2 constraint:1 llj:1 belonging:1 smaller:2 bap:1 across:7 ninety:1 slightly:2 making:1 gradually:1 previously:3 mind:1 meister:1 available:2 fluctuating:1 magnetic:3 top:3 running:1 assumes:1 warland:1 establish:1 already:1 spike:6 primary:2 rt:7 responds:1 bialek:1 separate:4 sci:3 meg:21 relationship:2 ratio:4 ghahremani:1 mexico:3 difficult:2 october:2 fe:2 expense:1 decharms:1 rise:3 design:1 reliably:1 allowing:1 upper:1 vertical:1 neuron:11 behave:1 displayed:3 parietal:1 situation:2 variability:31 precise:3 head:2 varied:1 jousmaki:1 intensity:1 pair:2 identi:1 established:2 macaque:1 bar:2 pattern:8 iragui:1 reliable:4 event:5 difficulty:1 townsend:1 temporally:1 axis:2 carried:1 koberle:1 prior:2 literature:1 berry:1 lewen:1 relative:1 fully:1 versus:2 digital:1 foundation:1 consistent:2 principle:1 beylin:1 row:3 populational:10 accounted:1 supported:1 belouchrani:1 jung:3 side:1 bias:1 understand:1 barak:1 institute:1 characterizing:1 van:1 bs:6 calculated:1 cortical:1 sensory:3 san:1 hyvarinen:1 unm:4 hari:1 assumed:1 conclude:1 continuous:1 reality:1 channel:1 nature:1 robust:2 ruyter:1 career:1 eeg:3 neurosci:1 hely:4 noise:12 profile:1 neuronal:25 fig:5 referred:2 precision:3 late:22 tang:6 removing:1 remained:1 specific:1 showing:1 physiological:1 deconvolution:1 essential:1 consist:1 false:3 ci:2 nat:1 downward:1 signalto:1 simply:1 likely:2 visual:48 pedro:1 goal:1 identity:1 magnetoencephalography:2 consequently:1 careful:1 presentation:1 feasible:1 change:1 specifically:1 determined:1 averaging:2 specie:1 pas:1 invariance:1 experimental:3 attempted:1 indicating:1 support:1 mark:1 arises:1 frontal:1 dept:3 tested:2 correlated:1
773
1,708
The Parallel Problems Server: an Interactive Tool for Large Scale Machine Learning Charles Lee Isbell, Jr. Parry Husbands isbell @research.att.com AT&T Labs 180 Park Avenue Room A255 Florham Park, NJ 07932-0971 PIRHusbands@lbl.gov Lawrence Berkeley National LaboratorylNERSC 1 Cyclotron Road, MS 50F Berkeley, CA 94720 Abstract Imagine that you wish to classify data consisting of tens of thousands of examples residing in a twenty thousand dimensional space. How can one apply standard machine learning algorithms? We describe the Parallel Problems Server (PPServer) and MATLAB*P. In tandem they allow users of networked computers to work transparently on large data sets from within Matlab. This work is motivated by the desire to bring the many benefits of scientific computing algorithms and computational power to machine learning researchers. We demonstrate the usefulness of the system on a number of tasks. For example, we perform independent components analysis on very large text corpora consisting of tens of thousands of documents, making minimal changes to the original Bell and Sejnowski Matlab source (Bell and Sejnowski, 1995). Applying ML techniques to data previously beyond their reach leads to interesting analyses of both data and algorithms. 1 Introduction Real-world data sets are extremely large by the standards of the machine learning community. In text retrieval, for example, we often wish to process collections consisting of tens or hundreds of thousands of documents and easily as many different words. Naturally, we would like to apply machine learning techniques to this problem; however, the sheer size of the data makes this difficult. This paper describes the Parallel Problems Server (PPServer) and MATLAB *P. The PPServer is a "linear algebra server" that executes distributed memory algorithms on large data sets. Together with MATLAB*P, users can manipulate large data sets within Matlab transparently. This system brings the efficiency and power of highly-optimized parallel computation to researchers using networked machines but maintain the many benefits of interactive environments. We demonstrate the usefulness of the PPServer on a number of tasks. For example, we perform independent components analysis on very large text corpora consisting of tens of thousands of documents with minimal changes to the original Bell and Sejnowski Matlab source (Bell and Sejnowski, 1995). Applying ML techniques to datasets previously beyond C. L. Isbell, Jr. and P. Husbands 704 Libraries Computational & Interface Routines Machine) Machin~, MatlabS Local Variables ml . . . [ ??? .. ? ??????l m, ... [ ........ ..... m, ... [ ...... .. m, ... [............. Figure 1: Use of the PPServer by Matlab is almost completely transparent. PPServer variables are tied to the PPServer itself while Matlab maintains handles to the data. Using Matlab's object system, functions using PPServer variables invoke PPServer commands implicitly. their reach, we discover interesting analyses of both data and algorithms. 2 The Parallel Problems Server The Parallel Problems Server (PPServer) is the foundation of this work. The PPServer is a realization of a novel client-server model for computation on very large matrices. It is compatible with any Unix-like platform supporting the Message Passing Interface (MPI) library (Gropp, Lusk and Skjellum, 1994). MPI is the standard for multi-processor communication and is the most portable way for writing parallel code. The PPServer implements functions for creating and removing distributed matrices, loading and storing them from/to disk using a portable format, and performing elementary matrix operations. Matrices are two-dimensional single or double precision arrays created on the PPServer itself (functions are provided for transferring matrix sections to and from a client) . The PPServer supports both dense and sparse matrices. The PPServer communicates with clients using a simple request-response protocol. A client requests an action by issuing a command with the appropriate arguments, the server executes that command, and then notifies the client that the command is complete. The PPServer is directly extensible via compiled libraries called packages. The PPServer implements a robust protocol for communicating with packages. Clients (and other packages) can load and remove packages on-the-fty, as well as execute commands within packages. Package programmers have direct access to information about the PPServer and its matrices. Each package represents its own namespace, defining a set of visible function names. This supports data encapsulation and allows users to hide a subset of functions in one package by loading another that defines the same function names. Finally, packages support common parallel idioms (eg applying a function to every element of a matrix), making it easier to add common functionality. All but a few PPServer commands are implemented in packages, including basic matrix operations. Many highly-optimized public libraries have been realized as packages using appropriate wrapper functions. These packages include ScaLAPACK (Blackford et a1., 1997), S3L (Sun's optimized version of ScaLAPACK), PARPACK (Maschhoff and Sorensen, 1996), and PETSc (PETSc, ). Large Scale Machine Learning Using The Parallel Problems Server 1 function H=hilb(n) l:nj 2 J 3 J = J (ones (n, 1) , : ) I J' j 4 5 E = ones (n, n) j 6 H = E. / (I+J-1) j 705 j Figure 2: Matlab code for producing Hilbert matrices. When n is influenced by P, each of the constructors creates a PPServer object instead of a Matlab object. 3 MATLAB*P By directly using the PPServer's client communication interface, it is possible for other applications to use the PPServer's functionality. We have implemented a client interface for Matlab, called MA1LAB*P. MATLAB*P is a collection of Matlab 5 objects, Matlab m-files (Matlab's scripting language) and Matlab MEX programs (Matlab's extemallanguage API) that allows for the transparent integration of Matlab as a front end for the Parallel Problems Server. The choice of Matlab was influenced by several factors. It is the de facto standard for scientific computing, enjoying wide use in industry and academia. In the machine learning community, for example, algorithms are often written as Matlab scripts and made freely available. In the scientific computing community, algorithms are often first prototyped in Matlab before being optimized for languages such as Fortran. We endeavor to make interaction with the PPServer as transparent as possible for the user. In principle, a typical Matlab user should never have to make explicit calls to the PPServer. Further, current Matlab programs should not have to be rewritten to take advantage of the PPServer. Space does not permit a complete discussion of MA1LAB*P (we refer the reader to (Husbands and Isbell, 1999)); however, we will briefly discuss how to use prewritten Matlab scripts without modification. This is accomplished through the simple but innovative P notation. We use Matlab 5's object oriented features to create PPServer objects automatically. P is a special object we introduce in Matlab that acts just like the integer 1. A user typing a=ones (1000*P, 1000) or b=rand( 1000, 1000*P) obtains two 1000-by-lOOO matrices distributed in parallel. The reader can guess the use of P here: it indicates that a is distributed by rows and b by columns. To a user, a and b are matrices, but within Matlab, they are handles to special distributed types that exist on the PPServer. Any further references to these variables (e.g. via such commands as eig, svd, inv, *, +, -) are recognized as a call to the PPServer rather than as a traditional Matlab command. Figure 2 shows the code for Matlab's built in function hilb. The call hilb (n) produces the n x n Hilbert matrix (H ij = i+}-l)' When n is influenced by P, a parallel array results: ? J=l: n in line 2 creates the PPServer vector 1,2, ? . . , n and places a handle to it in J. Note that this behavior does not interfere with the semantics of for loops (for i=l: n) as Matlab assigns to i the value of each column of 1: n: the numbers 1,2, ... , n. ? ones (n, 1) in line 3 produces a PPServer matrix. ? Emulation of Matlab's indexing functions results in the correct execution of line 3. C. L. Isbell, Jr. and P. Husbands 706 ? Overloading of ' (the transpose operator) executes line 4 on the PPServer. ? In line 5, E is generated on the PPServer because ofthe overloading of ones. ? Overloading elementary matrix operations makes H a PPServer matrix (line 6). The Parallel Problems Server and MA1LAB *p have been tested extensively on a variety of platforms. They currently run on Cray supercomputers! , clusters of symmetric multiprocessors from Sun Microsystems and DEC as well as on clusters of networked Intel PCs. The PPServer has also been tested with other clients, including Common LISP. Although computational performance varies depending upon the platform, it is clear that the system provides distinct computational advantages. Communication overhead (in our experiments, roughly two milliseconds per PPServer command) is negligible compared to the computational and space advantage afforded by transparent access to highly-optimized linear algebra algorithms. 4 Applications in Text Retrieval In this section we demonstrate the efficacy of the PPServer on real-world machine learning problems. In particular we explore the use of the PPServer and MA1LAB*P in the text retrieval domain. The task in text retrieval is to find the subset of a collection of documents relevant to a user's information request. Standard approaches are based on the Vector Space Model (VSM). A document is a vector where each dimension is a count of the occurrence of a different word. A collection of documents is a matrix, D, where each column is a document vector di. The similarity between tw~ documents is their inner product, dj. Queries are just like documents, so the relevance of documents to a query, q, is DT q. Jf Typical small collections contain a thousand vectors in a ten thousand dimensional space, while large collections may contain 500,000 vectors residing in hundreds of thousands of dimensions. Clearly, well-understood standard machine learning techniques may exhibit unpredictable behavior under such circumstances, or simply may not scale at all. Classically, ML-like approaches try to construct a set of linear operators which extract the underlying "topic" structure of documents. Documents and queries are projected into that new (usually smaller) space before being compared using the inner product. The large matrix support in MA1LAB*P enables us to use matrix decomposition techniques for extracting linear operators easily. We have explored several different algorithms(Isbell and Viola, 1998). Below, we discuss two standard algorithms to demonstrate how the PPServer allows us to perform interesting analysis on large datasets. 4.1 Latent Semantic Indexing Latent Semantic Indexing (LSI) (Deerwester et al., 1990) constructs -a smaller document matrix by using the Singular Value Decomposition (SVD): D = U SV T . U contains the eigenvectors of the co-occurrence matrix while the diagonal elements of S (referred to as singular values) contain the square roots of their corresponding eigenValUes. The eigenvectors with the largest eigenvalues capture the axes of largest variation in the data. LSI projects documents onto the k-dimensional subspace spanned by the first k columns of U (denoted Uk) so thatthe documents are now: V[ = S;;lUk. Queries are similarly projected. Thus, the document-query scores for LSI can be obtained with simple Matlab code: 1 Although there is no Matlab for the Cray, we are still able to use it to "execute" Matlab code -in parallel. 707 Large Scale Machine Learning Using The Parallel Problems Server '" .. . IG . ,, ------------ - ---- - - Figure 3: The first 200 singular values of a collection of about 500,000 documents and 200,000 terms, and singular values for half of that collection. Computation for on the full collection took only 62 minutes using 32 processors on a Cray TIE. D=dsparse('term-doc'); %D SPARSE reads a sparse matrix Q=dsparse('queries'); [U,S,VJ=svds(D,k); % compute the k-SVD of D sc=getlsiscores(U,S,V,Q); % computes v*(l/s)*u'*q The scores that are returned can then be combined with relevance judgements to obtain precision/recall curves that are displayed in Matlab: r=dsparse('judgements'); [pr,reJ=precisionrecall(sc,r); plot (re ( , @' ) , pr ( , @' ) ) ; In addition to evaluating the performance of various techniques, we can also explore characteristics of the data itself. For example, many implementations of LSI on large collections use only a subset of the documents for computational reasons. This leads one to question how the SVD is affected. Figure 3 shows the first singular values for one large collection as well as for a random half of that collection. It shows that the shape of the curves are remarkably similar (as they are for the other half). This suggests that we can derive a projection matrix from just half of the collection. An evaluation of this technique can easily be performed using our system. Prernlinary experiments show nearly identical retrieval performance. 4.2 What are the Independent Components of Documents? Independent components analysis (ICA)(Bell and Sejnowski, 1995) also recovers linear projections from data. Unlike LSI, which finds principal components, ICA finds axes that are statistically independent. ICA's biggest success is probably its application to the blind source separation or cocktail party problem. In this problem, one observes the output of a number of microphones. Each microphone is assumed to be recording a linear mixture of a number of unknown sources. The task is to recover the original sources. There is a natural embedding of text retrieval within this framework. The words that are observed are like microphone signals, and underlying ''topics'' are the source signals that give rise to them. Figure 4 shows a typical distribution of words projected along axes found by ICA. 2 Most words have a value close to zero. The histogram shows only the words large positive or 2These results are from a collection containing transcripts of White House press releases from 1993. There are 1585 documents and 18,675 distinct words. C. L. Isbell, Jr. and P. Husbands 708 africa apartheid I -1 anc transition mandela continent elite ethiopia saharan -0.75 0.5 0.75 Figure 4: Distribution of words with large magnitude an ICA axis from White House text. negative values. One group of words is made up of highly-related terms; namely, "africa," "apartheid," and "mandela." The other group of words are not directly related, but each cooccurs with different individual words in the first group. For example, "saharan" and "africa" occur together many times, but not in the context of apartheid and South Africa; rather, in documents concerning US policy toward Africa in general. As it so happens, "saharan" acts as a discriminating word for these SUbtopics. As observed in (Isbell and Viola, 1998), it appears that ICA is finding a set of words, S, that selects for related documents, H, along with another set of words, T, whose elements do not select for H, but co-occur with elements of S. Intuitively, S selects for documents in a general subject area, and T removes a specific subset of those documents, leaving a small set of highly related documents. This suggests a straightforward algorithm to achieve the same goal directly. This local clustering approach is similar to an unsupervised version of Rocchio with Query Zoning (Singhal, 1997). Further analysis of ICA on similar collections reveals other interesting behavior on large datasets. For example, it is known that ICA will attempt to find an unmixing matrix that is full rank. This is in conflict with the notion that these collections actually reside in a much smaller subspace. We have found in our experiments with ICA that some axes are highly kurtotic while others produce gaussian-like distributions. We conjecture that any axis that results in a gaussian-like distribution will be split arbitrarily among all "empty" axes. For all intents and purposes, these axes are uninformative. This provides an automatic noisereduction technique for ICA when applied to large datasets. For the purposes of comparison, Figure 5 illustrates the performance of several algorithms (including ICA and various clustering techniques) on articles from the Wall Street Journal.3 5 Discussion We have shown that MATLAB *p enables portable, high-performance interactive supercomputing using the Parallel Problems Server, a powerful mechanism for writing and accessing optimized algorithms. Further, the client communication protocol makes it possible to implement transparent integration with sufficiently powerful clients, such as Matlab 5. With such a tool, researchers can now use Matlab as something more than just a way for prototyping algorithms and working on small problems. MATLAB*P makes it possible to interactively operate on and visualize large data sets. We have demonstrated this last claim by using the PPServer system to apply ML techniques to large datasets, allowing for analyses of both data and algorithms. MATLAB*P has also been used to implement versions of Diverse Density(Maron, 1998), MIMIC(DeBonet, Isbell and Viola, 1996), and gradient descent. 3The WSJ collection contains 42,652 documents and 89,757 words Large Scale Machine Learning Using The Parallel Problems Server - - - _ . - - T.... " " - 709 .71~~~-~;:=======il - 0.0 - --- l.SI ~_a.... ....... DocuMnIIt_ a....... teA ", Recall Figure 5: A comparison of different algorithms on the Wall Street Journal References Bell, A. and Sejnowski, T. (1995). An infonnation-maximizaton approach to blind source separation and blind deconvolution. Neural Computation, 7:1129-1159. Blackfor<~, L. S., Choi, J., Cleary, A, D' Azevedo, E., Demmel, J., Dhilon, I., Dongarra, 1., Hammarling, S. , Henry, G., Petitet, A, Stanley, K, Walker, D., and Whaley, R. (1997). ScaLAPACK Users' Guide. http://www.netlib.orglscalapacklsluglscalapack..slug.htrnl. DeBonet, J., Isbell, C., and Viola, P. (1996). Mimic: Finding optima by estimating probability densities. In Advances in Neural Information Processing Systems. Deerwester, S., Dumais, S. T., Landauer, T. K, Furnas, G. w., and Harshman, R. A (1990). Indexing by latent semantic analysis. Journal of the Society for Information Science, 41(6):391-407. Frakes, W. B. and Baeza-Yates, R., editors (1992). Information Retrieval: Data Structures and Algorithms. Prentice-Hall. Gropp, W., Lusk, E., and Skjellum, A (1994). Using MPI: Portable Parallel Programming with the Message-Passing Interface. The MIT Press. Husbands, P. and Isbell, C. (1999). MITMatlab: A tool for interactive supercomputing. In Proceedings of the Ninth SIAM Conference on Parallel Processingfor Scientific Computing. Isbell, C. and Viola, P. (1998). Restructuring sparse high dimensional data for effective retrieval. In Advances in Neural Information Processing Systems. Kwok, K L. (1996). A new method of weighting query tenns for ad-hoc retrieval. In Proceedings of the 19th ACMlSIGIR Conference,pages 187-195. Maron, O. (1998). A framework for multiple-instance learning. In Advances in Neural Information Processing Systems . Maschhoff, K J. and Sorensen, D. C. (1996). A Portable Implementation of ARPACK for Distributed Memory Parallel Computers. In Preliminary Proceedings of the Copper Mountain Conference on Iterative Methods. O'Brien, G. W. (1994). Infonnation management tools for updating an svd-encoded indexing scheme. Technical Report UT-CS-94-259, University of Tennessee. PETSc. The Portable, Extensible http://www.mcs.anl.gov/home/group/petsc.htrnl. Toolkit for Scientific Computation. PPServer. The Parallel Problems Server Web Page. http://www.ai.mit.edulprojects/ppserver. Sahami, M ., Hearst, M., and Saund, E. (1996). Applying the multiple cause mixture model to text categorization. In Proceedings of the 13th International Machine Learning Conference. Singhal, A (1997). Learning routing queries in a query zone. In Proceedings of the 20th International Conference on Research and Development in Information Retrieval.
1708 |@word luk:1 briefly:1 version:3 judgement:2 loading:2 disk:1 decomposition:2 cleary:1 wrapper:1 contains:2 att:1 efficacy:1 score:2 document:26 africa:5 brien:1 current:1 com:1 si:1 issuing:1 written:1 visible:1 academia:1 shape:1 enables:2 remove:2 plot:1 half:4 guess:1 provides:2 noisereduction:1 along:2 direct:1 cray:3 overhead:1 introduce:1 ica:11 roughly:1 behavior:3 multi:1 automatically:1 gov:2 unpredictable:1 tandem:1 provided:1 discover:1 notation:1 underlying:2 project:1 estimating:1 what:1 mountain:1 finding:2 nj:2 berkeley:2 every:1 act:2 interactive:4 tie:1 facto:1 uk:1 producing:1 harshman:1 before:2 negligible:1 understood:1 local:2 positive:1 api:1 saharan:3 suggests:2 co:2 statistically:1 blackford:1 implement:4 precisionrecall:1 area:1 bell:6 projection:2 word:15 road:1 onto:1 close:1 operator:3 prentice:1 context:1 applying:4 writing:2 www:3 demonstrated:1 straightforward:1 elite:1 assigns:1 communicating:1 array:2 spanned:1 embedding:1 handle:3 notion:1 variation:1 constructor:1 imagine:1 machin:1 user:9 programming:1 element:4 updating:1 observed:2 whaley:1 capture:1 thousand:8 svds:1 sun:2 observes:1 accessing:1 environment:1 algebra:2 creates:2 upon:1 efficiency:1 completely:1 easily:3 various:2 distinct:2 describe:1 effective:1 sejnowski:6 demmel:1 query:10 sc:2 whose:1 encoded:1 slug:1 florham:1 itself:3 hoc:1 advantage:3 eigenvalue:2 took:1 interaction:1 product:2 networked:3 loop:1 realization:1 relevant:1 achieve:1 double:1 cluster:2 empty:1 optimum:1 produce:3 unmixing:1 wsj:1 categorization:1 object:7 depending:1 derive:1 ij:1 transcript:1 implemented:2 c:1 emulation:1 correct:1 functionality:2 routing:1 programmer:1 public:1 maschhoff:2 transparent:5 wall:2 preliminary:1 elementary:2 sufficiently:1 residing:2 hall:1 lawrence:1 visualize:1 claim:1 rocchio:1 purpose:2 currently:1 infonnation:2 largest:2 create:1 tool:4 mit:2 clearly:1 gaussian:2 rather:2 tennessee:1 command:9 ax:6 release:1 rank:1 indicates:1 multiprocessor:1 transferring:1 selects:2 semantics:1 among:1 denoted:1 development:1 platform:3 integration:2 special:2 construct:2 never:1 identical:1 represents:1 park:2 unsupervised:1 nearly:1 mimic:2 others:1 report:1 few:1 oriented:1 national:1 individual:1 consisting:4 maintain:1 attempt:1 message:2 highly:6 evaluation:1 mixture:2 pc:1 sorensen:2 enjoying:1 lbl:1 re:1 minimal:2 instance:1 classify:1 industry:1 column:4 kurtotic:1 extensible:2 singhal:2 subset:4 hundred:2 usefulness:2 front:1 encapsulation:1 apartheid:3 varies:1 sv:1 combined:1 dumais:1 density:2 international:2 siam:1 discriminating:1 lee:1 invoke:1 together:2 interactively:1 containing:1 management:1 classically:1 creating:1 de:1 blind:3 ad:1 script:2 try:1 root:1 lab:1 performed:1 saund:1 recover:1 maintains:1 parallel:21 square:1 il:1 characteristic:1 subtopics:1 ofthe:1 mc:1 researcher:3 executes:3 processor:2 reach:2 influenced:3 husband:6 naturally:1 di:1 recovers:1 recall:2 ut:1 stanley:1 hilbert:2 routine:1 actually:1 appears:1 dt:1 response:1 rand:1 execute:2 just:4 thatthe:1 working:1 web:1 eig:1 interfere:1 defines:1 brings:1 maron:2 scientific:5 name:2 contain:3 read:1 symmetric:1 semantic:3 eg:1 white:2 mpi:3 m:1 complete:2 demonstrate:4 continent:1 bring:1 interface:5 novel:1 charles:1 common:3 refer:1 frakes:1 ai:1 automatic:1 similarly:1 language:2 dj:1 henry:1 toolkit:1 access:2 similarity:1 compiled:1 add:1 something:1 own:1 hide:1 server:15 success:1 arbitrarily:1 tenns:1 accomplished:1 gropp:2 freely:1 recognized:1 signal:2 full:2 multiple:2 technical:1 retrieval:10 concerning:1 manipulate:1 a1:1 basic:1 circumstance:1 histogram:1 mex:1 dec:1 addition:1 remarkably:1 uninformative:1 singular:5 source:7 leaving:1 walker:1 operate:1 unlike:1 file:1 probably:1 recording:1 south:1 subject:1 lisp:1 call:3 integer:1 extracting:1 split:1 baeza:1 variety:1 scalapack:3 inner:2 avenue:1 motivated:1 returned:1 passing:2 cause:1 action:1 matlab:43 cocktail:1 mandela:2 clear:1 eigenvectors:2 parry:1 ten:5 extensively:1 http:3 debonet:2 exist:1 lsi:5 millisecond:1 transparently:2 per:1 diverse:1 tea:1 affected:1 yates:1 group:4 sheer:1 deerwester:2 run:1 package:12 unix:1 you:1 powerful:2 hammarling:1 dongarra:1 place:1 almost:1 reader:2 separation:2 home:1 doc:1 occur:2 isbell:12 afforded:1 fty:1 argument:1 extremely:1 innovative:1 performing:1 format:1 conjecture:1 request:3 jr:4 describes:1 smaller:3 tw:1 making:2 modification:1 happens:1 intuitively:1 indexing:5 pr:2 previously:2 discus:2 count:1 mechanism:1 fortran:1 sahami:1 end:1 available:1 operation:3 rewritten:1 permit:1 apply:3 kwok:1 appropriate:2 occurrence:2 notifies:1 supercomputer:1 rej:1 original:3 clustering:2 include:1 society:1 question:1 realized:1 traditional:1 diagonal:1 exhibit:1 gradient:1 subspace:2 street:2 topic:2 portable:6 reason:1 toward:1 code:5 difficult:1 negative:1 rise:1 intent:1 implementation:2 policy:1 twenty:1 perform:3 unknown:1 allowing:1 datasets:5 descent:1 displayed:1 supporting:1 defining:1 scripting:1 communication:4 viola:5 ninth:1 hearst:1 community:3 inv:1 copper:1 namely:1 optimized:6 conflict:1 beyond:2 able:1 microsystems:1 usually:1 below:1 prototyping:1 program:2 built:1 including:3 memory:2 power:2 typing:1 client:11 natural:1 scheme:1 library:4 axis:2 created:1 extract:1 text:9 interesting:4 foundation:1 article:1 principle:1 editor:1 storing:1 row:1 compatible:1 looo:1 last:1 transpose:1 guide:1 allow:1 wide:1 sparse:4 benefit:2 distributed:6 curve:2 dimension:2 world:2 evaluating:1 transition:1 computes:1 reside:1 collection:17 made:2 projected:3 ig:1 supercomputing:2 party:1 obtains:1 implicitly:1 arpack:1 ml:5 reveals:1 corpus:2 assumed:1 anl:1 landauer:1 latent:3 iterative:1 robust:1 ca:1 anc:1 protocol:3 domain:1 hilb:3 vj:1 dense:1 intel:1 referred:1 biggest:1 vsm:1 furnas:1 precision:2 wish:2 explicit:1 house:2 tied:1 communicates:1 weighting:1 removing:1 minute:1 ethiopia:1 load:1 specific:1 choi:1 explored:1 deconvolution:1 overloading:3 magnitude:1 execution:1 illustrates:1 easier:1 simply:1 explore:2 petitet:1 desire:1 restructuring:1 prototyped:1 netlib:1 goal:1 endeavor:1 room:1 jf:1 change:2 typical:3 principal:1 microphone:3 called:2 svd:5 zone:1 select:1 support:4 relevance:2 tested:2
774
1,709
Bayesian averaging is well-temperated Lars Kai Hansen Department of Mathematical Modelling Technical University of Denmark B321 DK-2800 Lyngby, Denmark lkhansen@imm .dtu.dk Abstract Bayesian predictions are stochastic just like predictions of any other inference scheme that generalize from a finite sample. While a simple variational argument shows that Bayes averaging is generalization optimal given that the prior matches the teacher parameter distribution the situation is less clear if the teacher distribution is unknown. I define a class of averaging procedures, the temperated likelihoods, including both Bayes averaging with a uniform prior and maximum likelihood estimation as special cases. I show that Bayes is generalization optimal in this family for any teacher distribution for two learning problems that are analytically tractable: learning the mean of a Gaussian and asymptotics of smooth learners. 1 Introduction Learning is the stochastic process of generalizing from a random finite sample of data. Often a learning problem has natural quantitative measure of generalization. If a loss function is defined the natural measure is the generalization error, i.e., the expected loss on a random sample independent of the training set. Generalizability is a key topic of learning theory and much progress has been reported. Analytic results for a broad class of machines can be found in the litterature [8, 12, 9, 10] describing the asymptotic generalization ability of supervised algorithms that are continuously parameterized. Asymptotic bounds on generalization for general machines have been advocated by Vapnik [11]. Generalization results valid for finite training sets can only be obtained for specific learning machines, see e.g. [5]. A very rich framework for analysis of generalization for Bayesian averaging and other schemes is defined in [6]. A veraging has become popular as a tool for improving generalizability of learning machines . In the context of (time series) forecasting averaging has been investigated intensely for decades [3]. Neural network ensembles were shown to improve generalization by simple voting in [4] and later work has generalized these results to other types of averaging. Boosting, Bagging, Stacking, and Arcing are recent examples of averaging procedures based on data resampling that have shown useful see [2] for a recent review with references. However, Bayesian averaging in particular is attaining a kind of cult status. Bayesian averaging is indeed provably optimal in a L. K. Hansen 266 number various ways (admissibility, the likelihood principle etc) [1]. While it follows by construction that Bayes is generalization optimal if given the correct prior information, i.e., the teacher parameter distribution, the situation is less clear if the teacher distribution is unknown. Hence, the pragmatic Bayesians downplay the role of the prior. Instead the averaging aspect is emphasized and "vague" priors are invoked. It is important to note that whatever prior is used Bayesian predictions are stochastic just like predictions of any other inference scheme that generalize from a finite sample. In this contribution I analyse two scenarios where averaging can improve generalizability and I show that the vague Bayes average is in fact optimal among the averaging schemes investigated. Averaging is shown to reduce variance at the cost of introducing bias, and Bayes happens to implement the optimal bias-variance trade-off. 2 Bayes and generalization Consider a model that is smoothly parametrized and whose predictions can be described in terms of a density function 1 . Predictions in the model are based on a given training set: a finite sample D = {Xa}~=l of the stochastic vector x whose density - the teacher - is denoted p(xIOo). In other words the true density is assumed to be defined by a fixed, but unknown, teacher parameter vector 00 . The model, denoted H, involves the parameter vector and the predictive density is given by p(xID, H) = ! ? p(xIO, H)p(OID, H)dO (1) p(OID, H) is the parameter distribution produced in training process. In a maximum likelihood scenario this distribution is a delta function centered on the most likely parameters under the model for the given data set. In ensemble averaging approaches, like boosting bagging or stacking, the distribution is obtained by training on resampled traning sets. In a Bayesian scenario, the parameter distribution is the posterior distribution, p(DIO, H)p(OIH) p(OID, H) = f p(DIO', H)p(O'IH)dO' (2) where p(OIH) is the prior distribution (probability density of parameters if D is empty). In the sequel we will only consider one model hence we suppress the model conditioning label H. The generalization error is the average negative log density (also known as simply the "log loss" - in some applied statistics works known as the "deviance") r(DIOo) = ! -logp(xID)p(xIOo)dx, (3) The expected value of the generalization error for training sets produced by the given teacher is given by f(Oo) = !! -logp(xID)p(xIOo)dxp(DIOo)dD. (4) lThis does not limit us to conventional density estimation; pattern recognition and many functional approximations problems can be formulated as density estimation problems as well. 267 Bayesian Averaging is Well-Temperated Playing the game of "guessing a probability distribution" [6] we not only face a random training set, we also face a teacher drawn from the teacher distribution p( Bo) . The teacher averaged generalization must then be defined as r= J f(Bo)p(Bo)dBo . (5) This is the typical generalization error for a random training set from the randomly chosen teacher - produced by the model H. The generalization error is minimized by Bayes averaging if the teacher distribution is used as prior. To see this, form the Lagrangian functional ?[q(xID)] = JJJ -logq(xID)p(xIBo)dxp(DIBo)dDp(Bo)dBo+A J q(xID)dx (6) defined on positive functions q(xID). The second term is used to ensure that q(xID) is a normalized density in x . Now compute the variational derivative to obtain 6? 1 6q(xID) = - q(xID) J p(xIBo)p(DIBo)p(Bo)dBo + A. (7) Equating this derivative to zero we recover the predictive distribution of Bayesian averagmg, p(DIB)p(B) (8) q(xID) = p(xIB) Jp(DIB')p(B')dB' dB, J =J where we used that A p(DIB)p(B)dB is the appropriate normalization constant. It is easily verified that this is indeed the global minimum of the averaged generalization error. We also note that if the Bayes average is performed with another prior than the teacher distribution p( Bo), we can expect a higher generalization error . The important question from a Bayesian point of view is then: Are there cases where averaging with generic priors (e.g. vague or uniform priors) can be shown to be optimal? 3 Temperated likelihoods To come closer to a quantative statement about when and why vague Bayes is the better procedure we will analyse two problems for which some analytical progress is possible. We will consider a one-parameter family of learning procedures including both a Bayes and the maximum likelihood procedure, v(DIB) p(BI!3,D,H) = Jpf3(DIB')dB" (9) where !3 is a positive parameter (plying the role of an inverse temperature). The family of procedures are all averaging procedures, and !3 controls the width of the average. Vague Bayes (here used synonymously with Bayes with a uniform prior) is recoved for !3 = 1, while the maximum posterior procedure is obtained by cooling to zero width !3 --+ 00 . In this context the generalization design question can be frased as follows : is there an optimal temperature in the family of the temperated likelihoods? 3.1 Example: ID normal variates Let the teacher distribution be given by p(xIBo) = ~exp (-~(X 211"<72 2<7 - Bo)2) (10) L.K. Hansen 268 The model density is of the same form with (J unknown and u 2 assumed to be known. For N examples the posterior (with a uniform prior) is, p(OID) = J 2:U2 exp (- ::2 (11) (x - (J)2) , = with x 1/ N Eo: Xo:. The temperated likelihood is obtained by raising to the ,8'th power and normalizing, p((JID,,8) = Vf7iN ~ exp (,8N - 2u 2 (x - (J)2 ) . (12) The predictive distribution is found by integrating w.r.t. (J, p(xID,,8) = !P(ZIB)P(BID,~)dB; ~exp (--212 (x- X)2) 21!'u$ uf3 , (13) with u~ = u 2(1+1/,8N). We note that this distribution is wider for all the averaging procedures than it is for maximum likelihood (,8 -T (0), i.e., less variant. For very small ,8 the predictive distribution is almost independent of the data set, hence highly biased. It is straightforward to compute the generalization error of the predictive distribution for general,8. First we compute the generalization error for the specific training set D, r(D,,8, (Jo) = ! -logp(xID, ,8)p(xl(Jo)dx = log )21!'u$ + 2~~ ((x - (JO)2 + ( 2) , (14) The average generalization error is then found by averaging w.r.t the sampling distribution using x"" N((Jo, u 2/N)., r(,8) = ! r(D, ,8)dDp(DI(Jo) = log )21!'u$ + 2:$ (~ + 1) , (15) We first note that the generalization error is independent of the teacher (Jo parameter, this happened because (J is a "location" parameter. The ,8-dependency of the averaged generalization error is depicted in Figure 1. Solving 8r(,8) /8,8 = 0 we find that the optimal ,8 solves u$=U2(,8~+I)=U2(~+I) :=} ,8=1 (16) Note that this result holds for any N and is independent of the teacher parameter. The Bayes averaging at unit temperature is optimal for any given value of (Jo, hence, for any teacher distribution. We may say that the vague Bayes scheme is robust to the teacher distribution in this case. Clearly this is a much stronger optimality than the more general result proven above. 3.2 Bias-variance tradeoff It is interesting to decompose the generalization error in Eq. 15 in bias and variance components. We follow Heskes [7] and define the bias error as the generalization error of the geometric average distribution, B(,8) ! = -logp(x)p(xl(Jo)dx, (17) 269 Bayesian Averaging is Wel/-Temperated 0.7 0.& GENERALIZATION 0.5 A 1? 04 ?, v 03 02 0.1 0 0 0.5 15 2 25 3 35 45 TEMPERATURE Figure 1: Bias-variance trade-off as function of the width of the temperated likelihood ensemble (temperature = 1/(3) for N = 1. The bias is computed as the generalization error of the predictive distribution obtained from the geometric average distribution w.r.t. training set fluctuations as proposed by Heskes. The predictive distribution produced by Bayesian averaging corresponds to unit temperature (vertical line) and it achieves the minimal generalization error. Maximum-likelihood estimation for reference is recovered as the zero width/temperature limit. with p(x) = Z-l exp ( / 10g(P(X 1D)]P(D I 80 )dD) . (18) Inserting from Eq. (13), we find p(z) = ~exp (-~(X 27r0'~ 0'f3 - 0)2) . 8 (19) Integrating over the teacher distribution we find, 1 B(f3) = -2 log 27r0'~ 0'2 + -2 20'f3 (20) The variance error is given by V(f3) = r(f3) - B(f3) , 0'2 V (f3) = 2N O'~ (21) We can now quantify the statements above. By averaging a bias is introduced -the predictive distribution becomes wider- which decrease the variance contribution initially so that the generalization error being the sum of the two decreases. At still higher temperatures the bias becomes too strong and the generalization error start to increase. The Bayes average at unit temperature is the optimal trade-off within the given family of procedures. 270 3.3 L. K. Hansen Asymptotics for smoothly parameterized models We now go on to show that a similar result also holds for general learning problems in limit of large data sets. We consider a system parameterized by a finite dimensional parameter vector O. For a given large training set and for a smooth likelihood function, the temperated likelihood is approximately Gaussian centered at the maximum posterior parameters[13]' hence the normalized temperated posterior reads P(OI(3D,H) = I(3NA(~OML) lexp (_(3; 60'A(D,OML)60) (22) where 60 = O-OML, with OML = OML(D) denoting the maximum likelihood solution for the given training sample. The second derivative or Hessian matrix is given by 1 N A(D,O) N LA(xa,O) (23) a=l {)2 A(x,O) = ()O{)O' - log p( x 10 ) (24) p(xIO)p(OI(3, D)dO (25) The predictive distribution is given by p(xl(3, D) = ! we write p(xIO) = exp(-f(xIO)) and expand f(xIO) around OML to second order, we find p(xIO) ~ p(XIOML) exp (-a(xIOML)'60 - ~60' A(xIOML)60) . (26) We are then in position to perform the integration over the posterior to find the normalized predictive distribution, p(xl(3, D) = p(XIOML) I(3N A(D)I 1 , I(3NA(D) + A(x)1 exp ( 2'a(xIOML) A(xIOML)a(xIOML)). (27) Proceeding as above, we compute the generalization error f((3, ( 0 ) = !! -logp(xl(3, D)p(xIOo)dxp(DIOo)dD (28) For sufficiently smooth likelihoods, fluctuations in the maximum likelihood parameters will be asymptotic normal, see e.g. [8], and furthermore fluctuations in A(D) can be neglected, this means that we can approximate, A(x) + A(D) ~ (~ + l)Ao, Ao = ! A(xIOo)p(xIOo)dx (29) where Ao is the averaged Fisher information matrix. With these approximations (valid as N --+ (0) the generalization error can be found, 1) d 1+ ~ d ( f((3, ( 0 ) ~ f(oo) + 2 log 1 + (3N - 21 + (3N' (30) with d = dim(O) denoting the dimension of the parameter vector. Like in the ID example (Eq. (15)) we find the generalization error is asymptotically independent of the teacher parameters. It is minimized for (3 1 and we conclude that Bayes is well-temperated in the asymptotics and that this holds for any teacher distribution. In the Bayes literature this is refered to as the prior is overwhelmed by data [1]. Decomposing the errors in bias and variance contributions we find similar results as for in ID example, Bayes introduces the optimal bias by averaging at unit temperature. = Bayesian Averaging is Well-Temperated 4 271 Discussion We have seen two examples of Bayes averaging being optimal, in particular improving on maximum likelihood estimation. We found that averaging introduces a bias and reduces variance so that the generalization error (being the sum of bias and variance) initially decrease. Bayesian averaging at unit temperature is the optimal width of the averaging distribution. For larger temperatures (widths) the bias is too strong and the generalization error increases. Both examples were special in the sense that they lead to generalization errors that are independent of the random teacher parameter. This is not generic, of course, rather the generic case is that a mis-specified prior can lead to arbitrary large learning catastrophes. Acknowledgments I thank the organizers of the 1999 Max Planck Institute Workshop on Statistical Physics of Neural Networks Michael Biehl, Wolfgang Kinzel and Ido Kanter, where this work was initiated. I thank Carl Edward Rasmussen, Jan Larsen, and Manfred Opper for stimulating discussions on_Bayesian averaging. This work was funded by the Danish Research Councils through the Computational Neural Network Center CONNECT and the THOR Center for Neuroinformatics. References [1] C.P. Robert: The Bayesian Choice - A Decision- Theoretic Motivation. Springer Texts in Statistics, Springer Verlag, New York (1994). A. Ohagan: Bayesian Inference. Kendall's Advanced Theory of Statistics. Vol 2B. The University Press, Cambridge (1994). [2] L. Breiman: Using adaptive bagging to debias regressions. Technical Report 547, Statistics Dept. U.C. Berkeley, (1999) . [3] R.T. Clemen Combining forecast: A review and annotated bibliography. Journal of Forecasting 5, 559 (1989). [4] L.K. Hansen and P. Salamon: Neural Network Ensembles. IEEE Transactions on Pattern Analysis and Machine Intelligence, 12, 993-1001 (1990). [5] L.K . Hansen: Stochastic Linear Learning: Exact Test and Training Error Averages. Neural Networks 6, 393-396, (1993) [6] D. Haussler and M. Opper: Mutual Information, Metric Entropy, and Cumulative Relative Entropy Risk Annals of Statistics 25 2451-2492 (1997) [7] T . Heskes: Bias/Variance Decomposition for Likelihood-Based Estimators. Neural Computation 10, pp 1425-1433, (1998) . [8] L. Ljung: System Identification: Theory for the User. Englewood Cliffs, New Jersey: Prentice-Hall, (1987). [9] J . Moody: "Note on Generalization, Regularization, and Architecture Selection in Nonlinear Learning Systems," in B.H. Juang, S.Y. Kung & C.A. Kamm (eds.) Proceedings of the first IEEE Workshop on Neural Networks for Signal Processing, Piscataway, New Jersey: IEEE, 1-10, (1991). [10] N . Murata, S. Yoshizawa & S. Amari: Network Information Criterion - Determining the Number of Hidden Units for an Artificial Neural Network Model. IEEE Transactions on Neural Networks, vol. 5, no. 6, pp. 865-872 , 1994. [11] V. Vapnik: Estimation of Dependences Based on Empirical Data. Springer-Verlag New York (1982). [12] H . White, "Consequences and Detection of Misspecified Nonlinear Regression Models," Journal of the American Statistical Association, 76(374), 419-433, (1981). [13] D .J .C MacKay: Bayesian Interpolation, Neural Computation 4, 415-447, (1992) .
1709 |@word stronger:1 decomposition:1 series:1 denoting:2 recovered:1 dx:5 must:1 oml:6 analytic:1 resampling:1 intelligence:1 cult:1 manfred:1 boosting:2 location:1 mathematical:1 become:1 indeed:2 expected:2 kamm:1 becomes:2 kind:1 quantitative:1 berkeley:1 voting:1 whatever:1 control:1 unit:6 planck:1 positive:2 limit:3 consequence:1 id:3 initiated:1 cliff:1 fluctuation:3 interpolation:1 approximately:1 equating:1 bi:1 averaged:4 acknowledgment:1 implement:1 procedure:10 jan:1 asymptotics:3 empirical:1 word:1 deviance:1 integrating:2 selection:1 prentice:1 context:2 risk:1 conventional:1 lagrangian:1 center:2 straightforward:1 go:1 clemen:1 estimator:1 haussler:1 annals:1 construction:1 user:1 exact:1 carl:1 recognition:1 cooling:1 logq:1 role:2 trade:3 decrease:3 downplay:1 xio:6 neglected:1 solving:1 predictive:10 debias:1 learner:1 vague:6 easily:1 various:1 jersey:2 artificial:1 neuroinformatics:1 whose:2 kanter:1 kai:1 larger:1 biehl:1 say:1 amari:1 ability:1 statistic:5 analyse:2 dxp:3 analytical:1 inserting:1 combining:1 juang:1 empty:1 wider:2 oo:2 advocated:1 progress:2 eq:3 strong:2 edward:1 solves:1 involves:1 come:1 quantify:1 correct:1 annotated:1 lars:1 stochastic:5 centered:2 xid:13 dio:2 ao:3 generalization:38 decompose:1 hold:3 around:1 sufficiently:1 hall:1 normal:2 exp:9 achieves:1 estimation:6 label:1 hansen:6 council:1 tool:1 zib:1 clearly:1 gaussian:2 rather:1 dib:5 breiman:1 arcing:1 modelling:1 likelihood:18 sense:1 dim:1 inference:3 initially:2 hidden:1 expand:1 provably:1 among:1 denoted:2 special:2 integration:1 mutual:1 mackay:1 f3:7 sampling:1 broad:1 synonymously:1 minimized:2 report:1 randomly:1 detection:1 englewood:1 highly:1 introduces:2 closer:1 minimal:1 logp:5 stacking:2 cost:1 introducing:1 uniform:4 too:2 reported:1 dependency:1 connect:1 teacher:23 generalizability:3 ido:1 traning:1 density:10 sequel:1 off:3 physic:1 michael:1 continuously:1 moody:1 na:2 jo:8 american:1 derivative:3 attaining:1 later:1 performed:1 view:1 wolfgang:1 kendall:1 start:1 recover:1 bayes:20 contribution:3 oi:2 variance:11 ensemble:4 murata:1 generalize:2 bayesian:17 identification:1 produced:4 dbo:3 ed:1 danish:1 pp:2 larsen:1 yoshizawa:1 di:1 mi:1 popular:1 salamon:1 higher:2 supervised:1 follow:1 furthermore:1 just:2 xa:2 nonlinear:2 normalized:3 true:1 analytically:1 hence:5 regularization:1 read:1 white:1 game:1 width:6 criterion:1 generalized:1 theoretic:1 oid:4 temperature:12 variational:2 invoked:1 misspecified:1 functional:2 kinzel:1 conditioning:1 jp:1 association:1 intensely:1 cambridge:1 heskes:3 funded:1 xib:1 etc:1 posterior:6 recent:2 scenario:3 verlag:2 seen:1 minimum:1 eo:1 r0:2 signal:1 reduces:1 smooth:3 technical:2 match:1 prediction:6 variant:1 regression:2 metric:1 jid:1 normalization:1 biased:1 db:5 bid:1 variate:1 architecture:1 reduce:1 tradeoff:1 forecasting:2 hessian:1 york:2 useful:1 clear:2 happened:1 delta:1 write:1 vol:2 key:1 drawn:1 verified:1 asymptotically:1 sum:2 inverse:1 parameterized:3 family:5 almost:1 decision:1 bound:1 resampled:1 wel:1 ddp:2 lkhansen:1 bibliography:1 aspect:1 argument:1 optimality:1 jjj:1 department:1 piscataway:1 happens:1 organizer:1 refered:1 xo:1 lyngby:1 describing:1 lexp:1 tractable:1 decomposing:1 appropriate:1 generic:3 bagging:3 ensure:1 question:2 dependence:1 guessing:1 thank:2 parametrized:1 topic:1 denmark:2 robert:1 statement:2 negative:1 suppress:1 design:1 unknown:4 litterature:1 perform:1 vertical:1 finite:6 situation:2 arbitrary:1 introduced:1 specified:1 raising:1 pattern:2 including:2 max:1 power:1 natural:2 advanced:1 oih:2 scheme:5 improve:2 dtu:1 text:1 prior:15 review:2 geometric:2 literature:1 determining:1 asymptotic:3 relative:1 loss:3 admissibility:1 expect:1 ljung:1 interesting:1 plying:1 proven:1 principle:1 dd:3 playing:1 course:1 rasmussen:1 bias:15 institute:1 face:2 dimension:1 opper:2 valid:2 cumulative:1 rich:1 adaptive:1 transaction:2 approximate:1 veraging:1 status:1 thor:1 global:1 imm:1 assumed:2 conclude:1 decade:1 lthis:1 why:1 robust:1 improving:2 investigated:2 motivation:1 position:1 xl:5 specific:2 emphasized:1 dk:2 normalizing:1 workshop:2 ih:1 vapnik:2 overwhelmed:1 forecast:1 smoothly:2 generalizing:1 depicted:1 entropy:2 simply:1 likely:1 bo:7 u2:3 catastrophe:1 springer:3 corresponds:1 stimulating:1 formulated:1 fisher:1 typical:1 averaging:32 la:1 pragmatic:1 kung:1 dept:1
775
171
366 NEURONAL MAPS FOR SENSORY-MOTOR CONTROL IN THE BARN OWL C.D. Spence, J.C. Pearson, JJ. Gelfand, and R.M. Peterson David Sarnoff Research Center Subsidiary of SRI International CN5300 Princeton, New Jersey 08543-5300 W.E. Sullivan Department of Biology Princeton University Princeton, New Jersey 08544 ABSTRACT The bam owl has fused visual/auditory/motor representations of space in its midbrain which are used to orient the head so that visual or auditory stimuli are centered in the visual field of view. We present models and computer simulations of these structures which address various problems, inclu<lln~ the construction of a map of space from auditory sensory information, and the problem of driving the motor system from these maps. We compare the results with biological data. INTRODUCTION Many neural network models have little resemblance to real neural structures, partly because the brain's higher functions, which they attempt to imitate, are not yet experimentally accessible. Nevertheless, some neural-net researchers are finding that the accessible structures are interesting, and that their functions are potentially useful. Our group is modeling a part of the barn owl's nervous system which orients the head to visual and auditory stimuli. The bam owl's brain stem and midbrain contains a system that locates visual and auditory stimuli in space. The system constructs an auditory map of spatial direction from the non-spatial information in the output of the two cochlea. This map is in the external nucleus of the inferior colliculus, or ICx [Knudsen and Konishi, 1978]. The ICx, along with the visual system, projects to the optic tectum, producing a fused visual and auditory map of space [Knudsen and Knudsen, 1983]. The map in the tectum is the source of target position used by the motor system for orienting the head. In the last futeen years, biologists have determined all of the structures in the system which produces the auditory map of space in the ICx. This system provides sev- Neuronal Maps for Sensory-Motor Coritrol in the Barn Owl eral examples of neuronal maps, regions of tissue in which the response properties of neurons vary continuously with position in the map. (For reviews, see Knudsen, 1984; Knudsen, du Lac, and Esterly, 1987; and Konishi, 1986.) Unfortunately, the motor system and the projections from the tectum are not well known, but experimental study of them has recently begun [Masino and Knudsen, 1988]. We should eventually be able to model a complete system, from sensory input to motor output In this paper we present several models of different parts of the head orientation system. Fig. 1 is an overview of the structures we'll discuss. In the second section of this paper we discuss models for the construction of the auditory space map ~ the lex. In the third section we discuss how the optic tectum might drive the motor system. CONSTRUCTION OF AN AUDITORY MAP OF SPACE The bam owl uses two binaural cues to locate objects in space: azimuth is derived from inter-aural time or phase delay (not onset time difference), while elevation is derived from inter-aural intensity difference (due to vertical asymmetries in sensitivAcoustic Core of the ICc ~ Lateral Shell of the ICc ~ External Nucleus of the IC (ICx) Optic Tectum ~ Retina Figure 1. Overview of the neuronal system for target localization in the barn owl (head orients towards potential targets for closer scrutiny). The illustration focuses on the functional representations of the neuronal computation, and does not show all of the relevant connections. The grids represent the centrally synthesized neuronal maps and the patterns within them indicate possible patterns of neuronal activation in response to acoustic stimuli. 367 368 Spence, Pearson, Gelfand, Peterson and Sullivan ity). Corresponding to these two cues are two separate processing streams which produce maps of the binamal cues, which are shown in Figs. 2-5. The information on these maps must be merged in order to construct the map of space in the ICx. lTD ---. f lex :~:~: .~ ~ tiS ~-----~ ~ t _ _ _.-...IID Azimuth Figure 2. Standard Model for the construction of an auditory map of space from maps of the binaural cues. Shading represents activity level. lID is Inter-aural Intensity Difference, lTD is Inter-aural Time Delay. A simple model for combining the two maps is shown in Fig. 2. It has not been described explicitly in the literature, but it has been hinted at [Knudsen, et ai, 1987]. For this reason we have called it the standard model. Here all of the neurons representing a given time delay or azimuth in the lTD vs. frequency map project to all of the neurons representing that azimuth in the space map. Thus a stimulus with a certain lTD would excite a strip of cells representing the associated azimuth and all elevations. Similarly, all of the neurons representing a given intensity difference or elevation in the lID vs. frequency map project to all of the neurons representing that elevation in the space map. (TIle map of lID vs. frequency is constructed in the nucleus ventralis lemnisci lateralis pars posterior, or VLVp. VLVp neurons are said to be sensitive to intensity difference, that is they fire if the intensity difference is great enough. Neurons in the VLVp are spatially organized by their intensity difference threshold [Manley. et ai, 1988]. Thus, intensity difference has a bar-chart-like representation, and our model needs some mechanism to pick out the ends of the bars.) Only the neurons at the intersection of the two strips will fire if lateral inhibition allows only those neurons receiving the most stimulation to fue. In the third section we will present a model for connections of inhibitory inter-neurons which can be applied to this model. Part of the motivation for the standard model is the problem with phase ghosts. Phase ghosts occur when the barn owl's nervous system incorrectly associates the wave fronts arriving at the two ears at high frequency. In this case, neurons in the map of lTD vs. frequency become active at locations representing a time delay which Neuronal Maps for Sensory-Motor Control in the Barn Owl differs from the true time delay by an integer multiple of the period of the sound. Because the period varies inversely with the frequency, these phase ghosts will have apparent time delays that vary with frequency. Thus, for stimuli that are not pure tones, if the bam owl can compare the activities in the map at different frequencies, it can eliminate the ghosts. The standard model does this (Fig. 2). In the lTD vs. frequency map there are more neurons fIring at the position of the true lTD than at the ghost positions, so space map neurons representing the true position will receive the most stimulation. Only those neurons representing the true position will fIre because of the lateral inhibition. There is another kind of ghost which we call the multiple-source ghost (Fig. 3). If two sounds occur simultaneously, then space map neurons representing the time delay of one source and the intensity difference of the other will receive a large amount of stimulation. Lateral inhibition may suppress these ghosts, but if so, the owl should only be able to locate one source at a time. In addition, the true sources might be suppressed. The bam owl may actually suffer from this problem, although it seems unlikely if the owl has to function in a noisy environment The relevant behavioral experiments have not yet been done. Experimental evidence does not support the standard model. The ICx receives most of its input from the lateral shell of the central nucleus of the inferior colliculus lTD --.. ?t . 4~ ::. - / / ., ~Cx 1- "' ~" i/" /: ~ " ',/ I t liD Azimuth Figure 3. Multiple-source ghosts in the standard model for the construction of the auditory space map. For clarity, only two pure tone stimuli are represented, and their frequencies and locations are such that the "phase ghost" problem is not a factor. The black squares represent regions of cells that are above threshold. The circled regions are those that are fIring in response to the lTD of one stimulus and the lID of another. These regions correspond to phantom targets. 369 370 Spence, Pearson, Gelfand, Peterson and Sullivan (lateral shell of the ICc) [Knudsen, 1983]. Neurons in the lateral shell are tuned to frequency and time delay, and these parameters are mapped [Wagner. Takahashi. and Konishi, 1987]. However, they are also affected by intensity difference [Knudsen and Konishi. 1978, I. Fujita, private communication]. Thus the lateral shell does not fit the picture of the input to the ICx in the standard model. rather it is some intermediate stage in the processing. We have a model, called the lateral shell model. which does not suffer from multiple-source ghosts (Fig. 4). In this model, the lateral shell of the ICc is a tbreedimensional map of frequency vs. intensity difference vs. time delay. A neuron in the map of time delay vs. frequency in the ICc core projects to all of the neurons in lTD lID lex Azimuth Figure 4. Lateral shell model for the construction of the auditory map of space in the ICx. f: frequency. lTD: inter-aural time delay. lID: inter-aural intensity difference. Neuronal Maps for Sensory-Motor Control in the Barn Owl the three-dimensional map which represent the same time delay and frequency. As in the standard model, a strip of neurons is stimulated, but now the frequency tuning is preserved. The map of intensity difference vs. frequency in the nucleus ventralis lemnisci lateralis pars posterior (VLVp) [Manley, et al, 1988] projects to the threedimensional map in a similar fashion. Lateral inhibition picks out the regions of intersection of the strips. Neurons in the space map in the ICx receive input from the strip of neurons in the three-dimensional map which represent the appropriate time delay and intensity difference, or equivalently azimuth and elevation. Phase ghosts will be present in the three-dimensional map, but in the ICx lateral inhibition will suppress them. Multiple-source ghosts are eliminated in the lateral shell model because the sources are essentially tagged by their frequency spectra. If two sources with no common frequencies are present, there are no neurons which represent the time delay of one source, the intensity difference of another, and a frequency common to both. In the more likely case in which some frequencies are common to both sources, there will be fewer neurons fIring at the ghost positions than at the real positions, so again lateral inhibition in the ICx can suppress fIring at the ghost positions, exactly as it suppresses the phase ghosts. The fact that intensity and time delay information is combined before frequency tuning is lost in the lex suggests that the owl handles multiple-source ghosts by frequency tagging. A three dimensional map is not essential, but it is conceptually simple. Before ending this section, we should mention that others have independently thought of this model. In particular, M. Konishi and co-workers have looked for a spatial organization or mapping of intensity response properties in the lateml shell, but they have not found it They also have said that they can't yet rule it out [M. Konishi, I. Fujita, private communication]. DRMNG THE MOTOR SYSTEM FROM MAPS As mentioned before, all of the parts of the auditory system in the bmin stem and midbrain are known, up to the optic tectum. The optic tectum has a map of spatial direction which is common to both the visual and auditory systems. In addition, it drives the motor system, so if the tectum is stimulated at a point, the bam owl's head will move to face a certain direction. The new orientation is mapped in register with the auditory/visual map of spatial direction, e.g., stimulating a location which represents a stimulus eight degrees to the right of the current orientation of the head will cause the head to turn eight degrees to the right Little is known about the projections of the tectum, although work has started [Masino and Knudsen, 1988]. There was one earlier experiment. Two electrodes were placed in one of the tecta at positions representing sensory stimuli eight degrees and sixty degrees toward the side of the head opposite to this tectum. When either position was stimulated by itself, the alert owl moved its head as expected, by eight or 371 372 Spence, Pearson, Gelfand, Peterson and Sullivan sixty degrees. When both were stimulated together. the head moved about forty degrees [du Lac and Knudsen. 1987]. The averaging of activity in the tectum is easy to explain. In some motor models it should be produced naturally by the activation of an agonist-antagonist muscle pair (see. for example. Grossberg and Kuperstein. 1986). In the presence of two stimuli. the tension in each muscle is the sum of the tensions it would have for either stimulus alone. so the equilibrium position should be about the average position. We have a different model. which produces a map of the nection strengths from tectal cells to an averaging map with the difference in represented direction. A quadratic is very broad. however. so lateral inhibition is required fairly narrow. ? 1-D tectum - . 25 average position. The concell decrease quadratically distribution of stimulation to make the active region ? . . .... .... '" . ".;- . -.rI''....' .... I.' . o 0 .... . ' ' .-.' ..' . . . 0 . -.- -. . 00 , .......... ,.. -.. .. . .. . . o o oo. ? _ 0 ? . ? '. 0 .. . . o 0 . '. ......'" . '.. . . .. .- . .. _". I...... '. _ ? ??? ??? o~----------------~~~------------------------~ Position Figure 5. Averaging map simulation. The upper thin rectangle is a one-dimensional version of the space map in the optic tectum. The two marks represent the position of stimulating electrodes that are simultaneously active. The lower rectangle is the averaging map. with position represented horizontally and time increasing vertically. The squares represent cell ftrings. Note that the activity quickly becomes centered at the average position of the two active positions in the tectum. We have simulated a one-dimensional version of this model. with 128 cells in the tectum. and in the averaging map. 128 excitatory and 128 inhibitory cells. The excitatory cells and the inhibitory cells both receive the same quadratically weighted input from the tectum. Each inhibitory cell inhibits all of the other cells in the averaging map. both excitatory and inhibitory. except for those in a small local neighborhood. (The weights are actually proportional to one minus a gaussian with a maximum of one.) An excitatory cell receives exactly the same input as the inhibitory Neuronal Maps for Sensory-Motor Control in the Barn Owl cell at the same location, so their voltages are the same. Because of this we only show the excitatory cells in Fig. 5. This figure shows cell position horizontally and time increasing vertically. The black squares are plotted at the time a given cell frres. We are interested in whether an architecture like this is biologically plausible. For this reason we have tried to be fairly realistic in our model. The cells obey a membrane equation: dv - = -gl (v - VI) - ge (v - v ) - g. (v - v.) C dt el I in which C is the capacitance, the g's are conductances, I refers to leakage quantities. e refers to excitatory quantities, and i refers to inhibitory quantities. The output of a cell is not a real number, but spikes that occur randomly at an average rate that is a monotonic function of the voltage. We used the usual sigmoidal function in this simulation, although the membrane equation automatically limits the voltage and hence the firing rate. A cell that spikes on a given time step affects other cells by affecting the appropriate conductance. To get the effect of a post synaptic potential, the conductances obey the equation of a damped harmonic oscillator: d 2g _y dg -2 -- (If-ro g - dt 2 When a spike from an excitatory cell arrives, we increment the time derivative of g by some amount If the oscillator is overdamped or critically damped, the conductance goes up for a time and then decreases, approaching zero exponentially. We are not suggesting that a damped harmonic oscillator exists in the membranes of neurons, but it efficiently models the dynamics of synaptic transmission. The equations for the conductances also have the nice property that the effects of multiple spikes at different times add. With values for the cell parameters that agree well with experimental data, it takes about twenty milliseconds for the simulated map to settle into a fairly steady state, which is a reasonable time for the function of this map. Also, there was no need to frne tune the parameters; within a fairly wide range the effect of changing them is to change the width of the region of activity. We tried another architecture for the inhibitory intemeurons, in which they received their input from the excitatory neurons and did not inhibit other inhibitory neurons. The voltages in this architecture oscillated for a very long time, without picking out a maximum. The architecture we are now using is apparently superior. Since it is quick to pick out a maximum of a broad distribution of stimulation, it should work very well in other models requiring lateral inhibition. such as the lateral shell model discussed earlier. 373 374 Spence, Pearson, Gelfand, Peterson and Sullivan CONCLUSION We have presented models for two parts of the barn owl's visual/auditory localization and head orientation system. These models make experimentally testable predictions. and suggest architectures for artificial systems. One model constructs a map of stimulus position from maps of inter-aural intensity and timing differences. This model solves potential problems with ghosts. i.e.. the representation of false sources in the presence of certain kinds of real sources. Another model computes the average value of a quantity represented on a neural map when the activity on the map has a complex distribution. This model explains recent physiological experiments. A simulation with fairly realistic model neurons has shown that a biological structure could perform this function in this way. A common feature of these models is the use of neuronal maps. We have only mentioned a few of the maps in the barn owl. and they are extremely common in other nervous systems. We think this architecture shows great promise for applications in artificial processing systems. Acknowledgments This work was supported by internal funds of the David Sarnoff Research Center. References Grossberg. S. and M. Kuperstein (1986) Neural dynamics of adaptive motor control. North-Holland. Knudsen. E.I. (1983) J. Compo Neurology. 218:174-186. Knudsen. E.I. (1984) in Dynamical Aspects of Neocortical Function. G.M. Edelman, W.E. Gall, and W.M. Cowan, editors, Wiley, New York. Knudsen, E.I. andP.F. Knudsen (1983) J. Compo Neurology, 218:187-196. Knudsen, E.I. and M. Konishi (1978) J. Neurophys .? 41:870-884. Knudsen. E.!.. S. du Lac. and S. Esterly (1987) Ann. Rev. Neurosci., 10:41-56. Konishi. M. (1986) Trends in Neuroscience. April. du Lac. S. and E.!. Knudsen (1987) Soc. for Neurosci. Abstr., 112.10. Manley. G.A.? A.C. Koeppl. and M. Konishi (1988) J. Neurosci. 8:2665-2676 Masino. T .? and E.I. Knudsen (1988) Soc. for Neurosci. Abstr., 496.16. Wagner. H.? T. Takahashi. and M. Konishi (1987) J. Neurosci. 10:3105-3116
171 |@word private:2 version:2 sri:1 seems:1 simulation:4 tried:2 pick:3 mention:1 minus:1 shading:1 contains:1 tuned:1 current:1 neurophys:1 activation:2 yet:3 must:1 realistic:2 motor:15 fund:1 v:9 alone:1 cue:4 fewer:1 nervous:3 imitate:1 tone:2 core:2 compo:2 provides:1 location:4 sigmoidal:1 along:1 constructed:1 alert:1 become:1 edelman:1 behavioral:1 inter:8 tagging:1 expected:1 brain:2 automatically:1 little:2 increasing:2 becomes:1 project:5 kind:2 suppresses:1 finding:1 esterly:2 ti:1 exactly:2 ro:1 control:5 scrutiny:1 producing:1 before:3 vertically:2 local:1 timing:1 limit:1 firing:5 might:2 black:2 suggests:1 co:1 sarnoff:2 range:1 grossberg:2 acknowledgment:1 spence:5 lost:1 differs:1 sullivan:5 thought:1 projection:2 refers:3 suggest:1 get:1 map:62 phantom:1 center:2 quick:1 go:1 independently:1 oscillated:1 pure:2 rule:1 ity:1 konishi:10 handle:1 increment:1 construction:6 tectum:16 target:4 us:1 gall:1 associate:1 trend:1 region:7 decrease:2 inhibit:1 mentioned:2 environment:1 dynamic:2 localization:2 binaural:2 jersey:2 various:1 represented:4 masino:3 artificial:2 pearson:5 neighborhood:1 apparent:1 gelfand:5 plausible:1 think:1 noisy:1 itself:1 net:1 relevant:2 combining:1 moved:2 electrode:2 asymmetry:1 transmission:1 abstr:2 produce:3 object:1 oo:1 received:1 solves:1 soc:2 indicate:1 direction:5 merged:1 centered:2 settle:1 owl:20 explains:1 elevation:5 biological:2 hinted:1 barn:10 ic:1 great:2 equilibrium:1 mapping:1 driving:1 vary:2 sensitive:1 weighted:1 gaussian:1 rather:1 voltage:4 derived:2 focus:1 el:1 eliminate:1 unlikely:1 interested:1 fujita:2 orientation:4 spatial:5 fairly:5 biologist:1 field:1 construct:3 eliminated:1 biology:1 represents:2 broad:2 thin:1 others:1 stimulus:13 few:1 retina:1 randomly:1 dg:1 simultaneously:2 lac:4 phase:7 fire:3 attempt:1 conductance:5 organization:1 sixty:2 arrives:1 damped:3 closer:1 worker:1 plotted:1 modeling:1 earlier:2 delay:15 azimuth:8 front:1 varies:1 combined:1 international:1 accessible:2 receiving:1 picking:1 together:1 fused:2 continuously:1 quickly:1 again:1 central:1 ear:1 tile:1 external:2 derivative:1 takahashi:2 potential:3 suggesting:1 north:1 vlvp:4 explicitly:1 register:1 onset:1 stream:1 vi:1 view:1 apparently:1 wave:1 chart:1 square:3 efficiently:1 correspond:1 conceptually:1 produced:1 iid:1 agonist:1 critically:1 researcher:1 drive:2 tissue:1 explain:1 strip:5 synaptic:2 frequency:23 naturally:1 associated:1 auditory:17 begun:1 kuperstein:2 organized:1 actually:2 higher:1 dt:2 tension:2 response:4 april:1 done:1 stage:1 receives:2 resemblance:1 orienting:1 effect:3 intemeurons:1 requiring:1 true:5 tagged:1 hence:1 spatially:1 ll:1 width:1 inferior:2 steady:1 antagonist:1 complete:1 bmin:1 neocortical:1 harmonic:2 recently:1 common:6 superior:1 functional:1 stimulation:5 overview:2 exponentially:1 discussed:1 synthesized:1 ai:2 tuning:2 grid:1 similarly:1 aural:7 cn5300:1 inhibition:8 add:1 posterior:2 recent:1 certain:3 muscle:2 forty:1 period:2 tectal:1 multiple:7 sound:2 stem:2 long:1 post:1 locates:1 prediction:1 essentially:1 cochlea:1 represent:7 cell:21 preserved:1 receive:4 affecting:1 addition:2 source:15 cowan:1 integer:1 call:1 presence:2 intermediate:1 enough:1 easy:1 affect:1 fit:1 architecture:6 approaching:1 opposite:1 whether:1 ltd:11 suffer:2 york:1 cause:1 jj:1 useful:1 tune:1 amount:2 inhibitory:9 millisecond:1 neuroscience:1 promise:1 affected:1 group:1 nevertheless:1 threshold:2 clarity:1 changing:1 lateralis:2 rectangle:2 year:1 colliculus:2 sum:1 orient:3 reasonable:1 eral:1 centrally:1 quadratic:1 activity:6 strength:1 occur:3 optic:6 ri:1 lemniscus:2 aspect:1 extremely:1 subsidiary:1 inhibits:1 department:1 membrane:3 suppressed:1 lid:7 biologically:1 rev:1 midbrain:3 dv:1 equation:4 agree:1 discus:3 eventually:1 mechanism:1 turn:1 ge:1 end:1 eight:4 obey:2 appropriate:2 testable:1 threedimensional:1 leakage:1 ventralis:2 move:1 capacitance:1 lex:4 looked:1 quantity:4 spike:4 usual:1 said:2 manley:3 separate:1 mapped:2 lateral:18 simulated:2 reason:2 toward:1 illustration:1 equivalently:1 unfortunately:1 potentially:1 suppress:3 twenty:1 perform:1 upper:1 vertical:1 neuron:27 knudsen:19 incorrectly:1 communication:2 head:12 locate:2 intensity:17 david:2 pair:1 required:1 tecta:1 connection:2 acoustic:1 icx:11 narrow:1 quadratically:2 address:1 able:2 bar:2 andp:1 dynamical:1 bam:6 pattern:2 ghost:18 representing:10 inversely:1 picture:1 started:1 sev:1 review:1 literature:1 circled:1 icc:5 nice:1 par:2 interesting:1 proportional:1 nucleus:5 degree:6 editor:1 excitatory:8 placed:1 lln:1 last:1 arriving:1 gl:1 supported:1 side:1 wide:1 peterson:5 face:1 wagner:2 ending:1 computes:1 sensory:8 adaptive:1 active:4 excite:1 neurology:2 spectrum:1 stimulated:4 du:4 complex:1 did:1 neurosci:5 motivation:1 neuronal:11 fig:7 fashion:1 wiley:1 position:21 third:2 physiological:1 evidence:1 essential:1 exists:1 false:1 intersection:2 cx:1 likely:1 visual:10 horizontally:2 holland:1 monotonic:1 shell:11 stimulating:2 ann:1 towards:1 oscillator:3 experimentally:2 change:1 determined:1 except:1 averaging:6 called:2 partly:1 experimental:3 internal:1 support:1 mark:1 overdamped:1 princeton:3
776
1,710
Learning to Parse Images Geoffrey E. Hinton and Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London London, United Kingdom WC1N 3AR {hinton,zoubin}@gatsby.ucl.ac.uk Vee Whye Tah Department of Computer Science University of Toronto Toronto, Ontario, Canada M5S 3G4 ywteh@cs.utoronto.ca Abstract We describe a class of probabilistic models that we call credibility networks. Using parse trees as internal representations of images, credibility networks are able to perform segmentation and recognition simultaneously, removing the need for ad hoc segmentation heuristics. Promising results in the problem of segmenting handwritten digits were obtained. 1 Introd uction The task of recognition has been the main focus of attention of statistical pattern recognition for the past 40 years. The paradigm problem is to classify an object from a vector of features extracted from the image. With the advent of backpropagation [1], the choice of features and the choice of weights to put on these features became part of a single, overall optimization and impressive performance was obtained for restricted but important tasks such as handwritten character identification [2]. A significant weakness of many current recognition systems is their reliance on a separate preprocessing stage that segments one object out of a scene and approximately normalizes it. Systems in which segmentation precedes recognition suffer from the fact that the segmenter does not know the shape of the object it is segmenting so it cannot use shape information to help it. Also, by segmenting an image, we remove the object to be recognized from the context in which it arises. Although this helps in removing the clutter present in the rest of the image, it might also reduce the ability to recognize an object correctly because the context in which an object arises gives a great deal of information about the nature of the object. Finally, each object can be described in terms of its parts, which can also be viewed as objects in their own right. This raises the question of how fine-grained the segmentations should be. In the words of David Marr: "Is a nose an object? Is a head one? ... What about a man on a horseback?" [3]. G. E. Hinton, Z. Ghahramani and Y. W. Teh 464 The successes of structural linguistics inspired an alternative approach to pattern recognition in which the paradigm problem was to parse an image using a hierarchical grammar of scenes and objects. Within linguistics, the structural approach was seen as an advance over earlier statistical approaches and for many years linguists eschewed probabilities, even though it had been known since the 1970's that a version of the EM algorithm could be used to fit stochastic context free grammars. Structural pattern recognition inherited the linguists aversion to probabilities and as a result it never worked very well for real data. With the advent of graphical models it has become clear that structure and probabilities can coexist. Moreover, the "explaining away" phenomenon that is central to inference in directed acyclic graphical models is exactly what is needed for performing inferences about possible segmentations of an image. In this paper we describe an image interpretation system which combines segmentation and recognition into the same inference process. The central idea is the use of parse trees of images. Graphical models called credibility networks which describe the joint distribution over the latent variables and over the possible parse trees are used. In section 2 we describe some current statistical models of image interpretation. In section 3 we develop credibility networks and in section 4 we derive useful learning and inference rules for binary credibility networks. In section 5 we demonstrate that binary credibility networks are useful in solving the problem of classifying and segmenting binary handwritten digits. Finally in section 6 we end with a discussion and directions for future research. 2 Related work Neal [4] introduced generative models composed of multiple layers of stochastic logistic units connected in a directed acyclic graph. In general, as each unit has multiple parents, it is intractable to compute the posterior distribution over hidden variables when certain variables are observed. However, Neal showed that Gibbs sampling can be used effectively for inference [4]. Efficient methods of approximating the posterior distribution were introduced later [5, 6, 7] and these approaches were shown to yield good density models for binary images of handwritten digits [8] . The problem with these models which make them inappropriate for modeling images is that they fail to respect the 'single-parent' constraint: in the correct interpretation of an image of opaque objects each object-part belongs to at most one object - images need parse trees, not parse DAGs. Multiscale models [9] are interesting generative models for images that use a fixed tree structure. Nodes high up in the tree control large blocks of the image while bottom level leaves correspond to individual pixels. Because a tree structure is used, it is easy to compute the exact posterior distribution over the latent (non-terminal) nodes given an image. As a result, the approach has worked much better than Markov random fields which generally involve an intractable partition function . A disadvantage is that there are serious block boundary artifacts, though overlapping trees can be used to smooth the transition from one block to another [10]. A more serious disadvantage is that the tree cannot possibly correspond to a parse tree because it is the same for every image. Zemel, Mozer and Hinton [11] proposed a neural network model in which the activities of neurons are used to represent the instantiation parameters of objects or their parts, Le. the viewpoint-dependent coordinate transformation between an object's intrinsic coordinate system and the image coordinate system. The weights on connections are then used to represent the viewpoint-invariant relationship between the instantiation parameters of a whole, rigid object and the instantiation parame- Learning to Parse Images 465 ters of its parts. This model captures viewpoint invariance nicely and corresponds to the way viewpoint effects are handled in computer graphics, but there was no good inference procedure for hierarchical models and no systematic way of sharing modules that recognize parts of objects among multiple competing object models. Simard et al [12] noted that small changes in object instantiation parameters result in approximately linear changes in (real-valued) pixel intensities. These can be captured successfully by linear models. To model larger changes, many locally linear models can be pieced together. Hinton, Dayan and Revow [13] proposed a mixture of factor analyzers for this. Tipping and Bishop have recently shown how to make this approach much more computationally efficient [14]. To make the approach really efficient, however, it is necessary to have multiple levels of factor analyzers and to allow an analyzer at one level to be shared by several competing analyzers at the next level up. Deciding which subset of the analyzers at one level should be controlled by one analyzer at the level above is equivalent to image segmentation or the construction of part of a parse tree and the literature on linear models contains no proposals on how to achieve this. 3 A new approach to image interpretation We developed a class of graphical models called credibility networks in which the possible interpretations of an image are parse trees, with nodes representing objectparts and containing latent variables. Given a DAG, the possible parse trees of an image are constrained to be individual or collections of trees where each unit satisfies the single-parent constraint, with the leaves being the pixels of an image. Credibility networks describe a joint distribution over the latent variables and possible tree structures. The EM algorithm [15] can be used to fit credibility networks to data. Let i E I be a node in the graph. There are three random variables associated with i. The first is a multinomial variate ,xi = {.xij hEpa(i) which describes the parent of i from among the potential parents pa(i) : ,x . . _ 1J - {I if parent of i is j, 0 if parent of i is not j. (1) The second is a binary variate Si which determines whether the object i l is present (Si = 1) or not (Si = 0). The third is the latent variables Xi that describe the pose and deformation of the object. Let A = {,xi : i E I}, S = {Si : i E I} and X = {Xi: i E I}. Each connection j -+ i has three parameters also. The first, Cij is an unnormalized prior probability that j is i's parent given that object j is present. The actual prior probability is 7rij = C" S ' 1J J (2) 2:kEPa(i) CikSk We assume there is always a unit 1 E pa(i) such that SI = 1. This acts as a default parent when no other potential parent is present and makes sure the denominator in (2) is never O. The second parameter, Pij, is the conditional probability that object i is present given that j is i's parent (,xij = 1). The third parameter tij characterizes the distribution of Xi given ,xij = 1 and Xj' Let 0 = {Cij,Pij, tij : i E I,j E pa(i)}. Using Bayes' rule the joint distribution over A, S and X given 0 is peA, S, XIO) = peA, SIO)p(XIA, S, 0). Note that A and S together define a parse tree for the image. Given the parse tree the distribution over latent variables p(XIA, S, 0) can be ITechnically this should be the object represented by node i. G. E. Hinton, Z. Ghahramani and Y. W. Teh 466 efficiently inferred from the image. The actual form of p(XIA, S, 0) is unimportant. The joint distribution over A and S is = II P(A,SIO) II (3) (1I'ijP:j(1- Pij)I-8 i )Ai j iEi jEpa(i) 4 Binary credibility networks The simulation results in section 5 are based on a simplified version of credibility networks in which the latent variables X are ignored. Notice that we can sum out A from the joint distribution (3), so that P(SIO) = II L 1I'ijP:j (1 - (4) Pij)I-8 i iEi jEpa(i) Using Bayes' rule and dividing (3) by (4), we have II II ( P(AIS, 0) = iEi jEpa(i) CijSjP:j (1: Pij)I-8 i I:kEPa(i) cik s kPik(l - Pik) 1-8') (5) Aij ? )1-8 .. Let rij = CijP:j (1 - Pij We can view rij as the unnormalized posterior probability that j is i's parent given that object j is present. The actual posterior is the fraction in (5) : (6) Given some observations 0 c S, let 11 = S \ 0 be the hidden variables. We approximate the posterior distribution for 11 using a factored distribution Q(ll) = II O':i (1 - O'i)1-8 i (7) iEi EQ[-log P(SIO) + log Q(S)] is L(EQ[IOg L cijsj-Iog L The variational free energy, F(Q,O) = F(Q,O) = Ci jSjP:j(1-Pi j )1-8i]) iEi L jEpa(i) (O'i log O'i + (1 - + jEpa(i) O'i) log (1 - (8) O'i)) iEi The negative of the free energy -F is a lower bound on the log likelihood of generating the observations O. The variational EM algorithm improves this bound by iteratively improving -F with respect to Q (E-step) and to 0 (M-step). Let ch(i) be the possible children of i. The inference rules can be derived from (8) : EQ [log L O'i CijSjPij -log jEpa(i) = sigmoid + L EQ [log L lEch( i) L cijsj(l - Pij)] jEpa(i) jEpa(l) rljSj -log L J:::: (9) CljSj jEpa(I)' Let D be the training set and Qd be the mean field approximation to the posterior distribution over 11 given the training data (observation) dE D. Then the learning Learning to Parse Images 467 Figure 1: Sample images from the test set. The classes of the two digits in each image in a row are given to the left. rules are (10) (11) For an efficient implementation of credibility networks using mean field approximations, we still need to evaluate terms of the form E[logx] and ~[l/x] where x is a weighted sum of binary random variates. In our implementation we used the simplest approximations: E[logx] ~ 10gE[x] and E[l/x] ~ l/E[x]. Although biased the implementation works well enough in general. 5 Segmenting handwritten digits Hinton and Revow [16] used a mixture of factor analyzers model to segment and estimate the pose of digit strings. When the digits do not overlap, the model was able to identify the digits present and segment the image easily. The hard cases are those in which two or more digits overlap significantly. To assess the ability of credibility networks at segmenting handwritten digits, we used superpositions of digits at exactly the same location. This problem is much harder than segmenting digit strings in which digits partially overlap. The data used is a set of 4400 images of single digits from the classes 2, 3, 4 and 5 derived from the CEDAR CDROM 1 database [17]. Each image has size 16x16. The size of the credibility network is 256-64-4. The 64 middle layer units are meant to encode low level features, while each of the 4 top level units are meant to encode a digit class. We used 700 images of single digits from each class ? to train the network. So it was not trained to segment images. During training we clamped at 1 the activation of the top layer unit corresponding to the class of the digit in the current image while fixing the rest at O. After training, the network was first tested on the 1600 images of single digits not in the training set. The predicted class of each image was taken to be the G. E. Hinton, Z. Ghahramani and Y. W. Teh 468 Figure 2: Segmentations of pairs of digits. (To make comparisons easier we show the overlapping image in both columns of a)-I).) class corresponding to the top layer unit with the highest activation. The error rate was 5.5%. We then showed the network 120 images of two overlapping digits from distinct classes. There were 20 images per combination of two classes. Some examples are given in Figure 1. The predicted classes of the two digits are chosen to be the corresponding classes of the 2 top layer units with the highest activations. A human subject (namely the third author) was tested on the same test set. The network achieved an error rate of 21.7% while the author erred on 19.2% of the images. We can in fact produce a segmentation of each image into an image for each class present. Recall that given the values of S the posterior probability of unit j being pixel i's parent is Wij. Then the posterior probability of pixel i belonging to digit class k is L: j EQ[WijWjk). This gives a simple way to segment the image. Figure 2 shows a number of segmentations. Note that for each pixel, the sum of the probabilities of the pixel belonging to each digit class is 1. To make the picture clearer, a white pixel means a probability of :::; .1 of belonging to a class, while black means ~ .6 probability, and the intensity of a gray pixel describes the size of the probability if it is between .1 and .6. Figures 2a) to 2f) shows successful segmentations, while Figure 2g) to 21) shows unsuccessful segmentations. 6 Discussion Using parse trees as the internal representations of images, credibility networks avoid the usual problems associated with a bottom-up approach to image interpretation. Segmentation can be carried out in a statistically sound manner, removing the need for hand crafted ad hoc segmentation heuristics. The granularity problem for segmentation is also resolved since credibility networks use parse trees as internal representations of images. The parse trees describe the segmentations of the image at every level of granularity, from individual pixels to the whole image. We plan to develop and implement credibility networks in which each latent variable Xi is a multivariate Gaussian, so that a node can represent the position, orientation and scale of a 2 or 3D object, and the conditional probability models on the links can represent the relationship between a moderately deformable object and its parts. Learning to Parse Images 469 Acknowledgments We thank Chris Williams, Stuart Russell and Phil Dawid for helpful discussions and NSERC and ITRC for funding. References [1] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Learning internal representations by error propagation. In D. E. Rumelhart, J. L. McClelland, and the PDP Research Group, editors, Parallel Distributed Processing: Explorations in The Microstructure of Cognition. Volume 1 : Foundations. The MIT Press, 1986. [2] Y. Le Cun, B. Boser, J. S. Denker, S. SoH a, R. E. Howard, and L. D. Jackel. Back-propagation applied to handwritten zip code recognition. Neural Computation, 1(4):541-551, 1989. [3] D. Marr. Vision: A Computational Investigation into the Human Representation and Processing of Visual Information. W. H. Freeman and company, San Francisco, 1980. [4] R. M. Neal. Connectionist learning of belief networks. Artificial Intelligence, 56:71113, 1992. [5] P. Dayan, G. E. Hinton, R. M. Neal, and R. S. Zemel. Helmholtz machines. Neural Computation, 7:1022-1037, 1995. [6] G. E. Hinton, P. Dayan, B. J. Frey, and R. M. Neal. The wake-sleep algorithm for self-organizing neural networks. Science, 268:1158-1161, 1995. [7] L. K. Saul and M. I. Jordan. Attractor dynamics in feedforward neural networks. Submitted for publication. [8] B. J. Frey, G. E. Hinton, and P. Dayan. Does the wake-sleep algorithm produce good density estimators? In D. Touretzky, M. Mozer, and M. Hasselmo, editors, Advances in Neural Information Processing Systems, volume 8. The MIT Press, 1995. [9] M. R. Luettgen and A. S. Willsky. Likelihood calculation for a class of multiscale stochastic models, with application to texture discrimination. IEEE Transactions on Image Processing, 4(2):194-207, 1995. [10] W. W. Irving, P. W. Fieguth, and A. S. Willsky. An overlapping tree approach to multiscale stochastic modeling and estimation. IEEE Transactions on Image Processing, 1995. [11] R. S. Zemel, M. C. Mozer, and G. E. Hinton. TRAFFIC: Recognizing objects using hierarchical reference frame transformations. In Advances in Neural Information Processing Systems, volume 2. Morgan Kaufmann Publishers, San Mateo CA, 1990. [12] P. Simard, Y. Le Cun, and J. Denker. Efficient pattern recognition using a new transformation distance. In S. Hanson, J. Cowan, and L. Giles, editors, Advances in Neural Information Processing Systems, volume 5. Morgan Kaufmann Publishers, San Mateo CA, 1992. [13] G. E. Hinton, P. Dayan, and M. Revow. Modeling the manifolds of images of handwritten digits. IEEE Transactions on Neural Networks, 8:65-74, 1997. [14] M. E. Tipping and C. M. Bishop. Mixtures of probabilistic principal component analysis. Technical Report NCRG/97/003, Aston University, Department of Computer Science and Applied Mathematics, 1997. [15] A.P. Dempster, N.M. Laird, and D.B. Rubin. Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society B, 39:1-38, 1977. [16] G. E. Hinton and M. Revow. Using mixtures of factor analyzers for segmentation and pose estimation, 1997. [17] J. J. Hull. A database for handwritten text recognition research. IEEE Transactions on Pattern Analysis and Machine Intelligence, 16(5):550-554, 1994.
1710 |@word middle:1 version:2 simulation:1 harder:1 contains:1 united:1 past:1 current:3 si:5 activation:3 partition:1 shape:2 remove:1 discrimination:1 generative:2 leaf:2 intelligence:2 node:6 toronto:2 location:1 become:1 combine:1 manner:1 g4:1 terminal:1 inspired:1 freeman:1 company:1 actual:3 inappropriate:1 moreover:1 advent:2 what:2 string:2 developed:1 transformation:3 every:2 act:1 exactly:2 uk:1 control:1 unit:11 segmenting:7 frey:2 approximately:2 might:1 black:1 mateo:2 statistically:1 directed:2 acknowledgment:1 block:3 implement:1 backpropagation:1 digit:24 procedure:1 significantly:1 word:1 zoubin:2 cannot:2 coexist:1 put:1 context:3 equivalent:1 phil:1 williams:2 attention:1 factored:1 rule:5 estimator:1 marr:2 coordinate:3 construction:1 exact:1 pa:3 dawid:1 rumelhart:2 recognition:11 helmholtz:1 database:2 observed:1 bottom:2 module:1 rij:3 capture:1 pieced:1 connected:1 russell:1 highest:2 mozer:3 dempster:1 xio:1 moderately:1 dynamic:1 segmenter:1 raise:1 solving:1 segment:5 trained:1 easily:1 joint:5 resolved:1 represented:1 train:1 distinct:1 describe:7 london:2 precedes:1 zemel:3 artificial:1 heuristic:2 larger:1 valued:1 grammar:2 ability:2 laird:1 hoc:2 ucl:1 organizing:1 ontario:1 achieve:1 deformable:1 parent:13 produce:2 generating:1 object:29 help:2 derive:1 develop:2 ac:1 fixing:1 pose:3 clearer:1 eq:5 dividing:1 c:1 predicted:2 qd:1 direction:1 correct:1 stochastic:4 pea:2 exploration:1 human:2 hull:1 microstructure:1 really:1 investigation:1 deciding:1 great:1 cognition:1 estimation:2 superposition:1 jackel:1 hasselmo:1 soh:1 successfully:1 weighted:1 mit:2 always:1 gaussian:1 avoid:1 publication:1 encode:2 derived:2 focus:1 likelihood:3 helpful:1 inference:7 itrc:1 dependent:1 rigid:1 dayan:5 hidden:2 wij:1 pixel:10 overall:1 among:2 orientation:1 plan:1 constrained:1 field:3 never:2 nicely:1 sampling:1 stuart:1 future:1 connectionist:1 logx:2 report:1 serious:2 composed:1 simultaneously:1 recognize:2 individual:3 attractor:1 weakness:1 mixture:4 wc1n:1 necessary:1 tree:21 incomplete:1 deformation:1 classify:1 earlier:1 modeling:3 column:1 giles:1 ar:1 disadvantage:2 subset:1 cedar:1 recognizing:1 successful:1 graphic:1 density:2 probabilistic:2 systematic:1 together:2 eschewed:1 central:2 ijp:2 luettgen:1 containing:1 possibly:1 simard:2 potential:2 iei:6 de:1 ad:2 later:1 view:1 characterizes:1 traffic:1 bayes:2 parallel:1 inherited:1 ass:1 became:1 kaufmann:2 efficiently:1 correspond:2 yield:1 identify:1 handwritten:9 identification:1 m5s:1 submitted:1 touretzky:1 sharing:1 energy:2 associated:2 erred:1 recall:1 improves:1 segmentation:17 cik:1 back:1 tipping:2 though:2 stage:1 hand:1 parse:19 multiscale:3 overlapping:4 propagation:2 logistic:1 artifact:1 gray:1 effect:1 iteratively:1 neal:5 deal:1 white:1 ll:1 during:1 self:1 irving:1 noted:1 unnormalized:2 whye:1 demonstrate:1 image:54 variational:2 recently:1 funding:1 sigmoid:1 multinomial:1 volume:4 ncrg:1 interpretation:6 significant:1 gibbs:1 dag:2 ai:2 credibility:17 mathematics:1 analyzer:8 had:1 impressive:1 posterior:9 own:1 showed:2 multivariate:1 belongs:1 certain:1 binary:7 success:1 seen:1 captured:1 morgan:2 zip:1 recognized:1 paradigm:2 ii:6 multiple:4 sound:1 smooth:1 technical:1 calculation:1 iog:2 controlled:1 denominator:1 vision:1 represent:4 achieved:1 proposal:1 fine:1 wake:2 publisher:2 biased:1 rest:2 sure:1 subject:1 cowan:1 jordan:1 call:1 structural:3 granularity:2 feedforward:1 easy:1 enough:1 xj:1 fit:2 variate:3 competing:2 reduce:1 idea:1 whether:1 handled:1 introd:1 suffer:1 linguist:2 ignored:1 useful:2 generally:1 clear:1 involve:1 tij:2 unimportant:1 clutter:1 locally:1 mcclelland:1 simplest:1 xij:3 notice:1 neuroscience:1 correctly:1 per:1 group:1 reliance:1 graph:2 fraction:1 year:2 sum:3 opaque:1 pik:1 layer:5 bound:2 sleep:2 activity:1 constraint:2 worked:2 scene:2 ywteh:1 performing:1 department:2 combination:1 belonging:3 describes:2 em:4 character:1 cun:2 restricted:1 invariant:1 taken:1 computationally:1 fail:1 needed:1 know:1 nose:1 ge:1 end:1 denker:2 hierarchical:3 away:1 alternative:1 top:4 linguistics:2 graphical:4 ghahramani:4 approximating:1 society:1 question:1 usual:1 distance:1 separate:1 link:1 thank:1 parame:1 chris:1 manifold:1 willsky:2 code:1 relationship:2 vee:1 kingdom:1 cij:2 negative:1 implementation:3 perform:1 teh:3 neuron:1 observation:3 markov:1 howard:1 hinton:15 head:1 pdp:1 frame:1 canada:1 intensity:2 inferred:1 david:1 introduced:2 pair:1 namely:1 connection:2 hanson:1 boser:1 able:2 pattern:5 cdrom:1 unsuccessful:1 royal:1 belief:1 overlap:3 representing:1 aston:1 picture:1 carried:1 text:1 prior:2 literature:1 interesting:1 acyclic:2 geoffrey:1 foundation:1 aversion:1 pij:7 rubin:1 viewpoint:4 editor:3 classifying:1 pi:1 normalizes:1 row:1 free:3 aij:1 allow:1 explaining:1 saul:1 distributed:1 boundary:1 default:1 xia:3 transition:1 author:2 collection:1 preprocessing:1 simplified:1 san:3 transaction:4 approximate:1 instantiation:4 francisco:1 xi:6 latent:8 promising:1 nature:1 ca:3 improving:1 main:1 whole:2 child:1 crafted:1 gatsby:2 x16:1 position:1 clamped:1 third:3 grained:1 removing:3 bishop:2 utoronto:1 intractable:2 intrinsic:1 uction:1 effectively:1 ci:1 texture:1 easier:1 visual:1 nserc:1 partially:1 ters:1 fieguth:1 ch:1 corresponds:1 satisfies:1 determines:1 extracted:1 conditional:2 viewed:1 revow:4 man:1 shared:1 change:3 hard:1 principal:1 called:2 tah:1 invariance:1 sio:4 college:1 internal:4 arises:2 meant:2 evaluate:1 tested:2 phenomenon:1
777
1,711
Probabilistic methods for Support Vector Machines Peter Sollich Department of Mathematics, King's College London Strand, London WC2R 2LS, U.K. Email: peter.sollich@kcl.ac.uk Abstract I describe a framework for interpreting Support Vector Machines (SVMs) as maximum a posteriori (MAP) solutions to inference problems with Gaussian Process priors. This can provide intuitive guidelines for choosing a 'good' SVM kernel. It can also assign (by evidence maximization) optimal values to parameters such as the noise level C which cannot be determined unambiguously from properties of the MAP solution alone (such as cross-validation error) . I illustrate this using a simple approximate expression for the SVM evidence. Once C has been determined, error bars on SVM predictions can also be obtained. 1 Support Vector Machines: A probabilistic framework Support Vector Machines (SVMs) have recently been the subject of intense research activity within the neural networks community; for tutorial introductions and overviews of recent developments see [1, 2, 3]. One of the open questions that remains is how to set the 'tunable' parameters of an SVM algorithm: While methods for choosing the width of the kernel function and the noise parameter C (which controls how closely the training data are fitted) have been proposed [4, 5] (see also, very recently, [6]), the effect of the overall shape of the kernel function remains imperfectly understood [1]. Error bars (class probabilities) for SVM predictions important for safety-critical applications, for example - are also difficult to obtain. In this paper I suggest that a probabilistic interpretation of SVMs could be used to tackle these problems. It shows that the SVM kernel defines a prior over functions on the input space, avoiding the need to think in terms of high-dimensional feature spaces. It also allows one to define quantities such as the evidence (likelihood) for a set of hyperparameters (C, kernel amplitude Ko etc). I give a simple approximation to the evidence which can then be maximized to set such hyperparameters. The evidence is sensitive to the values of C and Ko individually, in contrast to properties (such as cross-validation error) of the deterministic solution, which only depends on the product CKo. It can thfrefore be used to assign an unambiguous value to C, from which error bars can be derived. P. Sollich 350 I focus on two-class classification problems. Suppose we are given a set D of n training examples (Xi, Yi) with binary outputs Yi = ?1 corresponding to the two classes. The basic SVM idea is to map the inputs X onto vectors c/>(x) in some high-dimensional feature space; ideally, in this feature space, the problem should be linearly separable. Suppose first that this is true. Among all decision hyperplanes w?c/>(x) + b = 0 which separate the training examples (Le. which obey Yi(W'c/>(Xi) + b) > 0 for all Xi E D x , Dx being the set of training inputs), the SVM solution is chosen as the one with the largest margin, Le. the largest minimal distance from any of the training examples. Equivalently, one specifies the margin to be one and minimizes the squared length of the weight vector IIwI1 2 [1], subject to the constraint that Yi(W'c/>(Xi) + b) 2:: 1 for all i. If the problem is not linearly separable, 'slack variables' ~i 2:: 0 are introduced which measure how much the margin constraints are violated; one writes Yi(W'c/>(Xi) + b) 2:: 1 - ~i' To control the amount of slack allowed, a penalty term C Ei ~i is then added to the objective function ~ IIwI1 2, with a penalty coefficient C. Training examples with Yi(w ?c/>(xd + b) 2:: 1 (and hence ~i = 0) incur no penalty; all others contribute C[l - Yi(W ' c/>(Xi) + b)] each. This gives the SVM optimization problem: Find wand b to minimize ~llwl12 + C Ei l(Yi[W ' c/>(Xi) + b]) (1) where l(z) is the (shifted) 'hinge loss', l(z) = (1- z)8(1- z). To interpret SVMs probabilistically, one can regard (1) as defining a (negative) log-posterior probability for the parameters wand b of the SVM, given a training set D. The first term gives the prior Q(w,b) "" exp(-~llwW - ~b2B-2). This is a Gaussian prior on W; the components of W are uncorrelated with each other and have unit variance. I have chosen a Gaussian prior on b with variance B2; the flat prior implied by (1) can be recovered! by letting B -+ 00. Because only the 'latent variable' values O(x) = w?c/>(x) + b - rather than wand b individually - appear in the second, data dependent term of (1), it makes sense to express the prior directly as a distribution over these. The O(x) have a joint Gaussian distribution because the components ofw do, with covariances given by (O(x)O(x')) = ((c/>(x) ?w) (w?c/>(x'))) + B2 = c/>(x)?c/>(x') + B2. The SVM prior is therefore simply a Gaussian process (GP) over the functions 0, with covariance function K(x,x') = c/>(x) ?c/>(x') + B2 (and zero mean). This correspondence between SVMs and GPs has been noted by a number of authors, e.g. [6, 7, 8, 9, 10J . The second term in (1) becomes a (negative) log-likelihood if we define the probability of obtaining output Y for a given X (and 0) as Q(y =?llx, 0) = ~(C) exp[-Cl(yO(x))] (2) We set ~(C) = 1/[1 + exp(-2C)] to ensure that the probabilities for Y ?1 never add up to a value larger than one. The likelihood for the complete data set is then Q(DIO) = It Q(Yilxi, O)Q(Xi), with some input distribution Q(x) which remains essentially arbitrary at this point. However, this likelihood function is not normalized, because lI(O(x)) = Q(llx, 0) + Q( -llx, 0) = ~(C){ exp[ -Cl(O(x))] + exp[-Cl( -O(x))]} < 1 lIn the probabilistic setting, it actually makes more sense to keep B finite (and small); for B -+ 00, only training sets with all Yi equal have nonzero probability. 351 Probabilistic Methods for Support Vector Machines except when IO(x)1 = 1. To remedy this, I write the actual probability model as P(D,9) = Q(DI9)Q(9)/N(D) . (3) Its posterior probability P(9ID) '" Q(DI9)Q(9) is independent Qfthe normalization factor N(D); by construction, the MAP value of 9 is therefore the SVM solution. The simplest choice of N(D) which normalizes P(D, 9) is D-independent: N = Nn = Jd9Q(9)Nn(9), N(9) = JdxQ(x)lI(O(x)). (4) Conceptually, this corresponds to the following procedure of sampling from P(D, 9): First, sample 9 from the GP prior Q(9) . Then, for each data point, sample x from Q(x). Assign outputs Y = ?1 with probability Q(ylx,9), respectively; with the remaining probability l-lI(O(x)) (the 'don't know' class probability in [11]), restart the whole process by sampling a new 9. Because lI(O(x)) is smallest 2 inside the 'gap' IO(x)1 < 1, functions 9 with many values in this gap are less likely to 'survive' until a dataset of the required size n is built up. This is reflected in an n-dependent factor in the (effective) prior, which follows from (3,4) as P(9) '" Q(9)Nn(9). Correspondingly, in the likelihood P(ylx,9) = Q(ylx, 9)/1I(O(x)), P(xI9) '" Q(x) lI(O(x)) (5) (which now is normalized over y = ?1), the input density is influenced by the function 9 itself; it is reduced in the 'uncertainty gaps' IO(x)1 < 1. To summarize, eqs. (2-5) define a probabilistic data generation model whose MAP solution 9* = argmax P(9ID) for a given data set D is identical to a standard SVM. The effective prior P(9) is a GP prior modified by a data set size-dependent factor; the likelihood (5) defines not just a conditional output distribution, but also an input distribution (relative to some arbitrary Q(x)). All relevant properties of the feature space are encoded in the underlying GP prior Q(9), with covariance matrix equal to the kernel K(x, Xl). The log-posterior of the model In P(9ID) = -t J dx dxl O(X)K-l(X, Xl) O(XI) - C 'Ei l(YiO(xi)) + const (6) is just a transformation of (1) from wand b to 9. By differentiating w.r.t. the O(x) for non-training inputs, one sees that its maximum is of the standard form O*(x) = Ei (}:iYiK(X, Xi); for YiO*(Xi) > 1, < 1, and = lone has (}:i = 0, (}:i = C and (}:i E [0, C] respectively. I will call the training inputs Xi in the last group marginal; they form a subset of all support vectors (the Xi with (}:i > 0). The sparseness of the SVM solution (often the number of support vectors is ? n) comes from the fact that the hinge loss l(z) is constant for z > 1. This contrasts with other uses of GP models for classification (see e.g. [12]), where instead of the likelihood (2) a sigmoidal (often logistic) 'transfer function' with nonzerO gradient everywhere is used. Moreover, in the noise free limit, the sigmoidal transfer function becomes a step function, and the MAP values 9* will tend to the trivial solution O*(x) = O. This illuminates from an alternative point of view why the margin (the 'shift' in the hinge loss) is important for SVMs. Within the probabilistic framework, the main effect of the kernel in SVM classification is to change the properties of the underlying GP prior Q(9) in P(9) '" 2This is true for C > In 2. For smaller C, v( O( x? is actually higher in the gap, and the model makes less intuitive sense. P. Sollich 352 (e) (h) Figure 1: Samples from SVM priors; the input space is the unit square [0,1]2. 3d plots are samples 8(x) from the underlying Gaussian process prior Q(8). 2d greyscale plots represent the output distributions obtained when 8(x) is used in the likelihood model (5) with C = 2; the greyscale indicates the probability of y = 1 (black: 0, white: 1). (a,b) Exponential (Ornstein-Uhlenbeck) kernel/covariance function Koexp(-Ix - x/l/l), giving rough 8(x) and decision boundaries. Length scale l = 0.1, Ko = 10. (c) Same with Ko = 1, i.e. with a reduced amplitude of O(x); note how, in a sample from the prior corresponding to this new kernel, the grey 'uncertainty gaps' (given roughly by 18(x)1 < 1) between regions of definite outputs (black/white) have widened. (d,e) As first row, but with squared exponential (RBF) kernel Ko exp[-(x - X I )2/(2l 2)], yielding smooth 8(x) and decision boundaries. (f) Changing l to 0.05 (while holding Ko fixed at 10) and taking a new sample shows how this parameter sets the typical length scale for decision regions. (g,h) Polynomial kernel (1 + x?xl)P, with p = 5; (i) p = 10. The absence of a clear length scale and the widely differing magnitudes of 8(x) in the bottom left (x = [0,0]) and top right (x = [1,1]) corners of the square make this kernel less plausible from a probabilistic point of view. 353 Probabilistic Methods for Support Vector Machines Q(O)Nn(o). Fig. 1 illustrates this with samples from Q(O) for three different types of kernels. The effect of the kernel on smoothness of decision boundaries, and typical sizes of decision regions and 'uncertainty gaps' between them, can clearly be seen. When prior knowledge about these properties of the target is available, the probabilistic framework can therefore provide intuition for a suitable choice of kernel. Note that the samples in Fig. 1 are from Q(O), rather than from the effective prior P(O). One finds, however, that the n-dependent factor Nn(o) does not change the properties of the prior qualitatively3. 2 Evidence and error bars Beyond providing intuition about SVM kernels, the probabilistic framework discussed above also makes it possible to apply Bayesian methods to SVMs. For example, one can define the evidence, i.e. the likelihood of the data D, given the model as specified by the hyperparameters C and (some parameters defining) K(x, x'). It follows from (3) as P(D) = Q(D)/Nn, Q(D) = JdO Q(DIO)Q(O). (7) The factor Q(D) is the 'naive' evidence derived from the unnormalized likelihood model; the correction factor Nn ensures that P(D) is normalized over all data sets. This is crucial in order to guarantee that optimization of the (log) evidence gives optimal hyperparameter values at least on average (M Opper, private communication). Clearly, P(D) will in general depend on C and K(x,x') separately. The actual SVM solution, on the other hand, i.e. the MAP values 0*, can be seen from (6) to depend on the product C K (x, x') only. Properties of the deterministically trained SVM alone (such as test or cross-validation error) cannot therefore be used to determine C and the resulting class probabilities (5) unambiguously. I now outline how a simple approximation to the naive evidence can be derived. Q(D) is given by an integral over all B(x), with the log integrand being (6) up to an additive constant. After integrating out the Gaussian distributed B( x) with x ? Dx , an intractable integral over the B(Xi) remains. However, progress can be made by expanding the log integrand around its maximum B*(Xi)' For all non-marginal training inputs this is equivalent to Laplace's approximation: the first terms in the expansion are quadratic in the deviations from the maximum and give simple Gaussian integrals. For the remaining B(Xi), the leading terms in the log integrand vary linearly near the maximum. Couplings between these B(Xi) only appear at the next (quadratic) order; discarding these terms as subleading, the integral factorizes over the B(xd and can be evaluated. The end result of this calculation is: InQ(D) ~ -! LiYi<liB*(Xi) - CLil(YiB*(xd) - nln(l + e- 2C ) - ! Indet(LmKm) (8) The first three terms represent the maximum of the log integrand, In Q(DIO*); the last one comes from the integration over the fluctuations of the B(x). Note that it only contains information about the marginal training inputs: Km is the corresponding submatrix of K(x, x'), and Lm is a diagonal matrix with entries 3Quantitative changes arise because function values with IO(x)1 < 1 are 'discouraged' for large nj this tends to increase the size of the decision regions and narrow the uncertainty gaps. I have verified this by comparing samples from Q(O) and P(O). P. Sollich 354 0 -0.1 -0.2 -0.3 -0.4 -0.5 9(x) 2 1 o 1 2 0.8 P(y=llx) I 0.6 I 0.4 I \ 0.2 -1 -2 0.2 0.4 x 0.6 0.8 1 o I I I I I \. o 0.2 C J 0.4 x 0.6 4 3 I I I I , \. }; 0.8 1 Figure 2: Toy example of evidence maximization. Left: Target 'latent' function 8(x) (solid line). A SVM with RBF kernel K(x, Xl) = Ko exp[-(x - X I )2 /(2[2)], [ = 0.05, CKo = 2.5 was trained (dashed line) on n = 50 training examples (circles). Keeping CKo constant, the evidence P(D) (top right) was then evaluated as a function of C using (7,8). Note how the normalization factor Nn shifts the maximum of P(D) towards larger values of C than in the naive evidence Q(D). Bottom right: Class probability P(y = 11x) for the target (solid), and prediction at the evidence maximum C ~ 1.8 (dashed) . The target was generated from (3) with C=2 . 27r[ai(C -ai)/C]2. Given the sparseness ofthe SVM solution, these matrices should be reasonably small, making their determinants amenable to numerical computation or estimation [12]. Eq. (8) diverges when ai -+ a or -+ C for one of the marginal training inputs; the approximation of retaining only linear terms in the log integrand then breaks down. I therefore adopt the simple heuristic of replacing det(LmKm) by det(1 + LmKm), which prevents these spurious singularities (I is the identity matrix) . This choice also keeps the evidence continuous when training inputs move in or out of the set of marginal inputs as hyperparameters are varied. Fig. 2 shows a simple application of the evidence estimate (8) . For a given data set, the evidence P(D) was evaluated4 as a function of C. The kernel amplitude Ko was varied simultaneously such that C Ko and hence the SVM solution itself remained unchanged. Because the data set was generated artificially from the probability model (3), the 'true' value of C = 2 was known; in spite of the rather crude approximation for Q(D), the maximum of the full evidence P(D) identifies C ~ 1.8 quite close to the truth . The approximate class probability prediction P(y = 11x, D) for this value of C is also plotted in Fig. 2; it overestimates the noise in the target somewhat. Note that P(ylx, D) was obtained simply by inserting the MAP values 8*(x) into (5). In a proper Bayesian treatment, an average over the posterior distribution P(OID) should of course be taken; I leave this for future work. 4The normalization factor Nn was estimated, for the assumed uniform input density Q(x) of the example, by sampling from the GP prior Q(9) . If Q(x) is unknown, the empirical training input distribution can be used as a proxy, and one samples instead from a multivariate Gaussian for the 9(xd with covariance matrix K(Xi , Xj). This gave very similar values of In Nn in the example, even when only a subset of 30 training inputs was used. Probabilistic Methods for Support Vector Machines 355 In summary, I have described a probabilistic framework for SVM classification. It gives an intuitive understanding of the effect of the kernel, which determines a Gaussian process prior. More importantly, it also allows a properly normalized evidence to be defined; from this, optimal values of hyperparameters such as the noise parameter C, and corresponding error bars, can be derived. Future work will have to include more comprehensive experimental tests of the simple Laplacetype estimate of the (naive) evidence Q(D) that I have given, and comparison with other approaches. These include variational methods; very recent experiments with a Gaussian approximation for the posterior P(9ID), for example, seem promising [6]. Further improvement should be possible by dropping the restriction to a 'factor-analysed' covariance form [6]. (One easily shows that the optimal Gaussian covariance matrix is (D + K- 1 )-1, parameterized only by a diagonal matrix D.) It will also be interesting to compare the Laplace and Gaussian variational results for the evidence with those from the 'cavity field' approach of [10]. Acknowledgements It is a pleasure to thank Tommi Jaakkola, Manfred Opper, Matthias Seeger, Chris Williams and Ole Winther for interesting comments and discussions, and the Royal Society for financial support through a Dorothy Hodgkin Research Fellowship. References [1] C J C Burges. A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, 2:121-167, 1998. [2] A J Smola and B Scholkopf. A tutorial on support vector regression. 1998. Neuro COLT Technical Report TR-1998-030; available from http://svm.first.gmd.de/. [3] B Scholkopf, C Burges, and A J Smola. Advances in Kernel Methods: Su.pport Vector Machines . MIT Press, Cambridge, MA , 1998. [4) B Scholkopf, P Bartlett, A Smola, and R Williamson. Shrinking the tube: a new support vector regression algorithm. In NIPS 11. [5] N Cristianini, C Campbell, and J Shawe-Taylor. Dynamically adapting kernels in support vector machines. In NIPS 11. [6] M Seeger. Bayesian model selection for Support Vector machines, Gaussian processes and other kernel classifiers. Submitted to NIPS 12. [7] G Wahba. Support vector machines, reproducing kernel Hilbert spaces and the randomized GACV. Technical Report 984, University of Wisconsin, 1997. [8] T S Jaakkola and D Haussler. Probabilistic kernel regression models. In Proceedings of The 7th International Workshop on Artificial Intelligence and Statistics. To appear. [9] A J Smola, B Scholkopf, and K R Muller. The connection between regularization operators and support vector kernels. Neu.ral Networks, 11:637-649, 1998. [10] M Opper and 0 Winther. Gaussian process classification and SVM: Mean field results and leave-one-out estimator. In Advances in Large Margin Classifiers. MIT Press. To appear. [11] P Sollich. Probabilistic interpretation and Bayesian methods for Support Vector Machines. Submitted to ICANN 99. [12] C K I Williams. Prediction with Gaussian processes: From linear regression to linear prediction and beyond. In M I Jordan, editor, Learning and Inference in Graphical Models, pages 599-621. Kluwer Academic, 1998.
1711 |@word private:1 determinant:1 polynomial:1 open:1 grey:1 km:1 covariance:7 tr:1 solid:2 contains:1 recovered:1 comparing:1 analysed:1 dx:3 numerical:1 additive:1 shape:1 plot:2 alone:2 intelligence:1 manfred:1 contribute:1 hyperplanes:1 sigmoidal:2 scholkopf:4 inside:1 roughly:1 actual:2 lib:1 becomes:2 underlying:3 moreover:1 minimizes:1 lone:1 differing:1 transformation:1 nj:1 guarantee:1 quantitative:1 tackle:1 xd:4 classifier:2 uk:1 control:2 unit:2 appear:4 overestimate:1 safety:1 understood:1 tends:1 limit:1 io:4 id:4 fluctuation:1 black:2 dynamically:1 definite:1 writes:1 procedure:1 empirical:1 adapting:1 yio:2 integrating:1 spite:1 suggest:1 cannot:2 onto:1 close:1 selection:1 operator:1 restriction:1 equivalent:1 map:8 deterministic:1 williams:2 l:1 estimator:1 haussler:1 importantly:1 financial:1 laplace:2 construction:1 target:5 suppose:2 gps:1 us:1 recognition:1 xi9:1 bottom:2 region:4 ensures:1 intuition:2 ideally:1 cristianini:1 trained:2 depend:2 incur:1 gacv:1 easily:1 joint:1 kcl:1 describe:1 london:2 effective:3 ole:1 artificial:1 choosing:2 whose:1 encoded:1 larger:2 widely:1 plausible:1 heuristic:1 quite:1 statistic:1 think:1 gp:7 itself:2 matthias:1 product:2 inserting:1 relevant:1 intuitive:3 diverges:1 leave:2 illustrate:1 coupling:1 ac:1 progress:1 eq:2 come:2 tommi:1 closely:1 dio:3 wc2r:1 assign:3 singularity:1 correction:1 around:1 exp:7 lm:1 yilxi:1 vary:1 adopt:1 smallest:1 dxl:1 estimation:1 sensitive:1 individually:2 largest:2 rough:1 clearly:2 mit:2 gaussian:16 modified:1 rather:3 factorizes:1 jaakkola:2 probabilistically:1 derived:4 focus:1 yo:1 properly:1 improvement:1 ral:1 likelihood:10 indicates:1 contrast:2 seeger:2 sense:3 posteriori:1 inference:2 dependent:4 nn:10 spurious:1 overall:1 classification:5 among:1 colt:1 retaining:1 development:1 integration:1 marginal:5 equal:2 once:1 never:1 field:2 sampling:3 identical:1 survive:1 future:2 others:1 report:2 simultaneously:1 comprehensive:1 argmax:1 mining:1 yielding:1 amenable:1 integral:4 intense:1 taylor:1 circle:1 plotted:1 minimal:1 fitted:1 maximization:2 deviation:1 subset:2 entry:1 imperfectly:1 uniform:1 density:2 winther:2 randomized:1 international:1 probabilistic:15 squared:2 tube:1 corner:1 leading:1 li:5 toy:1 de:1 b2:4 coefficient:1 depends:1 ornstein:1 view:2 break:1 di9:2 minimize:1 square:2 variance:2 maximized:1 ofthe:1 conceptually:1 bayesian:4 submitted:2 influenced:1 email:1 neu:1 tunable:1 dataset:1 treatment:1 knowledge:2 hilbert:1 amplitude:3 actually:2 campbell:1 higher:1 unambiguously:2 reflected:1 evaluated:2 just:2 smola:4 until:1 hand:1 ei:4 replacing:1 su:1 defines:2 logistic:1 effect:4 normalized:4 true:3 remedy:1 hence:2 regularization:1 b2b:1 nonzero:2 white:2 inq:1 width:1 unambiguous:1 noted:1 unnormalized:1 illuminates:1 outline:1 complete:1 oid:1 interpreting:1 variational:2 recently:2 overview:1 discussed:1 interpretation:2 kluwer:1 interpret:1 cambridge:1 llx:4 ai:3 smoothness:1 mathematics:1 shawe:1 etc:1 add:1 posterior:5 multivariate:1 recent:2 binary:1 nln:1 yi:9 muller:1 seen:2 somewhat:1 determine:1 dashed:2 full:1 smooth:1 technical:2 academic:1 calculation:1 cross:3 lin:1 prediction:6 neuro:1 regression:4 ko:9 basic:1 essentially:1 kernel:25 normalization:3 represent:2 uhlenbeck:1 fellowship:1 separately:1 dorothy:1 crucial:1 comment:1 subject:2 tend:1 seem:1 jordan:1 call:1 near:1 xj:1 gave:1 wahba:1 idea:1 det:2 shift:2 expression:1 bartlett:1 penalty:3 peter:2 clear:1 ylx:4 amount:1 svms:7 simplest:1 reduced:2 http:1 specifies:1 gmd:1 tutorial:3 shifted:1 estimated:1 write:1 hyperparameter:1 dropping:1 express:1 group:1 changing:1 verified:1 iyik:1 wand:4 everywhere:1 uncertainty:4 parameterized:1 hodgkin:1 decision:7 submatrix:1 correspondence:1 quadratic:2 activity:1 cko:3 constraint:2 flat:1 integrand:5 separable:2 department:1 smaller:1 sollich:6 making:1 taken:1 remains:4 slack:2 know:1 letting:1 end:1 available:2 apply:1 obey:1 alternative:1 top:2 remaining:2 ensure:1 include:2 graphical:1 hinge:3 const:1 giving:1 society:1 unchanged:1 implied:1 objective:1 move:1 question:1 quantity:1 added:1 diagonal:2 gradient:1 discouraged:1 distance:1 separate:1 pleasure:1 thank:1 restart:1 chris:1 trivial:1 length:4 providing:1 equivalently:1 difficult:1 greyscale:2 holding:1 negative:2 ofw:1 guideline:1 proper:1 unknown:1 finite:1 defining:2 communication:1 varied:2 reproducing:1 arbitrary:2 community:1 introduced:1 required:1 widened:1 specified:1 connection:1 narrow:1 nip:3 beyond:2 bar:5 pattern:1 summarize:1 built:1 royal:1 critical:1 suitable:1 identifies:1 naive:4 prior:22 understanding:1 acknowledgement:1 discovery:1 relative:1 wisconsin:1 loss:3 generation:1 interesting:2 validation:3 proxy:1 editor:1 uncorrelated:1 normalizes:1 row:1 course:1 summary:1 last:2 free:1 keeping:1 burges:2 taking:1 correspondingly:1 differentiating:1 distributed:1 regard:1 boundary:3 opper:3 author:1 made:1 approximate:2 keep:2 cavity:1 assumed:1 xi:20 don:1 continuous:1 latent:2 why:1 promising:1 transfer:2 reasonably:1 expanding:1 obtaining:1 expansion:1 williamson:1 cl:3 artificially:1 icann:1 main:1 linearly:3 whole:1 noise:5 hyperparameters:5 arise:1 allowed:1 fig:4 shrinking:1 yib:1 deterministically:1 exponential:2 xl:4 crude:1 ix:1 down:1 remained:1 discarding:1 svm:25 evidence:21 intractable:1 workshop:1 magnitude:1 illustrates:1 sparseness:2 margin:5 gap:7 simply:2 likely:1 prevents:1 strand:1 llwl12:1 corresponds:1 truth:1 determines:1 ma:1 conditional:1 identity:1 king:1 rbf:2 towards:1 absence:1 change:3 determined:2 except:1 typical:2 experimental:1 college:1 support:18 violated:1 avoiding:1
778
1,712
Bayesian Transduction Thore Graepel, Ralf Herbrich and Klaus Obermayer Department of Computer Science Technical University of Berlin Franklinstr. 28/29, 10587 Berlin, Germany {graepeI2, raith, oby} @cs.tu-berlin.de Abstract Transduction is an inference principle that takes a training sample and aims at estimating the values of a function at given points contained in the so-called working sample as opposed to the whole of input space for induction. Transduction provides a confidence measure on single predictions rather than classifiers - a feature particularly important for risk-sensitive applications. The possibly infinite number of functions is reduced to a finite number of equivalence classes on the working sample. A rigorous Bayesian analysis reveals that for standard classification loss we cannot benefit from considering more than one test point at a time. The probability of the label of a given test point is determined as the posterior measure of the corresponding subset of hypothesis space. We consider the PAC setting of binary classification by linear discriminant functions (perceptrons) in kernel space such that the probability of labels is determined by the volume ratio in version space. We suggest to sample this region by an ergodic billiard. Experimental results on real world data indicate that Bayesian Transduction compares favourably to the well-known Support Vector Machine, in particular if the posterior probability of labellings is used as a confidence measure to exclude test points of low confidence. 1 Introduction According to Vapnik [9], when solving a given problem one should avoid solving a more general problem as an intermediate step. The reasoning behind this principle is that in order to solve the more general task resources may be wasted or compromises may have to be made which would not have been necessary for the solution of the problem at hand. A direct application of this common-sense principle reduces the more general problem of inferring a functional dependency on the whole of input space to the problem of estimating the values of a function at given points (working sample), a paradigm referred to as transductive inference. More formally, given a probability measure PXY on the space of data X x y = X x {-I, +1}, a training sample S = {(Xl, yd, ... ,(Xl, Yi)} is generated i.i.d. according to PXY. Additional m data points W = {Xl+I, ... ,Xi+ m } are drawn: the working sample. The goal is to label the objects of the working sample W using a fixed set 1{ of functions 457 Bayesian Transduction f : X Jo--ot {-I, +I} so as to minimise a predefined loss. In . contrast, inductive inference, aims at choosing a single function It E 1i best suited to capture the dependency expressed by the unknown P XY . Obviously, if we have a transductive algorithm A (W, S, 1i) that assigns to each working sample W a set of labels given the training sample S and the set 1i offunctions, we can define a function fs : X Jo--ot {-I, +1} by fs (x) A ({x} ,S, 1i) as a result ofthe transduction algorithm. There are two crucial differences to induction, however: i) A ({ x} , S, 1i) is not restricted to select a single decision function f E 1i for each x, ii) a transduction algorithm can give performance guarantees on particular labellings instead of functions. In practical applications this difference may be of great importance. = After all, in risk sensitive applications (medical diagnosis, financial and critical control applications) it often matters to know how confident we are about a given prediction. In this case a general confidence measure of the classifier w.r. t. the whole input distribution would not provide the desired warranty at all. Note that for linear classifiers some guarantee can be obtained by the margin [7] which in Section 4 we will demonstrate to be too coarse a confidence measure. The idea of transduction was put forward in [8], where also first algorithmic ideas can be found . Later [1] suggested an algorithm for transduction based on linear programming and [3] highlighted the need for confidence measures in transduction. The paper is structured as follows: A Bayesian approach to transduction is formulated in Section 2. In Section 3 the function class of kernel perceptrons is introduced to which the Bayesian transduction scheme is applied . For the estimation of volumes in parameter space we present a kernel billiard as an efficient sampling technique. Finally, we demonstrate experimentally in Section 4 how the confidence measure for labellings helps Bayesian Transduction to achieve low generalisation error at a low rejection rate of test points and thus to outperform Support Vector Machines (SVMs). 2 Bayesian Transductive Classification Suppose we are given a training sample S = {(Xl, YI) , . .. , (Xl, Yl )} drawn i.i.d. from PXY and a working sample W = {XHI,' " , XHm} drawn i.i.d. from P x . Given a prior PH over the set 1i of functions and a likelihood P (Xy)lIH=f we obtain a posterior probability PHI(Xy)l=s ~f PHIS by Bayes' rule. This posterior measure induces a probability measure on labellings b E {-I, +l}m of the working sample byl (1) For the sake of simplicity let us assume a PAC style setting, i.e. there exists a in the space 1i such that PYlx=x (y) = 6 (y (x)). In this case one function can define the so-called version-space as the set of functions that is consistent with the training sample r r (2) outside which the posterior PHIS vanishes. Then Pymls,w (b) represents the prior measure of functions consistent with the training sample S and the labelling b on the working sample W normalised by the prior measure of functions consistent with S alone. The measure PH can be used to incorporate prior knowledge into 1 Note that the number of different labellings b implement able by 1l is bounded above by the value of the growth function IIu (JWI) [8, p . 321]. T. Graepe/, R. Herbrich and K. Obennayer 458 the inference process. If no such knowledge is available, considerations of symmetry may lead to "uninformative" priors. Given the measure PYFnIS,W over labellings, in order to arrive at a risk minimal decision w.r.t. the labelling we need to define a loss function I : ym X ym I---t IR+ between labellings and minimise its expectation, R (b, S, W) = EYFnIS,W [I (b, ym)] = I (b, b/) PYFnIS,W (b /) , (3) 2: {b'} 2m where the summation runs over all the sample. Let us consider two scenarios: possible labellings b ' of the working 1. A 0-1-loss on the exact labelling b, i.e. for two labellings band b ' m Ie (b, b/) = 1- II 6 (b i - bD ?} i=l Re (b, S, W) =1- PYFnIS,W (b) . (4) = In this case choosing the labelling be argminb Re (b, S, W) of the highest joint probability Pymls,w (b) minimises the risk. This non-labelwise loss is appropriate if the goal is to exactly identify a combination of labels, e.g. the combination of handwritten digits defining a postal zip code. Note that classical SVM transduction (see, e.g. [8, 1]) by maximising the margin on the combined training and working sample approximates this strategy and hence does not minimise the standard classification risk on single instances as intended. 2. A 0-1-10ss on the single labels bi, i.e. for two labellings band b ' 1 m 1$ (b, b/) = (1- 6 (b i - bD) , (5) 2: m i=l R$ (b, S, W) ! f 2: (1- 6 (bi - b~)) Pymls,w (b /) i=l {b'} 1 m - 2: (1- PHIs ({f: f(Xl+i) = bd)) . m i=l Due to the independent treatment of the loss at working sample points the risk R$ (b, S, W) is minimised by the labelling of highest marginal probability of the labels, i.e. bi = argmaXyEY PHIs ({f: f(Xl+i) = y}). Thus in the case of the labelwise loss (5) a working sample of m > 1 point does not offer any advantages over larger working samples w.r. t. the Bayes-optimal decision. Since this corresponds to the standard classification setting, we will restrict ourselves to working samples of size m = 1, i.e. to one working point Xl+1. 3 Bayesian Transduction by Volume 3.1 The Kernel Perceptron We consider transductive inference for the class of kernel perceptrons. The decision functions are given by f (x) = sign ?w, q, (xl) >') = sign (t a;k (x;, X)) l w = 2: (?itP (xd E :F , i=l 459 Bayesian Transduction Figure 1: Schematic view of data space (left) and parameter space ( right) for a classification toy example. Using the duality given by (w , 4> (x)):F = 0 data points on the left correspond to hyperplanes on the right, while hyperplanes on the left can be thought of as points on the right . where the mapping 4> : X t--+ :F maps from input space X to a feature space :F completely determined by the inner product function (kernel) k : X x X t--+ IR (see [9 , 10]) . Given a training sample S = {(Xi , Yi)}~=l we can define the version space - the set of all perceptrons compatible with the training data - as in (2) having the additional constraint Ilwll:F = 1 ensuring uniqueness. In order to obtain a prediction on the label b1 of the working point Xl+l we note that Xl+l may bisects the volume V of version space into two sub-volumes V+ and V-, where the perceptrons in V+ would classify Xl+l as b1 = +1 and those in V- as b1 = -l. The ratio p+ = V+ IV is the probability of the labelling b1 = +1 given a uniform prior PH over wand the class of kernel perceptrons, accordingly for b1 = -1 (see Figure 1) . Already Vapnik in [8, p. 323] noticed that it is troublesome to estimate sub- volumes of version space. As the solution to this problem we suggest to use a billiard algorithm . 3.2 Kernel Billiard for Volume Estimation The method of playing billiard in version space was first introduced by Rujan [6] for the purpose of estimating its centre of mass and consequently refined and extended to kernel spaces by [4]. For Bayesian Transduction the idea is to bounce the billiard ball in version space and to record how much time it spends in each of the sub-volumes of interest. Under the assumption of ergodicity [2] w.r .t. the uniform measure in the limit the accumulated flight times for each sub-volume are proportional to the sub-volume itself. Since the trajectory is located in :F each position wand direction v of the ball can be expressed as linear combinations of the 4> (xd , i.e. l l W= L Q:i4> (Xi) i=l v = L ,Bi4> (Xi) i=l l (w, v):F =L Q:i,Bjk (Xi, Xj) i,j=l where 0:, {3 are real vectors with f components and fully determine the state of the billiard. The algorithm for the determination of the label b1 of Xl+l proceeds as follows : 1. Initialise the starting position Wo in V (S) using any kernel perceptron algorithm that achieves zero training error (e .g . SVM [9]) . Set V+ = V- = O. T. Graepel, R. Herbrich and K. Obennayer 460 2. Find the closest boundary of V (S) starting from current w into direction v, where the flight times Tj for all points including Xl+1 are determined using (w,tP(Xj?:r (v,tP(Xj)):r . The smallest positive flight time Tc = minj :T;>o Tj in kernel space corresponds to the closest data point boundary tP (xc) on the hypersphere. Note, that if Tc -7 00 we randomly generate a direction v pointing towards version space, i.e. y (v, tP (x)):r > 0 assuming the last bounce was at tP (x). 3. Calculate the ball's new position w' according to W , = + TcV Ilw + Tcvll:r w . (1 - /2) Calculate the distance tf = Ilw - w'llsphere = arccos Ilw - w'lI;' on the hypersphere and add it to the volume estimate VY corresponding to the current label y = sign (w + w', tP (Xl+d):r)? If the test point tP (xl+d was hit, i.e. c = l + 1, keep the old direction vector v. Otherwise update to the reflection direction v', v' = v - 2 (v, tP (xc) ):r tP (xc) . Go back to step 2 unless the stopping criterion (8) is met. Note that in practice one trajectory can be calculated in advance and can be used for all test points. The estimators of the probability of the labellings are then given by p+ = V+ /(V+ + V-) and p = V- /(v+ + V-). Thus, the algorithm outputs b1 with confidence Ctrans according to ~ Ctrans def argmaXyEY iY' def (2 . max (pi" ,p) - 1) E [0, 1] . , (6) (7) Note that the Bayes Point Machine (BPM) [4] aims at an optimal approximation of the transductive classification (6) by a single function f E 1{ and that the well known SVM can be viewed as an approximation of the BPM by the centre of the largest ball in version space. Thus, treating the real valued output If(xl+1) I ~f G;nd of SVM classifiers as a confidence measure can be considered an approximation of (7). The consequences will be demonstrated experimentally in the following section. Disregarding the issue of mixing time [2] and the dependence of trajectories we assume for the stopping criterion that the fraction pt of time tt spent in volume V+ on trajectory i of length (tt + f;) is a random variable having expectation p+ . Hoeffding's inequality [5] bounds the probability of deviation from the expectation p+ by more than f, P (!; t ! - p+ <: ,) ~ exp (-2n,2) ~ ~. p (8) Thus if we want the deviation f from the true label probability to be less than f < 0.05 with probability at least 1 - T} = 0.99 we need approximately n R:j 1000 bounces. The computational effort of the above algorithm for a working set of size m is of order 0 (nl (m + l)). 461 Bayesian Transduction 1= 100--1 2 o . . o ~~----~----~--~----~--~ 0.00 0.05 0.10 0 lei 020 rejeclion rate (a) o~ 000 _ _~_ _~_ _~~_ _~_ _~~ 0.05 0 .10 0 lei 020 0.25 030 rejection rate (b) Figure 2: Generalisation error vs. rejection rate for Bayesian Transduction and SVMs for the thyroid data set (0' = 3) (a) and the heart data set (0' = 10). The error bars in both directions indicate one standard deviation of the estimated means. The upper curve depicts the result for the SVM algorithm; the lower curve is the result obtained by Bayesian Transduction. 4 Experimental Results We focused on the confidence Ctrans Bayesian Transduction provides together with the prediction b1 of the label. If the confidence Ctrans reflects reliability of a label estimate at a given test point then rejecting those test points whose predictions carry low confidence should lead to a reduction in generalisation error on the remaining test points . In the experiments we varied a rejection threshold () between [0, 1] thus obtaining for each () a rejeection rate together with an estimate of the generalisation error at non-rejected points. Both these curves were linked by their common ()-axis resulting in a generalisation error versus rejection rate plot. We used the UCI 2 data sets thyroid and heart because they are medical applications for which the confidence of single predictions is particularly important. Also a high rejection rate due to too conservative a confidence measure may incur considerable costs. We trained a Support Vector Machine using RBF kernels k (x, x') = exp ( -llx - x'1l2 /20'2) with 0' chosen such as to insure the existence of a version space. We used 100 different training samples obtained by random 60%:40% splits of the whole data set. The margin Clnd of each test point was calculated as a confidence measure of SVM classifications. For comparison we determined the labels b1 and resulting confidences Ctrans using the Bayesian Transduction algorithm (see Section 3) with the same value of the kernel parameter. Since the rejection for the Bayesian Transduction was in both cases higher than for SVMs at the same level () we determined ()max which achieves the same rejection rate for the SVM confidence measures as Bayesian Transduction achieves at () = 1 (thyroid: ()max = 2.15, heart: ()max 1.54). The results for the two data sets are depicted in Figure 2. = In the thyroid example Figure 2 (a) one can see that Ctrans is indeed an appropriate indicator of confidence: at a rejection rate of approximately 20% the generalisation error approaches zero at minimal variance. For any desired generalisation error Bayesian Transduction needs to reject significantly less examples of the test set as compared to SVM classifiers, e.g. 4% less at 2.3% generalisation error. The results of the heart data set show even more pronounced characteristics w.r.t. to the rejection 2UCI University of California at Irvine: Machine Learning Repository T. Graepe/, R. Herbrich and K. Obermayer 462 rate. Note that those confidence measures considered cannot capture the effects of noise in the data which leads to a generalisation error of 16.4% even at maximal rejection () = 1 corresponding to the Bayes error under the given function class. 5 Conclusions and FUture Work In this paper we a presented a Bayesian analysis of transduction. The required volume estimates for kernel perceptrons in version space are performed by an ergodic billiard in kernel space. Most importantly, transduction not only determines the label of a given point but also returns a confidence measure of the classification in the form of the probability of the label under the model. Using this confidence measure to reject test examples then lead to improved generalisation error over SVMs. The billiard algorithm can be extended to the case of non-zero training error by allowing the ball to penetrate walls, a property that is captured by adding a constant>. to the diagonal of the kernel matrix [4] . Further research will aim at the discovery of PAC-Bayesian bounds on the generalisation error of transduction. Acknowledgements We are greatly indebted to U. Kockelkorn for many interesting suggestions and discussions . This project was partially funded by Technical University of Berlin via FIP 13/41. References [1] K. Bennett. Advances in Kernel Methods - Support Vector Learning, chapter 19, Combining Support Vector and Mathematical Programming Methods for Classification, pages 307-326. MIT Press, 1998. [2] I. Cornfeld, S. Fomin, and Y. Sinai. Ergodic Theory. Springer Verlag, 1982. [3] A. Gammerman, V. Vovk, and V. Vapnik. Learning by transduction. In Proceedings of Uncertainty in AI, pages 148-155, Madison, Wisconsin, 1998. [4] R. Herbrich, T. Graepel, and C. Campbell. Bayesian learning in reproducing kernel Hilbert spaces. Technical report, Technical University Berlin, 1999. TR 99-1l. [5] W. Hoeffding. Probability inequalities for sums of bounded random variables. Journal of the American Statistical Association, 58:13-30, 1963. [6] P. Rujan. Playing billiard in version space. Neural Computation, 9:99-122, 1997. [7] J. Shawe-Taylor. Confidence estimates of classification accuracy on new examples. Technical report, Royal Holloway, University of London, 1996. NC2-TR-1996-054. [8] V. Vapnik. Estimation of Dependences Based on Empirical Data. Springer, 1982. [9] V. Vapnik. The Nature of Statistical Learning Theory. Springer, 1995. [10] G. Wahba. Spline Models for Observational Data. Society for Industrial and Applied Mathematics, Philadelphia, 1990.
1712 |@word repository:1 version:12 nd:1 tr:2 carry:1 reduction:1 itp:1 current:2 bd:3 offunctions:1 treating:1 plot:1 update:1 v:1 alone:1 accordingly:1 jwi:1 record:1 hypersphere:2 coarse:1 provides:2 postal:1 herbrich:5 billiard:10 hyperplanes:2 mathematical:1 direct:1 indeed:1 considering:1 project:1 estimating:3 bounded:2 insure:1 mass:1 spends:1 guarantee:2 growth:1 xd:2 exactly:1 classifier:5 hit:1 control:1 medical:2 positive:1 xhm:1 limit:1 consequence:1 troublesome:1 yd:1 approximately:2 argminb:1 equivalence:1 bi:3 bjk:1 practical:1 practice:1 implement:1 digit:1 empirical:1 thought:1 reject:2 significantly:1 confidence:22 suggest:2 cannot:2 put:1 risk:6 map:1 demonstrated:1 go:1 starting:2 ergodic:3 focused:1 iiu:1 simplicity:1 penetrate:1 assigns:1 fomin:1 rule:1 estimator:1 importantly:1 financial:1 ralf:1 initialise:1 pt:1 suppose:1 exact:1 programming:2 hypothesis:1 particularly:2 located:1 capture:2 calculate:2 region:1 highest:2 vanishes:1 trained:1 solving:2 compromise:1 incur:1 completely:1 joint:1 chapter:1 london:1 oby:1 klaus:1 choosing:2 outside:1 refined:1 whose:1 larger:1 solve:1 valued:1 s:1 otherwise:1 transductive:5 highlighted:1 itself:1 obviously:1 advantage:1 argmaxyey:2 product:1 maximal:1 tu:1 uci:2 combining:1 mixing:1 achieve:1 pronounced:1 object:1 help:1 spent:1 minimises:1 c:1 indicate:2 met:1 direction:6 observational:1 wall:1 summation:1 considered:2 exp:2 great:1 algorithmic:1 mapping:1 pointing:1 achieves:3 smallest:1 purpose:1 uniqueness:1 estimation:3 label:16 sensitive:2 largest:1 bisects:1 tf:1 reflects:1 mit:1 aim:4 rather:1 avoid:1 likelihood:1 greatly:1 contrast:1 rigorous:1 industrial:1 sense:1 inference:5 stopping:2 accumulated:1 bpm:2 germany:1 issue:1 classification:11 arccos:1 marginal:1 having:2 sampling:1 represents:1 future:1 report:2 spline:1 xhi:1 randomly:1 intended:1 ourselves:1 interest:1 nl:1 behind:1 tj:2 lih:1 predefined:1 necessary:1 xy:3 unless:1 iv:1 old:1 taylor:1 desired:2 re:2 minimal:2 instance:1 classify:1 tp:9 graepe:2 cost:1 deviation:3 subset:1 uniform:2 too:2 dependency:2 combined:1 confident:1 ie:1 yl:1 minimised:1 ym:3 iy:1 together:2 jo:2 opposed:1 possibly:1 hoeffding:2 american:1 style:1 return:1 toy:1 li:1 exclude:1 de:1 matter:1 later:1 view:1 performed:1 linked:1 bayes:4 nc2:1 ir:2 accuracy:1 variance:1 characteristic:1 correspond:1 ofthe:1 identify:1 bayesian:22 handwritten:1 rejecting:1 trajectory:4 indebted:1 minj:1 irvine:1 treatment:1 knowledge:2 graepel:3 hilbert:1 back:1 campbell:1 higher:1 improved:1 ergodicity:1 rejected:1 working:18 hand:1 flight:3 favourably:1 lei:2 thore:1 effect:1 true:1 inductive:1 hence:1 criterion:2 tt:2 demonstrate:2 reflection:1 reasoning:1 consideration:1 common:2 functional:1 volume:13 association:1 approximates:1 llx:1 ai:1 mathematics:1 centre:2 shawe:1 reliability:1 funded:1 add:1 posterior:5 closest:2 scenario:1 verlag:1 inequality:2 binary:1 yi:3 pylx:1 captured:1 additional:2 zip:1 determine:1 paradigm:1 ii:2 reduces:1 technical:5 determination:1 offer:1 labelwise:2 schematic:1 prediction:6 ensuring:1 expectation:3 kernel:18 uninformative:1 want:1 crucial:1 ot:2 intermediate:1 split:1 xj:3 restrict:1 wahba:1 inner:1 idea:3 minimise:3 bounce:3 effort:1 wo:1 f:2 byl:1 band:2 ph:3 induces:1 svms:4 reduced:1 generate:1 outperform:1 vy:1 sign:3 estimated:1 gammerman:1 diagnosis:1 threshold:1 drawn:3 wasted:1 fraction:1 sum:1 wand:2 run:1 uncertainty:1 franklinstr:1 arrive:1 decision:4 def:2 bound:2 pxy:3 i4:1 constraint:1 sake:1 thyroid:4 department:1 structured:1 according:4 combination:3 ball:5 labellings:11 restricted:1 heart:4 resource:1 know:1 available:1 appropriate:2 existence:1 remaining:1 madison:1 xc:3 classical:1 society:1 noticed:1 already:1 strategy:1 dependence:2 diagonal:1 obermayer:2 distance:1 berlin:5 discriminant:1 induction:2 maximising:1 assuming:1 code:1 length:1 ratio:2 ilw:3 obennayer:2 unknown:1 clnd:1 allowing:1 upper:1 finite:1 defining:1 extended:2 varied:1 reproducing:1 introduced:2 required:1 california:1 warranty:1 able:1 suggested:1 proceeds:1 bar:1 including:1 max:4 royal:1 critical:1 indicator:1 scheme:1 axis:1 philadelphia:1 prior:6 l2:1 discovery:1 acknowledgement:1 wisconsin:1 loss:7 fully:1 interesting:1 suggestion:1 proportional:1 versus:1 consistent:3 principle:3 playing:2 pi:1 compatible:1 last:1 sinai:1 normalised:1 perceptron:2 benefit:1 boundary:2 calculated:2 curve:3 world:1 forward:1 made:1 fip:1 keep:1 reveals:1 b1:9 xi:5 nature:1 symmetry:1 obtaining:1 rujan:2 whole:4 noise:1 referred:1 depicts:1 transduction:29 sub:5 inferring:1 position:3 xl:17 pac:3 disregarding:1 svm:8 exists:1 ilwll:1 vapnik:5 adding:1 importance:1 labelling:6 margin:3 rejection:11 suited:1 tc:2 depicted:1 expressed:2 contained:1 phi:5 partially:1 springer:3 corresponds:2 determines:1 goal:2 formulated:1 viewed:1 consequently:1 rbf:1 towards:1 bennett:1 considerable:1 experimentally:2 infinite:1 determined:6 generalisation:11 vovk:1 conservative:1 called:2 duality:1 experimental:2 perceptrons:7 formally:1 select:1 holloway:1 support:5 incorporate:1
779
1,713
Policy Gradient Methods for Reinforcement Learning with Function Approximation Richard S. Sutton, David McAllester, Satinder Singh, Yishay Mansour AT&T Labs - Research, 180 Park Avenue, Florham Park, NJ 07932 Abstract Function approximation is essential to reinforcement learning, but the standard approach of approximating a value function and determining a policy from it has so far proven theoretically intractable. In this paper we explore an alternative approach in which the policy is explicitly represented by its own function approximator, independent of the value function, and is updated according to the gradient of expected reward with respect to the policy parameters. Williams's REINFORCE method and actor-critic methods are examples of this approach. Our main new result is to show that the gradient can be written in a form suitable for estimation from experience aided by an approximate action-value or advantage function. Using this result, we prove for the first time that a version of policy iteration with arbitrary differentiable function approximation is convergent to a locally optimal policy. Large applications of reinforcement learning (RL) require the use of generalizing function approximators such neural networks, decision-trees, or instance-based methods. The dominant approach for the last decade has been the value-function approach, in which all function approximation effort goes into estimating a value function, with the action-selection policy represented implicitly as the "greedy" policy with respect to the estimated values (e.g., as the policy that selects in each state the action with highest estimated value). The value-function approach has worked well in many applications, but has several limitations. First, it is oriented toward finding deterministic policies, whereas the optimal policy is often stochastic, selecting different actions with specific probabilities (e.g., see Singh, Jaakkola, and Jordan, 1994). Second, an arbitrarily small change in the estimated value of an action can cause it to be, or not be, selected. Such discontinuous changes have been identified as a key obstacle to establishing convergence assurances for algorithms following the value-function approach (Bertsekas and Tsitsiklis, 1996). For example, Q-Iearning, Sarsa, and dynamic programming methods have all been shown unable to converge to any policy for simple MDPs and simple function approximators (Gordon, 1995, 1996; Baird, 1995; Tsitsiklis and van Roy, 1996; Bertsekas and Tsitsiklis, 1996). This can occur even if the best approximation is found at each step before changing the policy, and whether the notion of "best" is in the mean-squared-error sense or the slightly different senses of residual-gradient, temporal-difference, and dynamic-programming methods. In this paper we explore an alternative approach to function approximation in RL. 1058 R. S. Sutton, D. McAl/ester. S. Singh and Y. Mansour Rather than approximating a value function and using that to compute a deterministic policy, we approximate a stochastic policy directly using an independent function approximator with its own parameters. For example, the policy might be represented by a neural network whose input is a representation of the state, whose output is action selection probabilities, and whose weights are the policy parameters. Let 0 denote the vector of policy parameters and p the performance of the corresponding policy (e.g., the average reward per step). Then, in the policy gradient approach, the policy parameters are updated approximately proportional to the gradient: ap ~O~CtaO' (1) where Ct is a positive-definite step size. If the above can be achieved, then 0 can usually be assured to converge to a locally optimal policy in the performance measure p. Unlike the value-function approach, here small changes in 0 can cause only small changes in the policy and in the state-visitation distribution. In this paper we prove that an unbiased estimate of the gradient (1) can be obtained from experience using an approximate value function satisfying certain properties. Williams's (1988, 1992) REINFORCE algorithm also finds an unbiased estimate of the gradient, but without the assistance of a learned value function. REINFORCE learns much more slowly than RL methods using value functions and has received relatively little attention. Learning a value function and using it to reduce the variance of the gradient estimate appears to be ess~ntial for rapid learning. Jaakkola, Singh and Jordan (1995) proved a result very similar to ours for the special case of function approximation corresponding to tabular POMDPs. Our result strengthens theirs and generalizes it to arbitrary differentiable function approximators. Konda and Tsitsiklis (in prep.) independently developed a very simialr result to ours. See also Baxter and Bartlett (in prep.) and Marbach and Tsitsiklis (1998). Our result also suggests a way of proving the convergence of a wide variety of algorithms based on "actor-critic" or policy-iteration architectures (e.g., Barto, Sutton, and Anderson, 1983; Sutton, 1984; Kimura and Kobayashi, 1998). In this paper we take the first step in this direction by proving for the first time that a version of policy iteration with general differentiable function approximation is convergent to a locally optimal policy. Baird and Moore (1999) obtained a weaker but superficially similar result for their VAPS family of methods. Like policy-gradient methods, VAPS includes separately parameterized policy and value functions updated by gradient methods. However, VAPS methods do not climb the gradient of performance (expected long-term reward), but of a measure combining performance and valuefunction accuracy. As a result, VAPS does not converge to a locally optimal policy, except in the case that no weight is put upon value-function accuracy, in which case VAPS degenerates to REINFORCE. Similarly, Gordon's (1995) fitted value iteration is also convergent and value-based, but does not find a locally optimal policy. 1 Policy Gradient Theorem We consider the standard reinforcement learning framework (see, e.g., Sutton and Barto, 1998), in which a learning agent interacts with a Markov decision process (MDP). The state, action, and reward at each time t E {O, 1, 2, . . .} are denoted St E S, at E A, and rt E R respectively. The environment's dynamics are characterized by state transition probabilities, P:SI = Pr {St+ 1 = Sf I St = s, at = a}, and expected rewards 'R~ = E {rt+l 1st = s, at = a}, 'r/s, Sf E S, a E A. The agent's decision making procedure at each time is characterized by a policy, 1l'(s, a, 0) = Pr {at = alst = s, O}, 'r/s E S,a E A, where 0 E ~, for l ? lSI, is a parameter vector. We assume that 1l' is diffentiable with respect to its parameter, i.e., that a1f~~a) exists. We also usually write just 1l'(s, a) for 1l'(s, a, 0). Policy Gradient Methods for RL with Function Approximation 1059 With function approximation, two ways of formulating the agent's objective are useful. One is the average reward formulation, in which policies are ranked according to their long-term expected reward per step, p(rr): p(1I") = n-+oon lim .!.E{rl +r2 + ... +rn 11I"} = '" ?ff(s) "'1I"(s,a)'R.:, ~ ~ II Q where cP (s) = limt-+oo Pr {St = slso, 11"} is the stationary distribution of states under 11", which we assume exists and is independent of So for all policies. In the average reward formulation, the value of a state-action pair given a policy is defined as 00 Q1r(s,a) p(1I") I So = LE {rt - = s,ao = a,1I"}, Vs E S,a E A. t=l The second formulation we cover is that in which there is a designated start state So, and we care only about the long-term reward obtained from it. We will give our results only once, but they will apply to this formulation as well under the definitions p(1I") = E{t. "(t-lrt I 8 0 ,1I"} and Q1r(s,a) = E{t. "(k-lrt+k 1St = s,at = a, 11" }. where,,( E [0,1] is a discount rate ("( = 1 is allowed only in episodic tasks). In this formulation, we define d1r (8) as a discounted weighting of states encountered starting at So and then following 11": cP(s) = E:o"(tpr{st = slso,1I"}. Our first result concerns the gradient of the performance metric with respect to the policy parameter: Theorem 1 (Policy Gradient). For any MDP, in either the average-reward or start-state formulations, ap = "'.ftr( )'" a1l"(s,a)Q1r( ao ~u II s ~ ao ) s, a . (2) Q Proof: See the appendix. This way of expressing the gradient was first rtiscussed for the average-reward formulation by Marbach and Tsitsiklis (1998), based on a related expression in terms of the state-value function due to Jaakkola, Singh, and Jordan (1995) and Coo and Chen (1997). We extend their results to the start-state formulation and provide simpler and more direct proofs. Williams's (1988, 1992) theory of REINFORCE algorithms can also be viewed as implying (2). In any event, the key aspect of both expressions for the gradient is that their are no terms of the form adiJII): the effect of policy changes on the distribution of states does not appear. This is convenient for approximating the gradient by sampling. For example, if 8 was sampled from the distribution obtained by following 11", then Ea a1r~~,a) Q1r (s, a) would be an unbiased estimate of ~. Of course, Q1r(s, a) is also not normally known and must be estimated. One approach is to use the actual returns, R t = E~l rt+k - p(1I") (or Rt = E~l "(k-lrt+k in the start-state formulation) as an approximation for each Q1r (St, at). This leads to Williams's episodic REINFORCE algorithm, t::..Ot oc a1r~~,at2 Rt 7r (1 ) (the ~a St,at 7r\St,Ut) corrects for the oversampling of actions preferred by 11"), which is known to follow ~ in expected value (Williams, 1988, 1992). 2 Policy Gradient with Approximation Now consider the case in which Q1r is approximated by a learned function approximator. If the approximation is sufficiently good, we might hope to use it in place of Q1r R. S. Sutton, D. MeAl/ester, S. Singh and Y. Mansour 1060 in (2) and still point roughly in the direction of the gradient. For example, Jaakkola, Singh, and Jordan (1995) proved that for the special case of function approximation arising in a tabular POMDP one could assure positive inner product with the gradient, which is sufficient to ensure improvement for moving in that direction. Here we extend their result to general function approximation and prove equality with the gradient. Let fw : S x A - ~ be our approximation to Q7f, with parameter w. It is natural to learn f w by following 1r and updating w by a rule such as AWt oc I,u [Q7f (St, at) fw(st,at)]2 oc [Q7f(st,at) - fw(st,at)]alw~~,ad, where Q7f(st,at) is some unbiased estimator of Q7f(st, at), perhaps Rt. When such a process has converged to a local optimum, then LcF(s):E 1r(s,a)[Q7f (s,a) - fw(s,a)] 8f~~,a) = o. (3) a /I Theorem 2 (Policy Gradient with Function Approximation). If fw satisfies (3) and is compatible with the policy parameterization in the sense that l then 8fw(s, a) 81r(s, a) 1 = 8w 80 1r(s, a) , (4) 8p ~ ~ 81r(s, a) ao = ~cF(s) ~ ao fw(s,a). (5) a II Proof: Combining (3) and (4) gives Ld7f (s) L II 87r1~a) [Q 7f (s,a) - fw(s,a)] = 0 (6) a which tells us that the error in fw(s, a) is orthogonal to the gradient of the policy parameterization. Because the expression above is zero, we can subtract it from the policy gradient theorem (2) to yield ap ao = L cF(s) II L a1r1~ a) Q7f(s, a) - :E cF(s) :E a1r1~ a) [Q 7f (s, a) 11 ~ ~ ~cF(s)~ /I = 3 a1r(s,a) ao [Q7f(s,a)-Q7f(s,a)+fw(s,a)] a ~ ~ ~ cF(s) ~ II fw(s, a)] a II a1r(s,a) ao fw(s, a). Q.E.D. a Application to Deriving Algorithms and Advantages Given a policy parameterization, Theorem 2 can be used to derive an appropriate form for the value-function parameterization. For example, consider a policy that is a Gibbs distribution in a linear combination of features: 'is E S,s E A, ITsitsiklis (personal communication) points out that /w being linear in the features given on the righthand side may be the only way to satisfy this condition. 1061 Policy Gradient Methods for RL with Function Approximation where each <Psa is an i-dimensional feature vector characterizing state-action pair s, a. Meeting the compatibility condition (4) requires that ofw(s,a) _ o1r(s,a) 1 _ 00 ( ) OW 7rS,a A 'l'sa _ L (S, b)A'l'sb, b 1r so that the natural parameterization of fw is fw(s,a) ~w T ["',. - ~"(S,b)""bl ? In other words, fw must be linear in the same features as the policy, except normalized to be mean zero for each state. Other algorithms can easily be derived for a variety of nonlinear policy parameterizations, such as multi-layer backpropagation networks. The careful reader will have noticed that the form given above for f w requires that it have zero mean for each state: l:a 1r(s, a)fw(s, a) = 0, Vs E S . In this sense it is better to think of f w as an approximation of the advantage function, A7r(s,a) = Q7r(s,a) - V7r(s) (much as in Baird, 1993), rather than of Q7r . Our convergence requirement (3) is really that fw get the relative value of the actions correct in each state, not the absolute value, nor the variation from state to state. Our results can be viewed as a justification for the special status of advantages as the target for value function approximation in RL. In fact, our (2), (3), and (5), can all be generalized to include an arbitrary function of state added to the value function or its approximation. For example, (5) can be generalized to ~ = l:s d7r (s) l:a 87r~~,a) [fw(s, a) + v(s)] ,where v : S ---+ R is an arbitrary function. (This follows immediately because l:a 87r~~a) = 0, Vs E S.) The choice of v does not affect any of our theorems, but can substantially affect the variance of the gradient estimators. The issues here are entirely analogous to those in the use of reinforcement baselines in earlier work (e.g., Williams, 1992; Dayan, 1991; Sutton, 1984). In practice, v should presumably be set to the best available approximation of V7r. Our results establish that that approximation process can proceed without affecting the expected evolution of fw and 1r . 4 Convergence of Policy Iteration with Function Approximation Given Theorem 2, we can prove for the first time that a form of policy iteration with function approximation is convergent to a locally optimal policy. Theorem 3 (Policy Iteration with Function Approximation). Let 1r and fw be any differentiable function approximators for the policy and value function respectively that satisfy the compatibility condition (4) and for which 18;~~9;) I < B < 00. Let {Ok}~o be any step-size sequence such that limk-+oo Ok = 0 and l:k Ok = 00. Then, for any MDP with bounded rewards, the maxe,s,a ,i,j sequence {p(1rk)}r=o, defined by any 00, 1rk = 1r(.,., Ok), and w such that '"' ~crk(S) '"' ~ 1rk(s,a) [Q7r k (s,a) - fw(s,a) ]ofw(s,a) ow Wk s Ok+l = '"' = ? a '"' 01rk(S, 00 a) fWk(s,a), Ok+Ok~crk(S)~ s a converges such that limk-+oo 8P~;k) = o. Proof: Our Theorem 2 assures that the Ok update is in the direction of the gradient. 8 2 7r(s a) ....?..?....The bounds on 89;89j and on the MDP's rewards together assure us that 89i89j R. S. Sutton, D. MeAl/ester. S. Singh and Y. Mansour 1062 is also bounded. These, together with the step-size requirements, are the necessary conditions to apply Proposition 3.5 from page 96 of Bertsekas and Tsitsiklis (1996), which assures convergence to a local optimum. Q.E.D. Acknowledgements The authors wish to thank Martha Steenstrup and Doina Precup for comments, and Michael Kearns for insights into the notion of optimal policy under function approximation. References Baird, L. C. (1993) . Advantage Updating. Wright Lab. Technical Report WL-TR-93-1l46. Baird, L. C. (1995). Residual algorithms: Reinforcement learning with function approximation. Proc. of the Twelfth Int. Co,:4. on Machine Learning, pp. 30-37. Morgan Kaufmann. Baird, L. C., Moore, A. W . (1999) . Gradient descent for general reinforcement learning. NIPS 11. MIT Press. Barto, A. G., Sutton, R. S., Anderson, C. W. (1983). Neuronlike elements that can solve difficult learning control problems. IEEE 1rans. on Systems, Man, and Cybernetics 19:835. Baxter, J ., Bartlett, P. (in prep.) Direct gradient-based reinforcement learning: I. Gradient estimation algorithms. Bertsekas, D. P., Tsitsiklis, J. N. (1996). Neuro-Dynamic Programming. Athena Scientific. Cao, X.-R., Chen, H.-F. (1997) . Perturbation realization, potentials, and sensitivity analysis of Markov Processes, IEEE 1hlns. on Automatic Control 42{1O):1382-1393. Dayan, P. (1991). Reinforcement comparison. In D. S. Touretzky, J. L. Elman, T. J. Sejnowski, and G. E. Hinton (eds.), Connectionist Models: Proceedings of the 1990 Summer School, pp. 45-51. Morgan Kaufmann. Gordon, G. J. (1995). Stable function approximation in dynamic programming. Proceedings of the Twelfth Int. Conf. on Machine Learning, pp. 261-268. Morgan Kaufmann. Gordon, G. J. (1996). Chattering in SARSA(A). CMU Learning Lab Technical Report. Jaakkola, T., Singh, S. P., Jordan, M. I. (1995) Reinforcement learning algorithms for partially observable Markov decision problems, NIPS 7, pp. 345-352. Morgan Kaufman. Kimura, H., Kobayashi, S. (1998). An analysis of actor/critic algorithms using eligibility traces: Reinforcement learning with imperfect value functions. Proc. ICML-98, pp. 278-286. Konda, V. R., Tsitsiklis, J. N. (in prep.) Actor-critic algorithms. Marbach, P., Tsitsiklis, J. N. (1998) Simulation-based optimization of Markov reward processes, technical report LIDS-P-2411, Massachusetts Institute of Technology. Singh, S. P., Jaakkola, T., Jordan, M. I. (1994) . Learning without state-estimation in partially observable Markovian decision problems. Proc. ICML-94, pp. 284-292. Sutton, R. S. (1984). Temporal Credit Assignment in Reinforcement Learning. Ph.D. thesis, University of Massachusetts, Amherst. Sutton, R. S., Barto, A. G. (1998) . Reinforcement Learning: An Introduction. MIT Press. Tsitsiklis, J. N. Van Roy, B. (1996) . Feature-based methods for large scale dynamic programming. Machine Learning 22:59-94. Williams, R. J . (1988) . Toward a theory of reinforcement-learning connectionist systems. Technical Report NU-CCS-88-3, Northeastern University, College of Computer Science. Williams, R. J. (1992) . Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning 8:229-256. Appendix: Proof of Theorem 1 We prove the theorem first for the average-reward formulation and then for the startstate formulation. 8V1I'(s) def 8fJ a = ~ [87r(S, 8 Q11'( s,a) ] ~ 80 a) Q11'() s,a +7r(s,a)80 a 1063 Policy Gradient Methods for RL with Function Approximation Therefore, = '""" ap ae L- [a1r(S,a)Q1T( ) ( ),"""pa aV1T(S')]_ aV1T(s) ao s, a + 1r s, a L- ss' ao ae a ~ Summing both sides over the stationary distribution d1T , = '""" L- d 1T ( s, a) Q1T (s, a) + '""" a aV1T (s') s) '""" L- a1r(ae L- U..nr ( s) '""" L- 7r ( s, a) '""" L- Pss' ae a s s _ a s' L~(s)av;O(s), s but since ~ is stationary, _ Ld1T (s) av;o(s) s a1r~~a) Q1T(s,a). :: = Ld1T (s) L Q.E.D. a s For the start-state formulation: aV 1T ( s) def a '""" 1T ae = ae L- 1r(s, a)Q (s, a) 'risE S a [a1r~~ a) Q1T(S, a) + 1r(s, a) :e Q1T (s, a)] = L a ~ ~ [inr~~ a) Q'(s,a) +--(s,a) :0 ['R~ + ~ ~P:., V'(S')]] ~ ~ [inr~~ a) Q'(s,a) +--(s,a) ~ ~P:.,! V'(S')] = '"""~ LL- 'Y It Pr ( s x -+ (7) x, k, 1r ),"""a1r(x,a)Q1T() L- ae x, a , It=o a after several steps of unrolling (7), where Pr(s -+ x, k, 1r) is the probability of going from state s to state x in k steps under policy 1r. It is then immediate that ap ae I } a {~t-l a 71' = aoE 'Y rt So,1r = ae v (so) ti = '"""~ L- L- 'Y It Pr ( So s = k=O s, k, 1r) ,"",,87r(s,a)Q7I'( L- ae s, a) a L ~(s) L s -+ a a1r~~ a) Q7I'(s, a). Q.E.D.
1713 |@word q7f:9 version:2 twelfth:2 simulation:1 r:1 valuefunction:1 tr:1 selecting:1 ours:2 si:1 written:1 must:2 update:1 v:3 stationary:3 greedy:1 selected:1 assurance:1 implying:1 parameterization:5 es:1 parameterizations:1 vaps:5 simpler:1 direct:2 prove:5 theoretically:1 expected:6 rapid:1 roughly:1 elman:1 nor:1 multi:1 discounted:1 little:1 actual:1 unrolling:1 estimating:1 bounded:2 kaufman:1 substantially:1 developed:1 finding:1 nj:1 kimura:2 temporal:2 ti:1 iearning:1 control:2 normally:1 appear:1 bertsekas:4 before:1 positive:2 kobayashi:2 local:2 sutton:11 establishing:1 approximately:1 ap:5 might:2 suggests:1 co:1 practice:1 definite:1 backpropagation:1 procedure:1 episodic:2 convenient:1 word:1 d1t:1 get:1 selection:2 put:1 deterministic:2 williams:8 go:1 attention:1 independently:1 starting:1 pomdp:1 immediately:1 rule:1 estimator:2 insight:1 deriving:1 proving:2 notion:2 variation:1 justification:1 analogous:1 updated:3 yishay:1 target:1 programming:5 pa:1 assure:2 roy:2 satisfying:1 strengthens:1 approximated:1 updating:2 element:1 highest:1 environment:1 q1t:6 reward:15 dynamic:6 personal:1 singh:10 upon:1 easily:1 represented:3 d7r:1 q7r:3 sejnowski:1 tell:1 whose:3 solve:1 s:1 florham:1 think:1 advantage:5 differentiable:4 rr:1 sequence:2 product:1 cao:1 combining:2 realization:1 degenerate:1 convergence:5 optimum:2 r1:1 requirement:2 converges:1 oo:3 derive:1 school:1 received:1 sa:1 direction:4 discontinuous:1 correct:1 stochastic:2 mcallester:1 a1l:1 require:1 ao:10 really:1 proposition:1 sarsa:2 a7r:1 sufficiently:1 credit:1 wright:1 presumably:1 estimation:3 proc:3 wl:1 v1i:1 hope:1 mit:2 rather:2 barto:4 jaakkola:6 derived:1 improvement:1 ps:1 baseline:1 sense:3 dayan:2 sb:1 going:1 selects:1 lrt:3 compatibility:2 issue:1 denoted:1 special:3 once:1 sampling:1 park:2 icml:2 tabular:2 report:4 connectionist:3 gordon:4 richard:1 oriented:1 neuronlike:1 inr:2 righthand:1 sens:1 necessary:1 experience:2 orthogonal:1 tree:1 fitted:1 instance:1 earlier:1 obstacle:1 markovian:1 cover:1 assignment:1 at2:1 st:16 sensitivity:1 amherst:1 corrects:1 michael:1 together:2 precup:1 squared:1 thesis:1 q11:2 slowly:1 ester:3 conf:1 return:1 potential:1 wk:1 includes:1 int:2 baird:6 satisfy:2 explicitly:1 doina:1 a1r:10 ad:1 lab:3 start:5 accuracy:2 variance:2 kaufmann:3 yield:1 v7r:2 pomdps:1 cc:1 cybernetics:1 converged:1 touretzky:1 ed:1 definition:1 pp:6 proof:5 sampled:1 proved:2 massachusetts:2 lim:1 ut:1 ea:1 appears:1 ok:8 follow:1 formulation:12 anderson:2 just:1 nonlinear:1 perhaps:1 scientific:1 mdp:4 effect:1 normalized:1 unbiased:4 evolution:1 equality:1 moore:2 assistance:1 psa:1 ll:1 eligibility:1 oc:3 generalized:2 cp:2 fj:1 rl:8 extend:2 tpr:1 theirs:1 expressing:1 gibbs:1 meal:2 automatic:1 similarly:1 marbach:3 moving:1 stable:1 actor:4 dominant:1 own:2 certain:1 arbitrarily:1 approximators:4 meeting:1 morgan:4 care:1 q1r:8 converge:3 ii:7 technical:4 characterized:2 long:3 ftr:1 neuro:1 ae:10 metric:1 cmu:1 crk:2 iteration:7 limt:1 achieved:1 whereas:1 affecting:1 separately:1 ot:1 unlike:1 limk:2 comment:1 climb:1 jordan:6 baxter:2 variety:2 affect:2 architecture:1 identified:1 imperfect:1 reduce:1 inner:1 avenue:1 whether:1 expression:3 bartlett:2 effort:1 ntial:1 proceed:1 cause:2 action:11 useful:1 discount:1 locally:6 ph:1 lsi:1 oversampling:1 estimated:4 arising:1 per:2 write:1 visitation:1 key:2 changing:1 parameterized:1 place:1 family:1 reader:1 decision:5 appendix:2 entirely:1 layer:1 ct:1 bound:1 summer:1 def:2 convergent:4 encountered:1 occur:1 worked:1 aspect:1 formulating:1 relatively:1 designated:1 according:2 combination:1 slightly:1 lid:1 making:1 coo:1 pr:6 awt:1 assures:2 rans:1 generalizes:1 available:1 apply:2 appropriate:1 alternative:2 ensure:1 cf:5 include:1 konda:2 establish:1 approximating:3 bl:1 objective:1 noticed:1 added:1 rt:8 interacts:1 nr:1 gradient:34 ow:2 unable:1 reinforce:6 thank:1 athena:1 toward:2 difficult:1 trace:1 ofw:2 rise:1 policy:57 av:3 markov:4 descent:1 immediate:1 hinton:1 communication:1 mansour:4 rn:1 perturbation:1 arbitrary:4 david:1 pair:2 learned:2 nu:1 nip:2 usually:2 suitable:1 event:1 ranked:1 natural:2 fwk:1 residual:2 technology:1 mdps:1 acknowledgement:1 determining:1 relative:1 limitation:1 proportional:1 proven:1 approximator:3 agent:3 sufficient:1 critic:4 course:1 compatible:1 last:1 tsitsiklis:11 side:2 weaker:1 institute:1 wide:1 characterizing:1 absolute:1 van:2 superficially:1 transition:1 author:1 reinforcement:15 far:1 approximate:3 observable:2 implicitly:1 preferred:1 status:1 satinder:1 summing:1 decade:1 learn:1 assured:1 main:1 prep:4 allowed:1 ff:1 wish:1 sf:2 weighting:1 learns:1 northeastern:1 theorem:11 rk:4 specific:1 r2:1 concern:1 essential:1 intractable:1 exists:2 chen:2 subtract:1 generalizing:1 explore:2 chattering:1 partially:2 satisfies:1 viewed:2 careful:1 man:1 change:5 aided:1 fw:21 martha:1 except:2 oon:1 kearns:1 maxe:1 college:1 mcal:1 a1f:1
780
1,714
U nmixing Hyperspectral Data Lucas Parra, Clay Spence, Paul Sajda Sarnoff Corporation, CN-5300, Princeton, NJ 08543, USA {lparra, cspence,psajda} @sarnoff.com Andreas Ziehe, Klaus-Robert Miiller GMD FIRST.lDA, Kekulestr. 7, 12489 Berlin, Germany {ziehe,klaus}@first.gmd.de Abstract In hyperspectral imagery one pixel typically consists of a mixture of the reflectance spectra of several materials, where the mixture coefficients correspond to the abundances of the constituting materials. We assume linear combinations of reflectance spectra with some additive normal sensor noise and derive a probabilistic MAP framework for analyzing hyperspectral data. As the material reflectance characteristics are not know a priori, we face the problem of unsupervised linear unmixing. The incorporation of different prior information (e.g. positivity and normalization of the abundances) naturally leads to a family of interesting algorithms, for example in the noise-free case yielding an algorithm that can be understood as constrained independent component analysis (ICA). Simulations underline the usefulness of our theory. 1 Introduction Current hyperspectral remote sensing technology can form images of ground surface reflectance at a few hundred wavelengths simultaneously, with wavelengths ranging from 0.4 to 2.5 J.Lm and spatial resolutions of 10-30 m. The applications of this technology include environmental monitoring and mineral exploration and mining. The benefit of hyperspectral imagery is that many different objects and terrain types can be characterized by their spectral signature. The first step in most hyperspectral image analysis systems is to perform a spectral unmixing to determine the original spectral signals of some set of prime materials. The basic difficulty is that for a given image pixel the spectral reflectance patterns of the surface materials is in general not known a priori. However there are general physical and statistical priors which can be exploited to potentially improve spectral unmixing. In this paper we address the problem of unmixing hyperspectral imagery through incorporation of physical and statistical priors within an unsupervised Bayesian framework. We begin by first presenting the linear superposition model for the reflectances measured. We then discuss the advantages of unsupervised over supervised systems. Unmixing Hyperspectral Data 943 We derive a general maximum a posteriori (MAP) framework to find the material spectra and infer the abundances. Interestingly, depending on how the priors are incorporated, the zero noise case yields (i) a simplex approach or (ii) a constrained leA algorithm. Assuming non-zero noise our MAP estimate utilizes a constrained least squares algorithm. The two latter approaches are new algorithms whereas the simplex algorithm has been previously suggested for the analysis of hyperspectral data. Linear Modeling To a first approximation the intensities X (Xi>.) measured in each spectral band A = 1, ... , L for a given pixel i = 1, ... , N are linear combinations of the reflectance characteristics S (8 m >.) of the materials m = 1, ... , M present in that area. Possible errors of this approximation and sensor noise are taken into account by adding a noise term N (ni>'). In matrix form this can be summarized as X = AS + N, subject to: AIM = lL, A ~ 0, (1) where matrix A (aim) represents the abundance of material m in the area corresponding to pixel i, with positivity and normalization constraints. Note that ground inclination or a changing viewing angle may cause an overall scale factor for all bands that varies with the pixels. This can be incorporated in the model by simply replacing the constraint AIM = lL with AIM ~ lL which does does not affect the discussion in the remainder of the paper. This is clearly a simplified model of the physical phenomena. For example, with spatially fine grained mixtures, called intimate mixtures, multiple reflectance may causes departures from this first order model. Additionally there are a number of inherent spatial variations in real data, such as inhomogeneous vapor and dust particles in the atmosphere, that will cause a departure from the linear model in equation (1). Nevertheless, in practical applications a linear model has produced reasonable results for areal mixtures. Supervised vs. Unsupervised techniques Supervised spectral un mixing relies on the prior knowledge about the reflectance patterns S of candidate surface materials, sometimes called endmembers, or expert knowledge and a series of semiautomatic steps to find the constituting materials in a particular scene. Once the user identifies a pixel i containing a single material, i.e. aim = 1 for a given m and i, the corresponding spectral characteristics of that material can be taken directly from the observations, i.e., 8 m >. = Xi>. [4]. Given knowledge about the endmembers one can simply find the abundances by solving a constrained least squares problem. The problem with such supervised techniques is that finding the correct S may require substantial user interaction and the result may be error prone, as a pixel that actually contains a mixture can be misinterpreted as a pure endmember. Another approach obtains endmembers directly from a database. This is also problematic because the actual surface material on the ground may not match the database entries, due to atmospheric absorption or other noise sources. Finding close matches is an ambiguous process as some endmembers have very similar reflectance characteristics and may match several entries in the database. Unsupervised unmixing, in contrast, tries to identify the endmembers and mixtures directly from the observed data X without any user interaction. There are a variety of such approaches. In one approach a simplex is fit to the data distribution [7, 6, 2]. The resulting vertex points of the simplex represent the desired endmembers, but this technique is very sensitive to noise as a few boundary points can potentially change the location of the simplex vertex points considerably. Another approach by Szu [9] tries to find abundances that have the highest entropy subject to constraints that the amount of materials is as evenly distributed as possible - an assumption L. Parra, C. D. Spence, P Sajda, A. Ziehe and K.-R. Muller 944 which is clearly not valid in many actual surface material distributions. A relatively new approach considers modeling the statistical information across wavelength as statistically independent AR processes [1]. This leads directly to the contextual linear leA algorithm [5]. However, the approach in [1] does not take into account constraints on the abundances, noise, or prior information. Most importantly, the method [1] can only integrate information from a small number of pixels at a time (same as the number of endmembers). Typically however we will have only a few endmembers but many thousand pixels. 2 2.1 The Maximum A Posterior Framework A probabilistic model of unsupervised spectral unmixing Our model has observations or data X and hidden variables A, S, and N that are explained by the noisy linear model (1). We estimate the values of the hidden variables by using MAP (A SIX) p , = p(XIA, S)p(A, S) = Pn(XIA, S)Pa(A)ps(S) p(X) p(X) (2) with Pa(A), Ps(S), Pn(N) as the a priori assumptions of the distributions. With MAP we estimate the most probable values for given priors after observing the data, A MAP , SMAP = argmaxp(A, SIX) (3) A,S Note that for maximization the constant factor p(X) can be ignored. Our first assumption, which is indicated in equation (2) is that the abundances are independent of the reflectance spectra as their origins are completely unrelated: (AO) A and S are independent. The MAP algorithm is entirely defined by the choices of priors that are guided by the problem of hyperspectral unmixing: (AI) A represent probabilities for each pixel i. (A2) S are independent for different material m. (A3) N are normal i.i.d. for all i, A. In summary, our MAP framework includes the assumptions AO-A3. 2.2 Including Priors Priors on the abundances be represented as, Positivity and normalization of the abundances can (4) eo where 60 represent the Kronecker delta function and the step function. With this choice a point not satisfying the constraint will have zero a posteriori probability. This prior introduces no particular bias of the solutions other then abundance constraints. It does however assume the abundances of different pixels to be independent. Prior on spectra Usually we find systematic trends in the spectra that cause significant correlation. However such an overall trend can be subtracted and/or filtered from the data leaving only independent signals that encode the variation from that overall trend. For example one can capture the conditional dependency structure with a linear auto-regressive (AR) model and analyze the resulting "innovations" or prediction errors [3]. In our model we assume that the spectra represent independent instances of an AR process having a white innovation process em.>. distributed according to Pe(e). With a Toeplitz matrix T of the AR coefficients we 945 Unmixing Hyperspectral Data can write, em = Sm T. The AR coefficients can be found in a preprocessing step on the observations X. If S now represents the innovation process itself, our prior can be represented as, M Pe (S) <X Pe(ST) = L L II II Pe( L sm>.d>.>.,) , (5) m=1 >.=1 >.'=1 Additionally Pe (e) is parameterized by a mean and scale parameter and potentially parameters determining the higher moments of the distributions. For brevity we ignore the details of the parameterization in this paper. Prior on the noise As outlined in the introduction there are a number of problems that can cause the linear model X = AS to be inaccurate (e.g. multiple reflections, inhomogeneous atmospheric absorption, and detector noise.) As it is hard to treat all these phenomena explicitly, we suggest to pool them into one noise variable that we assume for simplicity to be normal distributed with a wavelength dependent noise variance a>., L p(XIA, S) = Pn(N) = N(X - AS,~) = II N(x>. - As>., a>.l) , (6) >.=1 where N (', .) represents a zero mean Gaussian distribution, and 1 the identity matrix indicating the independent noise at each pixel. 2.3 MAP Solution for Zero Noise Case Let us consider the noise-free case. Although this simplification may be inaccurate it will allow us to greatly reduce the number of free hidden variables - from N M + M L to M2 . In the noise-free case the variables A, S are then deterministically dependent on each other through a N L-dimensional 8-distribution, Pn(XIAS) = 8(X - AS). We can remove one of these variables from our discussion by integrating (2). It is instructive to first consider removing A p(SIX) <X I dA 8(X - AS)Pa(A)ps(S) = IS- 1IPa(XS- 1 )Ps(S). (7) We omit tedious details and assume L = M and invertible S so that we can perform the variable substitution that introduces the Jacobian determinant IS-II . Let us consider the influence of the different terms. The Jacobian determinant measures the volume spanned by the endmembers S. Maximizing its inverse will therefore try to shrink the simplex spanned by S. The term Pa(XS- 1 ) should guarantee that all data points map into the inside of the simplex, since the term should contribute zero or low probability for points that violate the constraint. Note that these two terms, in principle, define the same objective as the simplex envelope fitting algorithms previously mentioned [2]. In the present work we are more interested in the algorithm that results from removing S and finding the MAP estimate of A. We obtain (d. Eq.(7)) p(AIX) oc I dS 8(X - AS)Pa(A)ps(S) = IA -llps(A- 1 X)Pa(A). (8) For now we assumed N = M. 1 If Ps (S) factors over m , i.e. endmembers are independent, maximizing the first two terms represents the leA algorithm. However, lIn practice more frequently we have N > M. In that case the observations X can be mapped into a M dimensional subspace using the singular value decomposition (SVD) , X = UDV T , The discussion applies then to the reduced observations X = u1x with U M being the first M columns of U . L. Parra. C. D. Spence. P Sajda. A. Ziehe and K.-R. Muller 946 the prior on A will restrict the solutions to satisfy the abundance constraints and bias the result depending on the detailed choice of Pa(A), so we are led to constrained ICA. In summary, depending on which variable we integrate out we obtain two methods for solving the spectral unmixing problem: the known technique of simplex fitting and a new constrained ICA algorithm. 2.4 MAP Solution for the Noisy Case Combining the choices for the priors made in section 2.2 (Eqs.(4), (5) and (6)) with (2) and (3) we obtain AMAP, SMAP = "''i~ax ft {g N(x", - a,s" a,) ll. P,(t. 'm,d",) } , (9) subject to AIM = lL, A 2: O. The logarithm of the cost function in (9) is denoted by L = L(A, S). Its gradient with respect to the hidden variables is = _AT nm diag(O')-l - fs(sm) 88L Sm (10) where N = X - AS, nm are the M column vectors of N, fs(s) = - olnc;(s). In (10) fs is applied to each element of Sm. The optimization with respect to A for given S can be implemented as a standard weighted least squares (L8) problem with a linear constraint and positivity bounds. Since the constraints apply for every pixel independently one can solve N separate constrained LS problems of M unknowns each. We alternate between gradient steps for S and explicit solutions for A until convergence. Any additional parameters of Pe(e) such as scale and mean may be obtained in a maximum likelihood (ML) sense by maximizing L. Note that the nonlinear optimization is not subject to constraints; the constraints apply only in the quadratic optimization. 3 Experiments 3.1 Zero Noise Case: Artificial Mixtures In our first experiment we use mineral data from the United States Geological Survey (USGS)2 to build artificial mixtures for evaluating our unsupervised unmixing framework. Three target endmembers where chosen (Almandine WS479 , Montmorillonite+Illi CM42 and Dickite NMNH106242). A spectral scene of 100 samples was constructed by creating a random mixture of the three minerals. Of the 100 samples, there were no pure samples (Le. no mineral had more than a 80% abundance in any sample). Figure 1A is the spectra of the endmembers recovered by the constrained ICA technique of section 2.3, where the constraints were implemented with penalty terms added to the conventional maximum likelihood ICA algorithm. These are nearly identical to the spectra of the true endmembers, shown in figure 1B, which were used for mixing. Interesting to note is the scatter-plot of the 100 samples across two bands. The open circles are the absorption values at these two bands for endmembers found by the MAP technique. Given that each mixed sample consists of no more than 80% of any endmember, the endmember points on the scatter-plot are quite distant from the cluster. A simplex fitting technique would have significant difficulty recovering the endmembers from this clustering. 2see http://speclab.cr .usgs.gov /spectral.lib.456.descript/ decript04.html 947 Unmixing Hyperspectral Data found endmembers observed X and found S target endmembers o g 0.8 ~ ~0.6 ., ~ ~ 0.4 o O~------' 50 100 150 wavelength O~------' 200 A 50 100 150 wavelength B 200 0.2'---~------' 0.4 0.6 0.8 wavelength=30 C Figure 1: Results for noise-free artificial mixture. A recovered endmembers using MAP technique. B "true" target endmembers. C scatter plot of samples across 2 bands showing the absorption of the three endmembers computed by MAP (open circles). 3.2 Noisy Case: Real Mixtures To validate the noise model MAP framework of section 2.4 we conducted an experiment using ground truthed USGS data representing real mixtures. We selected lOxl0 blocks of pixels from three different regions 3 in the AVIRIS data of the Cuprite, Nevada mining district. We separate these 300 mixed spectra assuming two endmembers and an AR detrending with 5 AR coefficients and the MAP techniques of section 2.4. Overall brightness was accounted for as explain in the linear modeling of section 1. The endmembers are shown in figure 2A and B in comparison to laboratory spectra from the USGS spectral library for these minerals [8J . Figure 2C shows the corresponding abundances, which match the ground truth; region (III) mainly consists of Muscovite while regions (1)+(I1) contain (areal) mixtures of Kaolinite and Muscovite. 4 Discussion Hyperspectral unmixing is a challenging practical problem for unsupervised learning. Our probabilistic approach leads to several interesting algorithms: (1) simplex fitting, (2) constrained ICA and (3) constrained least squares that can efficiently use multi-channel information. An important element of our approach is the explicit use of prior information. Our simulation examples show that we can recover the endmembers, even in the presence of noise and model uncertainty. The approach described in this paper does not yet exploit local correlations between neighboring pixels that are well known to exist. Future work will therefore exploit not only spectral but also spatial prior information for detecting objects and materials. Acknowledgments We would like to thank Gregg Swayze at the USGS for assistance in obtaining the data. 3The regions were from the image plate2.cuprite95.alpha.2um.image.wlocals.gif in ftp:/ /speclab.cr.usgs.gov /pub/cuprite/gregg.thesis.images/, at the coordinates (265,710) and (275,697), which contained Kaolinite and Muscovite 2, and (143,661), which only contained Muscovite 2. L. Parra, C. D, Spence, P Sajda, A. Ziehe and K-R. Muller 948 Muscovite Kaolinite 0.8 0.7 0 .65 0.6 0.6 0.55 0.4 0.5 'c .?.? ", "'0 .. ' ., 0.45 0.3 0.4,--~--:-:-:-"-~----:--:--~ 160 190 200 waveleng1h A 210 220 180 190 200 wavelength B 210 220 C Figure 2: A Spectra of computed endmember (solid line) vs Muscovite sample spectra from the USGS data base library. Note we show only part of the spectrum since the discriminating features are located only between band 172 and 220. B Computed endmember (solid line) vs Kaolinite sample spectra from the USGS data base library. C Abundances for Kaolinite and Muscovite for three regions (lighter pixels represent higher abundance). Region 1 and region 2 have similar abundances for Kaolinite and Muscovite, while region 3 contains more Muscovite. References [1] J. Bayliss, J. A. Gualtieri, and R. Cromp. Analyzing hyperspectral data with independent component analysis. In J. M. Selander, editor, Proc. SPIE Applied Image and Pattern Recognition Workshop, volume 9, P.O. Box 10, Bellingham WA 98227-0010, 1997. SPIE. [2] J.W. Boardman and F.A. Kruse. Automated spectral analysis: a geologic example using AVIRIS data, north Grapevine Mountains, Nevada. In Tenth Thematic Conference on Geologic Remote Sensing, pages 407-418, Ann arbor, MI, 1994. Environmental Research Institute of Michigan. [3] S. Haykin. Adaptive Filter Theory. Prentice Hall, 1991. [4] F. Maselli, , M. Pieri, and C. Conese. Automatic identification of end-members for the spectral decomposition of remotely sensed scenes. Remote Sensing for Geography, Geology, Land Planning, and Cultural Heritage (SPIE) , 2960:104109,1996. [5] B. Pearlmutter and L. Parra. Maximum likelihood blind source separation: A context-sensitive generalization ofICA. In M. Mozer, M. Jordan, and T. Petsche, editors, Advances in Neural Information Processing Systems 9, pages 613-619, Cambridge MA, 1997. MIT Press. [6] J.J. Settle. Linear mixing and the estimation of ground cover proportions. International Journal of Remote Sensing, 14:1159-1177,1993. [7] M.O. Smith, J .B. Adams, and A.R. Gillespie. Reference endmembers for spectral mixture analysis. In Fifth Australian remote sensing conference, volume 1, pages 331-340, 1990. [8] U.S. Geological Survey. USGS digital spectral library. Open File Report 93-592, 1993. [9] H. Szu and C. Hsu. Landsat spectral demixing a la superresolution of blind matrix inversion by constraint MaxEnt neural nets. In Wavelet Applications IV, volume 3078, pages 147-160. SPIE, 1997.
1714 |@word determinant:2 inversion:1 proportion:1 underline:1 tedious:1 open:3 simulation:2 sensed:1 decomposition:2 brightness:1 solid:2 moment:1 substitution:1 series:1 contains:2 united:1 pub:1 interestingly:1 current:1 com:1 contextual:1 recovered:2 scatter:3 yet:1 additive:1 distant:1 remove:1 plot:3 v:3 selected:1 geologic:2 parameterization:1 smith:1 haykin:1 filtered:1 regressive:1 detecting:1 contribute:1 location:1 misinterpreted:1 district:1 constructed:1 consists:3 fitting:4 inside:1 ica:6 udv:1 frequently:1 planning:1 multi:1 gov:2 actual:2 boardman:1 lib:1 ofica:1 begin:1 unrelated:1 cultural:1 superresolution:1 vapor:1 mountain:1 gif:1 finding:3 corporation:1 nj:1 guarantee:1 every:1 um:1 omit:1 understood:1 local:1 treat:1 analyzing:2 aviris:2 challenging:1 sarnoff:2 statistically:1 practical:2 acknowledgment:1 spence:4 practice:1 block:1 bellingham:1 area:2 remotely:1 integrating:1 suggest:1 close:1 prentice:1 context:1 influence:1 conventional:1 map:17 maximizing:3 independently:1 l:1 survey:2 resolution:1 simplicity:1 pure:2 m2:1 importantly:1 spanned:2 variation:2 coordinate:1 target:3 user:3 lighter:1 origin:1 pa:7 szu:2 trend:3 satisfying:1 element:2 located:1 recognition:1 cspence:1 database:3 observed:2 ft:1 capture:1 thousand:1 region:8 remote:5 highest:1 loxl0:1 substantial:1 mentioned:1 mozer:1 signature:1 solving:2 completely:1 represented:2 sajda:4 artificial:3 klaus:2 quite:1 solve:1 toeplitz:1 noisy:3 itself:1 advantage:1 net:1 nevada:2 interaction:2 remainder:1 neighboring:1 combining:1 gregg:2 mixing:3 validate:1 convergence:1 cluster:1 p:6 unmixing:13 argmaxp:1 adam:1 object:2 ftp:1 derive:2 depending:3 geology:1 measured:2 eq:2 implemented:2 recovering:1 australian:1 guided:1 inhomogeneous:2 correct:1 filter:1 exploration:1 viewing:1 settle:1 material:17 require:1 atmosphere:1 ao:2 generalization:1 geography:1 probable:1 absorption:4 parra:5 hall:1 ground:6 normal:3 lm:1 a2:1 lparra:1 proc:1 estimation:1 superposition:1 sensitive:2 weighted:1 mit:1 clearly:2 sensor:2 gaussian:1 aim:6 pn:4 cr:2 encode:1 ax:1 likelihood:3 mainly:1 greatly:1 contrast:1 sense:1 posteriori:2 dependent:2 landsat:1 inaccurate:2 typically:2 hidden:4 interested:1 germany:1 i1:1 pixel:16 overall:4 html:1 denoted:1 priori:3 lucas:1 constrained:10 spatial:3 once:1 having:1 identical:1 represents:4 unsupervised:8 nearly:1 future:1 simplex:11 report:1 inherent:1 few:3 simultaneously:1 descript:1 mining:2 introduces:2 mixture:15 yielding:1 iv:1 logarithm:1 maxent:1 desired:1 circle:2 instance:1 column:2 modeling:3 ar:7 cover:1 maximization:1 cuprite:2 cost:1 vertex:2 entry:2 hundred:1 usefulness:1 conducted:1 dependency:1 varies:1 considerably:1 st:1 international:1 discriminating:1 probabilistic:3 systematic:1 pool:1 invertible:1 aix:1 imagery:3 thesis:1 nm:2 containing:1 smap:2 positivity:4 creating:1 expert:1 account:2 de:1 summarized:1 includes:1 coefficient:4 north:1 satisfy:1 explicitly:1 blind:2 try:3 observing:1 analyze:1 recover:1 psajda:1 square:4 kekulestr:1 ni:1 variance:1 characteristic:4 efficiently:1 correspond:1 yield:1 identify:1 bayesian:1 identification:1 produced:1 monitoring:1 detector:1 explain:1 naturally:1 spie:4 mi:1 hsu:1 knowledge:3 clay:1 actually:1 higher:2 supervised:4 shrink:1 box:1 correlation:2 d:1 until:1 replacing:1 nonlinear:1 lda:1 indicated:1 usa:1 contain:1 true:2 spatially:1 laboratory:1 white:1 ll:5 assistance:1 ambiguous:1 oc:1 presenting:1 pearlmutter:1 reflection:1 endmember:5 image:7 ranging:1 physical:3 volume:4 significant:2 areal:2 cambridge:1 ai:1 automatic:1 outlined:1 particle:1 had:1 surface:5 base:2 gualtieri:1 dust:1 posterior:1 prime:1 exploited:1 muller:3 additional:1 eo:1 determine:1 kruse:1 signal:2 ii:5 multiple:2 violate:1 infer:1 match:4 characterized:1 lin:1 prediction:1 basic:1 normalization:3 sometimes:1 represent:5 lea:3 whereas:1 fine:1 singular:1 source:2 leaving:1 envelope:1 file:1 subject:4 member:1 jordan:1 presence:1 iii:1 automated:1 variety:1 affect:1 fit:1 restrict:1 andreas:1 reduce:1 cn:1 six:3 penalty:1 f:3 miiller:1 cause:5 ignored:1 detailed:1 amount:1 band:6 gmd:2 reduced:1 http:1 exist:1 problematic:1 ipa:1 delta:1 write:1 nevertheless:1 changing:1 tenth:1 usgs:9 angle:1 parameterized:1 inverse:1 uncertainty:1 family:1 reasonable:1 utilizes:1 separation:1 entirely:1 bound:1 simplification:1 quadratic:1 incorporation:2 constraint:14 kronecker:1 scene:3 endmembers:24 relatively:1 according:1 alternate:1 combination:2 across:3 em:2 explained:1 taken:2 equation:2 previously:2 discus:1 amap:1 know:1 end:1 apply:2 spectral:19 petsche:1 subtracted:1 original:1 clustering:1 include:1 heritage:1 exploit:2 reflectance:11 build:1 objective:1 added:1 speclab:2 gradient:2 subspace:1 separate:2 mapped:1 berlin:1 geological:2 mineral:5 thank:1 evenly:1 considers:1 assuming:2 innovation:3 robert:1 potentially:3 unknown:1 perform:2 observation:5 sm:5 incorporated:2 intensity:1 atmospheric:2 inclination:1 address:1 suggested:1 usually:1 pattern:3 departure:2 including:1 ia:1 gillespie:1 difficulty:2 representing:1 improve:1 technology:2 library:4 identifies:1 auto:1 prior:18 determining:1 mixed:2 interesting:3 digital:1 integrate:2 principle:1 editor:2 land:1 prone:1 l8:1 summary:2 accounted:1 free:5 bias:2 allow:1 institute:1 face:1 fifth:1 kaolinite:6 benefit:1 distributed:3 boundary:1 xia:3 valid:1 evaluating:1 made:1 adaptive:1 preprocessing:1 simplified:1 constituting:2 alpha:1 obtains:1 ignore:1 ml:1 assumed:1 xi:2 spectrum:15 terrain:1 un:1 additionally:2 channel:1 obtaining:1 da:1 diag:1 noise:21 paul:1 bayliss:1 illi:1 thematic:1 deterministically:1 explicit:2 candidate:1 intimate:1 pe:6 muscovite:9 jacobian:2 wavelet:1 abundance:18 grained:1 removing:2 showing:1 sensing:5 x:2 a3:2 demixing:1 workshop:1 adding:1 hyperspectral:14 entropy:1 led:1 michigan:1 wavelength:8 simply:2 contained:2 applies:1 truthed:1 truth:1 environmental:2 relies:1 ma:1 conditional:1 identity:1 ann:1 change:1 hard:1 called:2 svd:1 arbor:1 la:1 indicating:1 ziehe:5 latter:1 brevity:1 phenomenon:2 princeton:1 instructive:1
781
1,715
U nmixing Hyperspectral Data Lucas Parra, Clay Spence, Paul Sajda Sarnoff Corporation, CN-5300, Princeton, NJ 08543, USA {lparra, cspence,psajda} @sarnoff.com Andreas Ziehe, Klaus-Robert Miiller GMD FIRST.lDA, Kekulestr. 7, 12489 Berlin, Germany {ziehe,klaus}@first.gmd.de Abstract In hyperspectral imagery one pixel typically consists of a mixture of the reflectance spectra of several materials, where the mixture coefficients correspond to the abundances of the constituting materials. We assume linear combinations of reflectance spectra with some additive normal sensor noise and derive a probabilistic MAP framework for analyzing hyperspectral data. As the material reflectance characteristics are not know a priori, we face the problem of unsupervised linear unmixing. The incorporation of different prior information (e.g. positivity and normalization of the abundances) naturally leads to a family of interesting algorithms, for example in the noise-free case yielding an algorithm that can be understood as constrained independent component analysis (ICA). Simulations underline the usefulness of our theory. 1 Introduction Current hyperspectral remote sensing technology can form images of ground surface reflectance at a few hundred wavelengths simultaneously, with wavelengths ranging from 0.4 to 2.5 J.Lm and spatial resolutions of 10-30 m. The applications of this technology include environmental monitoring and mineral exploration and mining. The benefit of hyperspectral imagery is that many different objects and terrain types can be characterized by their spectral signature. The first step in most hyperspectral image analysis systems is to perform a spectral unmixing to determine the original spectral signals of some set of prime materials. The basic difficulty is that for a given image pixel the spectral reflectance patterns of the surface materials is in general not known a priori. However there are general physical and statistical priors which can be exploited to potentially improve spectral unmixing. In this paper we address the problem of unmixing hyperspectral imagery through incorporation of physical and statistical priors within an unsupervised Bayesian framework. We begin by first presenting the linear superposition model for the reflectances measured. We then discuss the advantages of unsupervised over supervised systems. Unmixing Hyperspectral Data 943 We derive a general maximum a posteriori (MAP) framework to find the material spectra and infer the abundances. Interestingly, depending on how the priors are incorporated, the zero noise case yields (i) a simplex approach or (ii) a constrained leA algorithm. Assuming non-zero noise our MAP estimate utilizes a constrained least squares algorithm. The two latter approaches are new algorithms whereas the simplex algorithm has been previously suggested for the analysis of hyperspectral data. Linear Modeling To a first approximation the intensities X (Xi>.) measured in each spectral band A = 1, ... , L for a given pixel i = 1, ... , N are linear combinations of the reflectance characteristics S (8 m >.) of the materials m = 1, ... , M present in that area. Possible errors of this approximation and sensor noise are taken into account by adding a noise term N (ni>'). In matrix form this can be summarized as X = AS + N, subject to: AIM = lL, A ~ 0, (1) where matrix A (aim) represents the abundance of material m in the area corresponding to pixel i, with positivity and normalization constraints. Note that ground inclination or a changing viewing angle may cause an overall scale factor for all bands that varies with the pixels. This can be incorporated in the model by simply replacing the constraint AIM = lL with AIM ~ lL which does does not affect the discussion in the remainder of the paper. This is clearly a simplified model of the physical phenomena. For example, with spatially fine grained mixtures, called intimate mixtures, multiple reflectance may causes departures from this first order model. Additionally there are a number of inherent spatial variations in real data, such as inhomogeneous vapor and dust particles in the atmosphere, that will cause a departure from the linear model in equation (1). Nevertheless, in practical applications a linear model has produced reasonable results for areal mixtures. Supervised vs. Unsupervised techniques Supervised spectral un mixing relies on the prior knowledge about the reflectance patterns S of candidate surface materials, sometimes called endmembers, or expert knowledge and a series of semiautomatic steps to find the constituting materials in a particular scene. Once the user identifies a pixel i containing a single material, i.e. aim = 1 for a given m and i, the corresponding spectral characteristics of that material can be taken directly from the observations, i.e., 8 m >. = Xi>. [4]. Given knowledge about the endmembers one can simply find the abundances by solving a constrained least squares problem. The problem with such supervised techniques is that finding the correct S may require substantial user interaction and the result may be error prone, as a pixel that actually contains a mixture can be misinterpreted as a pure endmember. Another approach obtains endmembers directly from a database. This is also problematic because the actual surface material on the ground may not match the database entries, due to atmospheric absorption or other noise sources. Finding close matches is an ambiguous process as some endmembers have very similar reflectance characteristics and may match several entries in the database. Unsupervised unmixing, in contrast, tries to identify the endmembers and mixtures directly from the observed data X without any user interaction. There are a variety of such approaches. In one approach a simplex is fit to the data distribution [7, 6, 2]. The resulting vertex points of the simplex represent the desired endmembers, but this technique is very sensitive to noise as a few boundary points can potentially change the location of the simplex vertex points considerably. Another approach by Szu [9] tries to find abundances that have the highest entropy subject to constraints that the amount of materials is as evenly distributed as possible - an assumption L. Parra, C. D. Spence, P Sajda, A. Ziehe and K.-R. Muller 944 which is clearly not valid in many actual surface material distributions. A relatively new approach considers modeling the statistical information across wavelength as statistically independent AR processes [1]. This leads directly to the contextual linear leA algorithm [5]. However, the approach in [1] does not take into account constraints on the abundances, noise, or prior information. Most importantly, the method [1] can only integrate information from a small number of pixels at a time (same as the number of endmembers). Typically however we will have only a few endmembers but many thousand pixels. 2 2.1 The Maximum A Posterior Framework A probabilistic model of unsupervised spectral unmixing Our model has observations or data X and hidden variables A, S, and N that are explained by the noisy linear model (1). We estimate the values of the hidden variables by using MAP (A SIX) p , = p(XIA, S)p(A, S) = Pn(XIA, S)Pa(A)ps(S) p(X) p(X) (2) with Pa(A), Ps(S), Pn(N) as the a priori assumptions of the distributions. With MAP we estimate the most probable values for given priors after observing the data, A MAP , SMAP = argmaxp(A, SIX) (3) A,S Note that for maximization the constant factor p(X) can be ignored. Our first assumption, which is indicated in equation (2) is that the abundances are independent of the reflectance spectra as their origins are completely unrelated: (AO) A and S are independent. The MAP algorithm is entirely defined by the choices of priors that are guided by the problem of hyperspectral unmixing: (AI) A represent probabilities for each pixel i. (A2) S are independent for different material m. (A3) N are normal i.i.d. for all i, A. In summary, our MAP framework includes the assumptions AO-A3. 2.2 Including Priors Priors on the abundances be represented as, Positivity and normalization of the abundances can (4) eo where 60 represent the Kronecker delta function and the step function. With this choice a point not satisfying the constraint will have zero a posteriori probability. This prior introduces no particular bias of the solutions other then abundance constraints. It does however assume the abundances of different pixels to be independent. Prior on spectra Usually we find systematic trends in the spectra that cause significant correlation. However such an overall trend can be subtracted and/or filtered from the data leaving only independent signals that encode the variation from that overall trend. For example one can capture the conditional dependency structure with a linear auto-regressive (AR) model and analyze the resulting "innovations" or prediction errors [3]. In our model we assume that the spectra represent independent instances of an AR process having a white innovation process em.>. distributed according to Pe(e). With a Toeplitz matrix T of the AR coefficients we 945 Unmixing Hyperspectral Data can write, em = Sm T. The AR coefficients can be found in a preprocessing step on the observations X. If S now represents the innovation process itself, our prior can be represented as, M Pe (S) <X Pe(ST) = L L II II Pe( L sm>.d>.>.,) , (5) m=1 >.=1 >.'=1 Additionally Pe (e) is parameterized by a mean and scale parameter and potentially parameters determining the higher moments of the distributions. For brevity we ignore the details of the parameterization in this paper. Prior on the noise As outlined in the introduction there are a number of problems that can cause the linear model X = AS to be inaccurate (e.g. multiple reflections, inhomogeneous atmospheric absorption, and detector noise.) As it is hard to treat all these phenomena explicitly, we suggest to pool them into one noise variable that we assume for simplicity to be normal distributed with a wavelength dependent noise variance a>., L p(XIA, S) = Pn(N) = N(X - AS,~) = II N(x>. - As>., a>.l) , (6) >.=1 where N (', .) represents a zero mean Gaussian distribution, and 1 the identity matrix indicating the independent noise at each pixel. 2.3 MAP Solution for Zero Noise Case Let us consider the noise-free case. Although this simplification may be inaccurate it will allow us to greatly reduce the number of free hidden variables - from N M + M L to M2 . In the noise-free case the variables A, S are then deterministically dependent on each other through a N L-dimensional 8-distribution, Pn(XIAS) = 8(X - AS). We can remove one of these variables from our discussion by integrating (2). It is instructive to first consider removing A p(SIX) <X I dA 8(X - AS)Pa(A)ps(S) = IS- 1IPa(XS- 1 )Ps(S). (7) We omit tedious details and assume L = M and invertible S so that we can perform the variable substitution that introduces the Jacobian determinant IS-II . Let us consider the influence of the different terms. The Jacobian determinant measures the volume spanned by the endmembers S. Maximizing its inverse will therefore try to shrink the simplex spanned by S. The term Pa(XS- 1 ) should guarantee that all data points map into the inside of the simplex, since the term should contribute zero or low probability for points that violate the constraint. Note that these two terms, in principle, define the same objective as the simplex envelope fitting algorithms previously mentioned [2]. In the present work we are more interested in the algorithm that results from removing S and finding the MAP estimate of A. We obtain (d. Eq.(7)) p(AIX) oc I dS 8(X - AS)Pa(A)ps(S) = IA -llps(A- 1 X)Pa(A). (8) For now we assumed N = M. 1 If Ps (S) factors over m , i.e. endmembers are independent, maximizing the first two terms represents the leA algorithm. However, lIn practice more frequently we have N > M. In that case the observations X can be mapped into a M dimensional subspace using the singular value decomposition (SVD) , X = UDV T , The discussion applies then to the reduced observations X = u1x with U M being the first M columns of U . L. Parra. C. D. Spence. P Sajda. A. Ziehe and K.-R. Muller 946 the prior on A will restrict the solutions to satisfy the abundance constraints and bias the result depending on the detailed choice of Pa(A), so we are led to constrained ICA. In summary, depending on which variable we integrate out we obtain two methods for solving the spectral unmixing problem: the known technique of simplex fitting and a new constrained ICA algorithm. 2.4 MAP Solution for the Noisy Case Combining the choices for the priors made in section 2.2 (Eqs.(4), (5) and (6)) with (2) and (3) we obtain AMAP, SMAP = "''i~ax ft {g N(x", - a,s" a,) ll. P,(t. 'm,d",) } , (9) subject to AIM = lL, A 2: O. The logarithm of the cost function in (9) is denoted by L = L(A, S). Its gradient with respect to the hidden variables is = _AT nm diag(O')-l - fs(sm) 88L Sm (10) where N = X - AS, nm are the M column vectors of N, fs(s) = - olnc;(s). In (10) fs is applied to each element of Sm. The optimization with respect to A for given S can be implemented as a standard weighted least squares (L8) problem with a linear constraint and positivity bounds. Since the constraints apply for every pixel independently one can solve N separate constrained LS problems of M unknowns each. We alternate between gradient steps for S and explicit solutions for A until convergence. Any additional parameters of Pe(e) such as scale and mean may be obtained in a maximum likelihood (ML) sense by maximizing L. Note that the nonlinear optimization is not subject to constraints; the constraints apply only in the quadratic optimization. 3 Experiments 3.1 Zero Noise Case: Artificial Mixtures In our first experiment we use mineral data from the United States Geological Survey (USGS)2 to build artificial mixtures for evaluating our unsupervised unmixing framework. Three target endmembers where chosen (Almandine WS479 , Montmorillonite+Illi CM42 and Dickite NMNH106242). A spectral scene of 100 samples was constructed by creating a random mixture of the three minerals. Of the 100 samples, there were no pure samples (Le. no mineral had more than a 80% abundance in any sample). Figure 1A is the spectra of the endmembers recovered by the constrained ICA technique of section 2.3, where the constraints were implemented with penalty terms added to the conventional maximum likelihood ICA algorithm. These are nearly identical to the spectra of the true endmembers, shown in figure 1B, which were used for mixing. Interesting to note is the scatter-plot of the 100 samples across two bands. The open circles are the absorption values at these two bands for endmembers found by the MAP technique. Given that each mixed sample consists of no more than 80% of any endmember, the endmember points on the scatter-plot are quite distant from the cluster. A simplex fitting technique would have significant difficulty recovering the endmembers from this clustering. 2see http://speclab.cr .usgs.gov /spectral.lib.456.descript/ decript04.html 947 Unmixing Hyperspectral Data found endmembers observed X and found S target endmembers o g 0.8 ~ ~0.6 ., ~ ~ 0.4 o O~------' 50 100 150 wavelength O~------' 200 A 50 100 150 wavelength B 200 0.2'---~------' 0.4 0.6 0.8 wavelength=30 C Figure 1: Results for noise-free artificial mixture. A recovered endmembers using MAP technique. B "true" target endmembers. C scatter plot of samples across 2 bands showing the absorption of the three endmembers computed by MAP (open circles). 3.2 Noisy Case: Real Mixtures To validate the noise model MAP framework of section 2.4 we conducted an experiment using ground truthed USGS data representing real mixtures. We selected lOxl0 blocks of pixels from three different regions 3 in the AVIRIS data of the Cuprite, Nevada mining district. We separate these 300 mixed spectra assuming two endmembers and an AR detrending with 5 AR coefficients and the MAP techniques of section 2.4. Overall brightness was accounted for as explain in the linear modeling of section 1. The endmembers are shown in figure 2A and B in comparison to laboratory spectra from the USGS spectral library for these minerals [8J . Figure 2C shows the corresponding abundances, which match the ground truth; region (III) mainly consists of Muscovite while regions (1)+(I1) contain (areal) mixtures of Kaolinite and Muscovite. 4 Discussion Hyperspectral unmixing is a challenging practical problem for unsupervised learning. Our probabilistic approach leads to several interesting algorithms: (1) simplex fitting, (2) constrained ICA and (3) constrained least squares that can efficiently use multi-channel information. An important element of our approach is the explicit use of prior information. Our simulation examples show that we can recover the endmembers, even in the presence of noise and model uncertainty. The approach described in this paper does not yet exploit local correlations between neighboring pixels that are well known to exist. Future work will therefore exploit not only spectral but also spatial prior information for detecting objects and materials. Acknowledgments We would like to thank Gregg Swayze at the USGS for assistance in obtaining the data. 3The regions were from the image plate2.cuprite95.alpha.2um.image.wlocals.gif in ftp:/ /speclab.cr.usgs.gov /pub/cuprite/gregg.thesis.images/, at the coordinates (265,710) and (275,697), which contained Kaolinite and Muscovite 2, and (143,661), which only contained Muscovite 2. L. Parra, C. D, Spence, P Sajda, A. Ziehe and K-R. Muller 948 Muscovite Kaolinite 0.8 0.7 0 .65 0.6 0.6 0.55 0.4 0.5 'c .?.? ", "'0 .. ' ., 0.45 0.3 0.4,--~--:-:-:-"-~----:--:--~ 160 190 200 waveleng1h A 210 220 180 190 200 wavelength B 210 220 C Figure 2: A Spectra of computed endmember (solid line) vs Muscovite sample spectra from the USGS data base library. Note we show only part of the spectrum since the discriminating features are located only between band 172 and 220. B Computed endmember (solid line) vs Kaolinite sample spectra from the USGS data base library. C Abundances for Kaolinite and Muscovite for three regions (lighter pixels represent higher abundance). Region 1 and region 2 have similar abundances for Kaolinite and Muscovite, while region 3 contains more Muscovite. References [1] J. Bayliss, J. A. Gualtieri, and R. Cromp. Analyzing hyperspectral data with independent component analysis. In J. M. Selander, editor, Proc. SPIE Applied Image and Pattern Recognition Workshop, volume 9, P.O. Box 10, Bellingham WA 98227-0010, 1997. SPIE. [2] J.W. Boardman and F.A. Kruse. Automated spectral analysis: a geologic example using AVIRIS data, north Grapevine Mountains, Nevada. In Tenth Thematic Conference on Geologic Remote Sensing, pages 407-418, Ann arbor, MI, 1994. Environmental Research Institute of Michigan. [3] S. Haykin. Adaptive Filter Theory. Prentice Hall, 1991. [4] F. Maselli, , M. Pieri, and C. Conese. Automatic identification of end-members for the spectral decomposition of remotely sensed scenes. Remote Sensing for Geography, Geology, Land Planning, and Cultural Heritage (SPIE) , 2960:104109,1996. [5] B. Pearlmutter and L. Parra. Maximum likelihood blind source separation: A context-sensitive generalization ofICA. In M. Mozer, M. Jordan, and T. Petsche, editors, Advances in Neural Information Processing Systems 9, pages 613-619, Cambridge MA, 1997. MIT Press. [6] J.J. Settle. Linear mixing and the estimation of ground cover proportions. International Journal of Remote Sensing, 14:1159-1177,1993. [7] M.O. Smith, J .B. Adams, and A.R. Gillespie. Reference endmembers for spectral mixture analysis. In Fifth Australian remote sensing conference, volume 1, pages 331-340, 1990. [8] U.S. Geological Survey. USGS digital spectral library. Open File Report 93-592, 1993. [9] H. Szu and C. Hsu. Landsat spectral demixing a la superresolution of blind matrix inversion by constraint MaxEnt neural nets. In Wavelet Applications IV, volume 3078, pages 147-160. SPIE, 1997. Invariant Feature Extraction and Classification in Kernel Spaces Sebastian Mika l , Gunnar Ratsch l , Jason Weston 2 , Bernhard Sch8lkopf3, Alex Smola4 , and Klaus-Robert Muller l 1 GMD FIRST, Kekulestr. 7,12489 Berlin, Germany 2 Barnhill BioInformatics, 6709 Waters Av., Savannah, GR 31406, USA 3 Microsoft Research Ltd., 1 Guildhall Street, Cambridge CB2 3NH, UK 4 Australian National University, Canberra, 0200 ACT, Australia {mika, raetsch, klaus }@first.gmd.de, jasonw@dcs.rhbnc.ac.uk bsc@microsoft.com, Alex.Smola.anu.edu.au Abstract We incorporate prior knowledge to construct nonlinear algorithms for invariant feature extraction and discrimination. Employing a unified framework in terms of a nonlinear variant of the Rayleigh coefficient, we propose non-linear generalizations of Fisher's discriminant and oriented PCA using Support Vector kernel functions . Extensive simulations show the utility of our approach. 1 Introduction It is common practice to preprocess data by extracting linear or nonlinear features. The most well-known feature extraction technique is principal component analysis PCA (e.g. [3]). It aims to find an orthonormal, ordered basis such that the i-th direction describes as much variance as possible while maintaining orthogonality to all other directions. However, since PCA is a linear technique, it is too limited to capture interesting nonlinear structure in a data set and nonlinear generalizations have been proposed, among them Kernel PCA [14], which computes the principal components of the data set mapped nonlinearly into some high dimensional feature space F. Often one has prior information, for instance, we might know that the sample is corrupted by noise or that there are invariances under which a classification should not change. For feature extraction, the concepts of known noise or transformation invariance are to a certain degree equivalent, i.e. they can both be interpreted as causing a change in the feature which ought to be minimized. Clearly, invariance alone is not a sufficient condition for a good feature, as we could simply take the constant function. What one would like to obtain is a feature which is as invariant as possible while still covering as much of the information necessary for describing the particular data. Considering only one (linear) feature vector wand restricting to first and second order statistics of the data one arrives at a maximization of the so called Rayleigh coefficient (1) 527 Invariant Feature Extraction and Classification in Kernel Spaces where w is the feature vector and Sf, SN are matrices describing the desired and undesired properties of the feature , respectively (e.g. information and noise). If S/ is the data covariance and SN the noise covariance, we obtain oriented PCA [3J . If we leave the field of data description to perform supervised classification, it is common to choose S / as the separability of class centers (between class variance) and SN to be the within class variance. In that case , we recover the well known Fisher Discriminant [7J. The ratio in (1) is maximized when we cover much of the information coded by S/ while avoiding the one coded by SN . The problem is known to be solved, in analogy to PCA , by a generalized symmetric eigenproblem S/w = >"SNW [3], where>.. E ~ is the corresponding (biggest) eigenvalue. In this paper we generalize this setting to a nonlinear one. In analogy to [8, 14J we first map the data via some nonlinear mapping <l> to some high-dimensional feature space F and then optimize (1) in F . To avoid working with the mapped data explicitly (which might be impossible if F is infinite dimensional) we introduce support vector kernel functions [11], the well-known kernel trick. These kernel functions k(x , y) compute a dot product in some feature space F , i.e. k(x , y) = (<l>(x)? <l>(y)) . Formulating the algorithms in Fusing <l> only in dot products , we can replace any occurrence of a dot product by the kernel function k. Possible choices for k which have proven useful e.g. in Support Vector Machines [2] or Kernel PCA [14J are Gaussian RBF, k(x , y) = exp( -llx - yI12/ c), or polynomial kernels , k(x , y) = (x? y)d , for some positive constants c E ~ and dEN, respectively. The remainder of this paper is organized as follows: The next section shows how to formulate the optimization problem induced by (1) in feature space. Section 3 considers various ways to find Fisher's Discriminant in F; we conclude with extensive experiments in section 4 and a discussion of our findings. 2 Kernelizing the Rayleigh Coefficient To optimize (1) in some kernel feature space F we need to find a formulation which uses only dot products of <l>-images. As numerator and denominator are both scalars this can be done independently. Furthermore, the matrices S/ and SN are basically covariances and thus the sum over outer products of <l>-images. Therefore, and due to the linear nature of (1) every solution W E F can be written as an expansion in terms of mapped training datal, i.e. l W= L Cti<l>(Xi). (2) i =l To define some common choices in F let X = {Xl , .. . ,xe} be our training sample and, where appropriate, Xl U X2 = X , Xl n X2 = 0, two subclasses (with IXi I = ?i). We get the full covariance of X by C= 1 f L (<l>(x) - m)(<l>(x) - m)T with m = f1 L <l>(x) , ~EX (3) ~EX I SB and Sw are operators on a (finite-dimensional) subspace spanned by the CP(Xi ) (in a possibly infinite space). Let w = VI + V2, where VI E Span(CP(Xi) : i = 1, .. . , f) and V2 1. Span(CP(xi) : i = 1, ... , f) . Then for S Sw or S SB (which are both symmetric) = (w , Sw) ((VI ((VI = + V2) , S(VI + V2)) + V2)S, VI) (VI , SVI) As VI lies in the span of the cp(Xi) and S only operates on this subspace there exist an expansion of w which maximizes J(w) . S. Mika. G. Riitsch. J. Weston. B. Scholkopj, A. J. Smola and K.-R. Muller 528 which could be used as Sf in oriented Kernel PCA. For SN we could use an estimate of the noise covariance, analogous to the definition of C but over mapped patterns sampled from the assumed noise distribution. The standard formulation of the Fisher discriminant in F, yielding the Kernel Fisher Discriminant (KFD) [8] is given by Sw = L L (cJ>(x) - mi)(cJ>(x) - mdT and SB = (m2 - mt}(m2 - ml)T, i=I,2 xEX; the within-class scatter Sw (as SN), and the between class scatter SB ( as Sf). Here mi is the sample mean for patterns from class i. To incorporate a known invariance e.g. in oriented Kernel PCA, one could use the tangent covariance matrix [12], T 1 = ft 2 L (cJ>(x) - cJ>(?tx))(cJ>(x) - cJ>(?tx))T for some small t> O. (4) :IlEX Here ?t is a local I-parameter transformation. T is a finite difference approximation t of the covariance of the tangent of ?t at point cJ>(x) (details e.g. in [12]). Using Sf = C and SN = T in oriented Kernel PCA, we impose invariance under the local transformation ?t. Crucially, this matrix is not only constructed from the training patterns X. Therefore, the argument used to find the expansion (2) is slightly incorrect. Neverthless, we can assume that (2) is a reasonable approximation for describing the variance induced by T. Multiplying either of these matrices from the left and right with the expansion (2), we can find a formulation which uses only dot products. For the sake of brevity, we only give the explicit formulation of (1) in F for KFD (cf. [8] for details) . Defining (I-'i)j = L:IlEXi k(xj,x) we can write (1) for KFD as t J(a) = (aTI-') 2 aTNa aTMa aTNa' (5) where N = KKT - Li=1,2fil-'iI-'T, I-' = 1-'2 - 1-'1 ' M = I-'I-'T, and Kij = k(xi,xj). The results for other choices of Sf and SN in F as for the cases of oriented kernel PCA or transformation invariance can be obtained along the same lines. Note that we still have to maximize a Rayleigh coefficient. However, now it is a quotient in terms of expansion coefficients a, and not in terms of w E F which is a potentially infinite-dimensional space. Furthermore, it is well known that the solution for this special eigenproblem is in the direction of N- 1(1-'2 - 1-'1) [7), which can be solved using e.g. a Cholesky factorization of N. The projection of a new pattern x onto w in F can then be computed by l (w? cJ>(x)) = LQik(xi'x). (6) i=1 3 Algorithms Estimating a covariance matrix with rank up to f from f samples is ill-posed. Furthermore, by performing an explicit centering in F each covariance matrix loses one more dimension, i.e. it has only rank f - 1 (even worse, for KFD the matrix N has rank f - 2). Thus the ratio in (1) is not well defined anymore, as the denominator might become zero. In the following we will propose several ways to deal with this problem in KFD. Furthermore we will tackle the question how to solve the optimization problem of KFD more efficiently. So far, we have an eigenproblem of size .e x .e. If .e becomes large this is numerically demanding. Reformulations of the original problem allow to overcome some of these limitations. Finally, we describe the connection between KFD and RBF networks. Invariant Feature Extraction and Classification in Kernel Spaces 3.1 529 Regularization and Solution on a Subspace As noted before, the matrix N has only rank ? - 2. Besides numerical problems which can cause the matrix N to be not even positive, we could think of imposing some regularization to control capacity in F. To this end, we simply add a mUltiple of the identity matrix to N, Le. replace N by NJ1. where NJ1. := N + /-LI. (7) This can be viewed in different ways: (i) for /-L > 0 it makes the problem feasible and numerically more stable as NJ1. becomes positive; (ii) it can be seen as decreasing the bias in sample based estimation of eigenvalues (cf. [6)); (iii) it imposes a regularization on 11011 2, favoring solutions with small expansion coefficients. furthermore, one could use other regularization type additives to N, e.g. penalizing IIwl1 2 in analogy to SVM (by adding the kernel matrix Kij = k(xi' Xj)). To optimize (5) we need to solve an ? x ? eigenproblem, which might be intractable for large ?. As the solutions are not sparse one can not directly use efficient algorithms like chunking for Support Vector Machines (cf. [13]). To this end, we might restrict the solution to lie in a subspace, Le. instead of expanding w by (2) we write (8) i=l with m < l. The patterns Zi could either be a subset of the training patterns X or e.g. be estimated by some clustering algorithm. The derivation of (5) does not change, only K is now m x ? and we end up with m x m matrices N and M. Another advantage is, that it increases the rank of N (relative to its size) although there still might be some need for regularization. 3.2 Quadratic optimization and Sparsification Even if N has full rank, maximizing (5) is underdetermined: if 0 is optimal, then so is any multiple thereof. Since 0T M 0 = (0T J..L)2, M has rank one. Thus we can seek for a vector 0, such that oTNo is minimal for fixed OTJ..L (e.g. to 1). The solution is unique and we can find the optimal 0 by solving the quadratic optimization problem: (9) Although the quadratic optimization problem is not easier to solve than the eigenproblem, it has an appealing interpretation. The constraint 0 T J..L = 1 ensures, that the average class distance, projected onto the direction of discrimination, is constant, while the intra class variance is minimized, i.e. we maximize the average margin. Contrarily, the SVM approach [2] optimizes for a large minimal margin. Considering (9) we are able to overcome another shortcoming of KFD . The solutions 0 are not sparse and thus evaluating (6) is expensive. To solve this we can add an h-regularizer >'110111 to the objective function, where>. is a regularization parameter allowing us to adjust the degree of sparseness. 3.3 Connection to RBF Networks Interestingly, there exists a close connection between RBF networks (e.g. [9, 1)) and KFD. If we add no regularization and expand in all training patterns, we find that an optimal 0 is given by 0 = K- 1 y, where K is the symmetric, positive matrix of all kernel elements k(xi' Xj) and y the ?1 label vector2. A RBF-network with the 2To see this, note that N can be written as N = KDK where D = I -YIyT -Y2Y; has rank e- 2, while Yi is the vector of l/Vli's for patterns from class i and zero otherwise. 530 S. Mika, G. Ratsch, J. Weston, B. SchOlkopf, A. J. Smola and K.-R. Muller RBF AB ABR SVM KFD Banana 10.8?O.06 12.3?O.07 10.9?0.04 1l.5?O.07 10.8?O.05 B.Cancer 27.6?0.47 30.4?0.47 26.5?0.45 26.o?O.4725.8?0.46 Diabetes 24.3?O.19 26.5?O.23 23.8?O.18 23.5?0.17 23.2?O.16 24.7?O.24 27.5?O.25 24.3?O.21 23.6?O.21 23.1?0.22 German Heart 17.6?O.33 20.3?O.34 16.5?O.35 16.0?O.33 16.1?0.34 Image 3.3?O.06 2.1?O.01 2.1?O.06 3.o?O.06 4.8?O.06 1.7?O.02 1.9?O.03 1.6?0.01 1.7?O.01 1.5?O.01 Ringnorm 34.4?O.20 35.7?O.18 34.2?O.22 32.4?O.18 33.2?0.11 F.Sonar Splice 10.o?O.10 10.1?O.05 9 .5?O.01 10.9?O.07 10.5?O.06 Thyroid 4.5?O.21 4.4?O.22 4.6?O.22 4.8?O.22 4.2?O.21 23.3?O.13 22.6?O.12 22.6?O.12 22.4?O.10 23.2?O.20 Titanic Twonorm 2.9?O.03 3.0?O.03 2.1?O.02 3.0?O.02 2.6?O.02 Waveform 10.7?O.1l 10.8?O.06 9.8?O.08 9. 9? o. 04 9.9?O.04 Table 1: Comparison between KFD, single RBF classifier, AdaBoost (AB), regul. Ada(ABR) Boost and SVMs (see text) . Best result in bold face, second best in italics. same kernel at each sample and fixed kernel width gives the same solution, if the mean squared error between labels and output is minimized. Also for the case of restricted expansions (8) there exists a connection to RBF networks with a smaller number of centers (cf. [4]) . 4 Experiments Kernel Fisher Discriminant Figure 1 shows an illustrative comparison of the features found by KFD, and Kernel PCA. The KFD feature discriminates the two classes, the first Kernel PCA feature picks up the important nonlinear structure. To evaluate the performance of the KFD on real data sets we performed an extensive comparison to other state-of-the-art classifiers, whose details are reported in [8j.3 We compared the Kernel Fisher Discriminant and Support Vector Machines, both with Gaussian kernel, to AdaBoost [5], and regularized AdaBoost [10] (cf. table 1). For KFD we used the regularized within-class scatter (7) and computed projections onto the optimal direction w E :F by means of (6). To use w for classification we have to estimate a threshold. This can be done by e.g. trying all thresholds between two outputs on the training set and selecting the median of those with the smallest empirical error, or (as we did here) by computing the threshold which maximizes the margin on the outputs in analogy to a Support Vector Machine, where we deal with errors on the trainig set by using the SVM soft margin approach. A disadvantage of this is, however , that we have to control the regularization constant for the slack variables. The results in table 1 show the average test error and the standard If K has full rank, the null space of D , which is spanned by Yl and Y2' is the null space of N . For a = K- 1Y we get aTN a = 0 and aTJ.? =I O. As we are free to fix the constraint aT J.? to any positive constant (not just 1), a is also feasible. 3The breast cancer domain was obtained from the University Medical Center, Inst. of Oncology, Ljubljana, Yugoslavia. Thanks to M. Zwitter and M. Soklic for the data. All data sets used in the experiments can be obtained via http://www.first.gmd.de/-raetsch/ . Figure 1: Comparison of feature found by KFD (left) and first Kernel PCA feature (right). Depicted are two classes (information only used by KFD) as dots and crosses and levels of same feature value. Both with polynomial kernel of degree two, KFD with the regularized within class scatter (7) (/1 = 10- 3 ) . Invariant Feature Extraction and Classification in Kernel Spaces 531 deviation of the averages' estimation, over 100 runs with different realizations of the datasets. To estimate the necessary parameters, we ran 5-fold cross validation on the first five realizations of the training sets and took the model parameters to be the median over the five estimates (see [10] for details of the experimental setup). Using prior knowledge. A toy example (figure 2) shows a comparison of Kernel PCA and oriented Kernel PCA, which used S[ as the full covariance (3) and as noise matrix SN the tangent covariance (4) of (i) rotated patterns and (ii) along the x-axis translated patterns. The toy example shows how imposing the desired invariance yields meaningful invariant features. In another experiment we incorporated prior knowledge in KFD. We used the USPS database of handwritten digits, which consists of 7291 training and 2007 test patterns, ~ach 2?6 .dimensional gray scale ima~es of the digits 0 ... 9: We use? the regulanzed withm-class scatter (7) (p, = 10- ) as SN and added to It an multiple A of the tangent covariance (4), i.e. SN = NJj + AT. As invariance transformations we have chosen horizontal and vertical translation, rotation, and thickening (cf. [12]), where we simply averaged the matrices corresponding to each transformation. The feature was extracted by using the restricted expansion (8), where the patterns Zi were the first 3000 training samples. As kernel we have chosen a Gaussian of width 0.3?256, which is optimal for SVMs [12]. For each class we trained one KFD which classified this class against the rest and computed the 10-class error by the winnertakes-all scheme. The threshold was estimated by minimizing the empirical risk on the normalized outputs of KFD. Without invariances, i.e. A = 0, we achieved a test error of 3.7%, slightly better than a plain SVM with the same kernel (4.2%) [12]. For A = 10- 3 , using the tangent covariance matrix led to a very slight improvement to 3.6%. That the result was not significantly better than the corresponding one for KFD (3.7%) can be attributed to the fact that we used the same expansion coefficients in both cases. The tangent covariance matrix, however, lives in a slightly different subspace. And indeed, a subsequent experiment where we used vectors which were obtained by clustering a larger dataset, including virtual examples generated by the appropriate invariance transformation, led to 3.1 %, comparable to an SVM using prior knowledge (e.g. [12]; best SVM result 2.9% with local kernel and virtual support vectors). 5 Conclusion In the task of learning from data it is equivalent to have prior knowledge about e.g. invariances or about specific sources of noise. In the case of feature extraction, we seek features which are sufficiently (noise-) invariant while still describing interesting structure. Oriented PCA and, closely related, Fisher's Discriminant, use particularly simple features, since they only consider first and second order statistics for maximizing the Rayleigh coefficient (1). Since linear methods can be too restricted in many real-world applications, we used Support Vector Kernel functions to obtain nonlinear versions of these algorithms, namely oriented Kernel PCA and Kernel Fisher Discriminant analysis. Our experiments show that the Kernel Fisher Discriminant is competitive or in Figure 2: Comparison of first features found by Kernel PCA and oriented Kernel PCA (see text); from left to right: KPCA, OKPCA with rotation and translation invariance; all with Gaussian kernel. 532 S. Mika, G. Riitsch, J. Weston, B. SchOlkopf, A. J. Smola and K.-R. Muller some cases even superior to the other state-of-the-art algorithms tested. Interestingly, both SVM and KFD construct a hyperplane in :F which is in some sense optimal. In many cases, the one given by the solution w of KFD is superior to the one of SVMs. Encouraged by the preliminary results for digit recognition, we believe that the reported results can be improved, by incorporating different invariances and using e.g. local kernels [12]. Future research will focus on further improvements on the algorithmic complexity of our new algorithms, which is so far larger than the one of the SVM algorithm, and on the connection between KFD and Support Vector Machines (cf. [16, 15]) . Acknowledgments This work was partially supported by grants of the DFG (JA 379/5-2,7-1,9-1) and the EC STORM project number 25387 and carried out while BS and AS were with GMD First. References [1) C.M. Bishop. Neural Networks for Pattern Recognition. Oxford Univ. Press, 1995. [2] B. Boser, 1. Guyon, and V.N. Vapnik. A training algorithm for optimal margin classifiers. In D. Haussler, editor, Proc. COLT, pages 144- 152. ACM Press, 1992. [3) K.I . Diamantaras and S.Y. Kung. Principal Component Neural Networks. Wiley, New York,1996. [4] B.Q. Fang and A.P. Dawid. Comparison of full bayes and bayes-least squares criteria for normal discrimination. Chinese Journal of Applied Probability and Statistics, 12:401- 410, 1996. [5] Y. Freund and R.E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. In EuroCOLT 94. LNCS, 1994. [6] J.H. Friedman. Regularized discriminant analysis. Journal of the American Statistical Association, 84(405):165- 175, 1989. [7] K Fukunaga. Introduction to Statistical Pattern Recognition. Academic Press, San Diego, 2nd edition, 1990. [8) S. Mika, G. Ratsch, J. Weston, B. Scholkopf, and K-R. Muller. Fisher discriminant analysis with kernels. In Y.-H. Hu, J . Larsen, E. Wilson, and S. Douglas, editors, Neural Networks for Signal Processing IX, pages 41-48. IEEE, 1999. [9] J. Moody and C. Darken. Fast learning in networks of locally-tuned processing units. Neural Computation, 1(2):281-294, 1989. [10] G. Ratsch, T. Onoda, and K-R. Muller. Soft margins for adaboost. Technical Report NC-TR-1998-021, Royal Holloway College, University of London, UK, 1998. [11] S. Saitoh. Theory of Reproducing Kernels and its Applications. Longman Scientific & Technical, Harlow, England, 1988. [12] B. Scholkopf. Support vector learning. Oldenbourg Verlag, 1997. [13) B. Scholkopf, C.J.C. Burges, and A.J. Smola, editors. Advances in Kernel Methods Support Vector Learning. MIT Press, 1999. [14] B. Scholkopf, A.J. Smola, and K-R. Muller. Nonlinear component analysis as a kernel eigenvalue problem. Neural Computation, 10:1299-1319, 1998. [15] A. Shashua. On the relationship between the support vector machine for classification and sparsified fisher's linear discriminant. Neural Processing Letters, 9(2):129- 139, April 1999. [16) S. Tong and D. Koller. Bayes optimal hyperplanes --+ maximal margin hyperplanes. Submitted to IJCA1'99 WorkshOp on Support Vector Machines (robotics. stanford. edurkoller/), 1999.
1715 |@word determinant:2 version:1 inversion:1 polynomial:2 proportion:1 underline:1 nd:1 tedious:1 open:3 hu:1 simulation:3 seek:2 sensed:1 covariance:14 crucially:1 decomposition:2 brightness:1 pick:1 tr:1 solid:2 moment:1 substitution:1 series:1 contains:2 united:1 selecting:1 pub:1 tuned:1 interestingly:3 ati:1 riitsch:2 recovered:2 current:1 contextual:1 com:2 yet:1 scatter:8 written:2 oldenbourg:1 subsequent:1 distant:1 additive:2 numerical:1 remove:1 plot:3 xex:1 discrimination:3 alone:1 v:3 selected:1 geologic:2 parameterization:1 smith:1 haykin:1 filtered:1 regressive:1 detecting:1 boosting:1 contribute:1 location:1 hyperplanes:2 district:1 misinterpreted:1 five:2 atj:1 along:2 constructed:2 become:1 scholkopf:6 incorrect:1 consists:4 fitting:4 inside:1 introduce:1 indeed:1 ica:6 udv:1 frequently:1 planning:1 multi:1 y2y:1 eurocolt:1 decreasing:1 gov:2 actual:2 boardman:1 considering:2 lib:1 becomes:2 ofica:1 begin:1 unrelated:1 estimating:1 maximizes:2 project:1 cultural:1 null:2 what:1 vapor:1 mountain:1 superresolution:1 gif:1 interpreted:1 unified:1 finding:4 transformation:7 corporation:1 sparsification:1 ought:1 nj:1 guarantee:1 every:2 act:1 subclass:1 tackle:1 um:1 classifier:3 uk:3 control:2 unit:1 grant:1 omit:1 diamantaras:1 medical:1 before:1 positive:5 understood:1 local:5 treat:1 oxford:1 analyzing:2 mika:6 might:6 au:1 aviris:2 challenging:1 ringnorm:1 limited:1 factorization:1 sarnoff:2 statistically:1 averaged:1 acknowledgment:2 unique:1 practical:2 spence:4 practice:2 block:1 bellingham:1 svi:1 cb2:1 digit:3 lncs:1 area:2 empirical:2 remotely:1 significantly:1 projection:2 integrating:1 suggest:1 get:2 onto:3 close:2 operator:1 prentice:1 context:1 impossible:1 risk:1 influence:1 optimize:3 www:1 equivalent:2 map:18 conventional:1 center:3 maximizing:5 independently:2 l:1 survey:2 resolution:1 formulate:1 simplicity:1 pure:2 m2:3 haussler:1 importantly:1 spanned:4 orthonormal:1 fang:1 variation:2 coordinate:1 analogous:1 target:3 diego:1 user:3 lighter:1 us:2 ixi:1 origin:1 diabetes:1 pa:7 element:3 dawid:1 satisfying:1 recognition:4 located:1 trend:3 trick:1 expensive:1 cspence:1 szu:2 particularly:1 database:4 observed:2 ft:2 solved:2 capture:2 thousand:1 region:8 ensures:1 remote:5 highest:1 loxl0:1 ran:1 substantial:1 mentioned:1 discriminates:1 mozer:1 complexity:1 instructive:1 signature:1 trained:1 solving:3 basis:1 completely:1 usps:1 translated:1 represented:2 various:1 tx:2 regularizer:1 derivation:1 sajda:4 univ:1 fast:1 describe:1 shortcoming:1 london:1 artificial:3 klaus:4 quite:1 whose:1 stanford:1 posed:1 larger:2 solve:5 otherwise:1 toeplitz:1 statistic:3 think:1 noisy:3 itself:1 advantage:2 eigenvalue:3 net:1 nevada:2 propose:2 took:1 interaction:2 product:6 maximal:1 remainder:2 neighboring:1 causing:1 combining:1 gregg:2 realization:2 mixing:3 description:1 validate:1 nj1:3 convergence:1 p:6 ach:1 cluster:1 argmaxp:1 unmixing:13 adam:1 leave:1 rotated:1 object:2 ftp:1 derive:2 depending:3 ac:1 geology:1 measured:2 eq:2 implemented:2 recovering:1 quotient:1 australian:2 direction:5 guided:1 inhomogeneous:2 waveform:1 closely:1 njj:1 filter:1 correct:1 exploration:1 australia:1 viewing:1 settle:1 material:17 virtual:2 require:1 ja:1 atmosphere:1 f1:1 generalization:4 geography:1 preliminary:1 ao:2 fix:1 probable:1 absorption:4 parra:5 underdetermined:1 fil:1 sufficiently:1 hall:1 ground:6 normal:4 guildhall:1 exp:1 algorithmic:1 mapping:1 lm:1 a2:1 smallest:1 lparra:1 proc:2 estimation:3 label:2 superposition:1 sensitive:2 weighted:1 bsc:1 clearly:3 sensor:2 gaussian:5 mit:2 aim:7 pn:4 avoid:1 cr:2 wilson:1 encode:1 ax:1 focus:1 improvement:2 rank:9 likelihood:3 mainly:1 greatly:1 contrast:1 sense:2 inst:1 posteriori:2 savannah:1 dependent:2 landsat:1 sb:4 inaccurate:2 typically:2 hidden:4 koller:1 favoring:1 expand:1 interested:1 i1:1 germany:2 pixel:16 overall:4 classification:8 ill:1 among:1 html:1 priori:3 lucas:1 colt:1 denoted:1 art:2 spatial:3 constrained:10 special:1 field:1 construct:2 once:1 eigenproblem:5 having:1 extraction:8 encouraged:1 identical:1 represents:4 unsupervised:8 nearly:1 future:2 simplex:11 report:2 minimized:3 inherent:1 few:3 oriented:10 simultaneously:1 national:1 descript:1 dfg:1 ima:1 microsoft:2 friedman:1 ab:2 kfd:25 mining:2 intra:1 adjust:1 introduces:2 mixture:15 arrives:1 yielding:2 necessary:2 iv:1 logarithm:1 maxent:1 circle:2 desired:3 minimal:2 instance:2 column:2 modeling:3 kij:2 soft:2 ar:7 disadvantage:1 cover:2 maximization:2 kpca:1 cost:1 fusing:1 vertex:2 subset:1 deviation:1 ada:1 hundred:1 usefulness:1 cuprite:2 entry:2 conducted:1 gr:1 too:2 reported:2 dependency:1 varies:1 corrupted:1 considerably:1 ljubljana:1 st:1 thanks:1 international:1 discriminating:1 twonorm:1 probabilistic:3 systematic:1 yl:1 aix:1 pool:1 invertible:1 moody:1 thesis:1 imagery:3 squared:1 nm:2 containing:1 choose:1 possibly:1 smap:2 positivity:4 worse:1 creating:1 expert:1 american:1 li:2 toy:2 account:2 de:3 bold:1 summarized:1 includes:1 coefficient:12 north:1 satisfy:1 explicitly:2 blind:2 vi:8 performed:1 try:3 jason:1 analyze:1 observing:1 shashua:1 competitive:1 bayes:3 recover:2 psajda:1 square:5 ni:1 kekulestr:2 variance:6 characteristic:4 efficiently:2 maximized:1 correspond:1 preprocess:1 yield:2 identify:1 generalize:1 bayesian:1 handwritten:1 identification:1 produced:1 basically:1 monitoring:1 multiplying:1 classified:1 submitted:1 detector:1 explain:1 barnhill:1 sebastian:1 definition:1 centering:1 against:1 vli:1 larsen:1 storm:1 thereof:1 naturally:1 spie:4 mi:3 vector2:1 attributed:1 sampled:1 hsu:1 dataset:1 knowledge:8 organized:1 cj:8 clay:1 actually:1 higher:2 supervised:5 adaboost:4 improved:1 april:1 formulation:4 done:2 shrink:1 box:1 furthermore:5 just:1 smola:6 until:1 d:1 working:1 correlation:2 horizontal:1 replacing:1 nonlinear:11 atn:1 lda:1 gray:1 indicated:1 scientific:1 believe:1 usa:2 contain:1 true:2 concept:1 y2:1 normalized:1 regularization:8 spatially:1 symmetric:3 laboratory:1 deal:2 undesired:1 white:1 ll:5 assistance:1 width:2 numerator:1 covering:1 illustrative:1 ambiguous:1 noted:1 oc:1 criterion:1 generalized:1 trying:1 presenting:1 theoretic:1 pearlmutter:1 cp:4 reflection:1 endmember:5 ranging:1 image:10 common:3 rotation:2 superior:2 mt:1 physical:3 volume:4 nh:1 association:1 interpretation:1 slight:1 numerically:2 raetsch:2 significant:2 cambridge:2 imposing:2 ai:1 llx:1 areal:2 automatic:1 outlined:1 particle:1 had:1 dot:6 stable:1 surface:5 base:2 add:3 saitoh:1 gualtieri:1 dust:1 posterior:1 optimizes:1 prime:1 certain:1 verlag:1 life:1 xe:1 yi:1 muller:10 exploited:1 seen:1 additional:1 impose:1 eo:1 determine:1 maximize:2 kruse:1 signal:3 ii:8 multiple:5 full:5 violate:1 infer:1 technical:2 match:4 academic:1 england:1 characterized:1 cross:2 lin:1 coded:2 yugoslavia:1 prediction:1 variant:1 basic:1 denominator:2 breast:1 represent:5 kernel:46 sometimes:1 normalization:3 robotics:1 achieved:1 lea:3 whereas:1 fine:1 ratsch:4 singular:1 source:3 median:2 leaving:1 envelope:1 rest:1 contrarily:1 file:1 subject:4 induced:2 member:1 jordan:1 extracting:1 presence:1 iii:2 automated:1 variety:1 xj:4 fit:1 zi:2 winnertakes:1 affect:1 restrict:2 reduce:1 andreas:1 cn:1 otj:1 six:3 pca:20 rhbnc:1 utility:1 ltd:1 penalty:1 f:3 miiller:1 york:1 cause:6 ignored:1 harlow:1 useful:1 detailed:1 amount:1 band:6 locally:1 svms:3 gmd:6 reduced:1 http:2 schapire:1 exist:2 problematic:1 ipa:1 estimated:2 delta:1 zwitter:1 write:3 gunnar:1 datal:1 threshold:4 nevertheless:1 changing:1 penalizing:1 douglas:1 tenth:1 longman:1 usgs:9 sum:1 wand:1 run:1 angle:1 inverse:1 letter:1 uncertainty:1 parameterized:1 family:1 reasonable:2 guyon:1 utilizes:1 separation:1 decision:1 comparable:1 entirely:1 bound:1 simplification:1 fold:1 quadratic:4 constraint:16 incorporation:2 kronecker:1 alex:2 orthogonality:1 x2:2 scene:3 sake:1 endmembers:24 thyroid:1 argument:1 fukunaga:1 span:3 formulating:1 performing:1 relatively:1 according:1 alternate:1 combination:2 across:3 describes:1 em:2 separability:1 slightly:3 smaller:1 appealing:1 b:1 explained:1 restricted:3 den:1 invariant:8 taken:2 chunking:1 heart:1 equation:2 previously:2 slack:1 discus:1 amap:1 describing:4 german:1 know:2 abr:2 end:4 reformulations:1 apply:2 v2:5 appropriate:2 spectral:19 petsche:1 occurrence:1 anymore:1 kernelizing:1 subtracted:1 original:2 clustering:3 cf:7 include:1 heritage:1 maintaining:1 sw:5 exploit:2 reflectance:11 build:1 chinese:1 objective:2 added:2 mdt:1 question:1 speclab:2 italic:1 gradient:2 subspace:6 distance:1 separate:2 thank:1 berlin:2 geological:2 mineral:5 capacity:1 street:1 outer:1 evenly:1 mapped:5 considers:2 discriminant:13 water:1 assuming:2 besides:1 relationship:1 ratio:2 minimizing:1 innovation:3 nc:1 setup:1 robert:2 potentially:4 unknown:1 perform:3 allowing:1 av:1 vertical:1 observation:5 datasets:1 sm:5 darken:1 finite:2 sparsified:1 defining:1 incorporated:3 banana:1 dc:1 reproducing:1 oncology:1 intensity:1 atmospheric:2 namely:1 nonlinearly:1 extensive:3 connection:5 inclination:1 boser:1 boost:1 address:1 able:1 suggested:1 usually:1 pattern:17 departure:2 yi12:1 royal:1 including:2 gillespie:1 ia:1 demanding:1 difficulty:2 regularized:4 representing:1 scheme:1 improve:1 technology:2 titanic:1 library:4 identifies:1 axis:1 carried:1 auto:1 sn:12 text:2 prior:24 tangent:6 determining:1 relative:1 freund:1 mixed:2 interesting:5 limitation:1 thickening:1 analogy:4 proven:1 iiwl1:1 digital:1 validation:1 integrate:2 degree:3 sufficient:1 imposes:1 principle:1 editor:5 land:1 translation:2 cancer:2 prone:1 l8:1 accounted:1 summary:2 supported:1 free:6 bias:3 allow:2 burges:1 institute:1 face:2 fifth:1 sparse:2 kaolinite:6 distributed:3 benefit:1 overcome:2 xia:3 boundary:1 world:1 valid:1 plain:1 dimension:1 evaluating:2 kdk:1 made:1 adaptive:1 san:1 preprocessing:1 projected:1 computes:1 far:2 ec:1 employing:1 simplified:1 constituting:2 alpha:1 obtains:1 ignore:1 bernhard:1 ml:2 kkt:1 assumed:2 conclude:1 xi:11 spectrum:15 terrain:1 un:1 sonar:1 table:3 additionally:2 channel:1 onoda:1 nature:1 expanding:1 trainig:1 obtaining:1 expansion:9 domain:1 diag:1 da:1 did:1 noise:30 paul:1 edition:1 canberra:1 biggest:1 bayliss:1 tong:1 wiley:1 illi:1 thematic:1 deterministically:1 explicit:4 sf:5 xl:3 lie:2 candidate:1 intimate:1 muscovite:9 pe:6 jacobian:2 wavelet:1 grained:1 splice:1 abundance:18 removing:2 ix:1 specific:1 bishop:1 showing:1 sensing:5 x:2 svm:9 a3:2 demixing:1 intractable:1 incorporating:1 exists:2 restricting:1 adding:2 workshop:2 vapnik:1 hyperspectral:14 sparseness:1 margin:7 anu:1 easier:1 entropy:1 depicted:1 michigan:1 rayleigh:5 wavelength:8 simply:5 led:3 contained:2 ordered:1 partially:1 scalar:1 applies:1 truthed:1 loses:1 truth:1 relies:1 acm:1 ma:1 environmental:2 weston:5 conditional:1 cti:1 identity:2 extracted:1 viewed:1 ann:1 soklic:1 rbf:8 replace:2 fisher:12 feasible:2 change:4 hard:1 infinite:3 operates:1 hyperplane:1 principal:3 called:3 invariance:13 svd:1 experimental:1 arbor:1 la:1 meaningful:1 e:1 indicating:1 holloway:1 college:1 ziehe:5 support:13 cholesky:1 latter:1 kung:1 brevity:2 bioinformatics:1 avoiding:1 incorporate:2 evaluate:1 princeton:1 tested:1 phenomenon:2 ex:2
782
1,716
Bayesian Map Learning in Dynamic Environments Kevin P. Murphy Computer Science Division University of California Berkeley, CA 94720-1776 murphyk@cs.berkeley.edu Abstract We consider the problem of learning a grid-based map using a robot with noisy sensors and actuators. We compare two approaches: online EM, where the map is treated as a fixed parameter, and Bayesian inference, where the map is a (matrix-valued) random variable. We show that even on a very simple example, online EM can get stuck in local minima, which causes the robot to get "lost" and the resulting map to be useless. By contrast, the Bayesian approach, by maintaining multiple hypotheses, is much more robust. We then introduce a method for approximating the Bayesian solution, called Rao-Blackwellised particle filtering. We show that this approximation, when coupled with an active learning strategy, is fast but accurate. 1 Introduction The problem of getting mobile robots to autonomously learn maps of their environment has been widely studied (see e.g., [9] for a collection of recent papers). The basic difficulty is that the robot must know exactly where it is (a problem called localization), so that it can update the right part of the map. However, to know where it is, the robot must already have a map: relying on dead-reckoning alone (Le., integrating the motor commands) is unreliable because of noise in the actuators (slippage and drift). One obvious solution is to use EM, where we alternate between estimating the location given the map (the E step), and estimating the map given the location (the M step). Indeed, this approach has been successfully used by several groups [8, 11, 12]. However, in all of these works, the trajectory of the robot was specified by hand, and the map was learned off-line. For fully autonomous operation, and to cope with dynamic environments, the map must be learned online. We consider two approaches to online learning: online EM, and Bayesian inference, K. P Murphy 1016 a c Figure 1: (a) The POMDP represented as a graphical model. L t is the location, Mt(i) is the label of the i'th grid cell, At is the action, and Zt is the observation. Dotted circles denote variables that EM treats as parameters. (b) A one-dimensional grid with binary labels (white = 0, black = 1). (c) A two-dimensional grid, with four labels (closed doors, open doors, walls, and free space). where we treat the map as a random variable. In Section 3, we show that the Bayesian approach can lead to much better results than online EM; unfortunately, it is computationally intractable, so in Section 4, we discuss an approximation based on Rao-BIackwellised particle filtering. 2 The model We now precisely define the model that we will use in this paper; it is similar to, but much simpler than, the occupancy grid model in [12]. The map is defined to be a grid, where each cell has a label which represents what the robot would see at that point. More formally, the map at time t is a vector of discrete random variables, M t (i) E {I, ... , No}, where 1 ::; i ::; N L. Of course, the map is not observed directly, and nor is the robot's location, L t E {I, ... , NL}. What is observed is Zt E {l, ... ,No}, the label of the cell at the robot's current location, and At E {I, ... ,NA}, the action chosen by the robot just before time t. The conditional independence assumptions we are making are illustrated in Figure l(a). We start by considering the very simple one-dimensional grid shown in Figure l(b), where there are just two actions, move right (-+) and move left (f-), and just two labels, off (0) and on (1). This is sufficiently small that we can perform exact Bayesian inference. Later, we will generalize to two dimensions. The prior for the location is a delta function with all its mass on the first (left-most) cell, independent of AI. The transition model for the location is as follows. Pa P r ( L t = J?1 Lt-I =~,. A t =-+ ) = { 11 - Pa o if j = i + 1, j < N if j = i, j < N if j = i = N otherwise where Pa is the probability of a successful action, i.e., 1 - Pa is the probability that the robot's wheels slip. There is an analogous equation for the case when At =f-. Note that it is not possible to pass through the "rightmost" cell; the robot can use this information to help localize itself. The prior for the map is a product of the priors for each cell, which are uniform. (We could model correlation between neighboring cells using a Markov Random Field, although this is computationally expensive.) The transition model for the map is a product of the transition models for each cell, which are defined as follows: Bayesian Map Learning in Dynamic Environments 1017 the probability that a 0 becomes a 1 or vice versa is Pc (probability of change), and hence the probability that the cell label remains the same is 1 - Pc. Finally, the observation model is Pr(Zt = klMt = (mI , ... , mNL), L t = i) = { Po 1- Po if mi = k otherwise where Po is the probability of a succesful observation, Le. , 1 - Po is the probability of a classification error. Another way of writing this, that will be useful later, is to introduce the dummy deterministic variable, which has the following distribution: Pr(Z: = klMt = (mI, ... ,mNL) , L t = i) = 8(k,mi) , where 8(a, b) = 1 if a = b and is 0 otherwise. Thus acts just like a multiplexer, selecting out a component of M t as determined by the "gate" Lt. The output of the multiplexer is then passed through a noisy channel, which flips bits with probability 1 - Po, to produce Zt. Z:, Z: 3 Bayesian learning compared to EM For simplicity, we assume that the parameters Po, Pa and Pc, are all known. (In this section, we use Po = 0.9, Pa = 0.8 and Pc = 0, so the world is somewhat "slippery", but static in appearance.) The state estimation problem is to compute the belief state Pr(L t , MtIYl:t), where Yt = (Zt, At) is the evidence at time t; this is equivalent to performing online inference in the graphical model shown in Figure 1(a). Unfortunately, even though we have assumed that the components of M t are a priori independent, they become correlated by virtue of sharing a common child, Zt. That is, since the true location of the robot is unknown, all of the cells are possible causes of the observation, and they "compete" to "explain" the data. Hence all of the hidden variables become coupled, and the belief state has size O(NL2NL). If the world is static (Le. , Pc = 0) , we can treat M as a fixed , but unknown, parameter; this can then be combined with the noisy sensor model to define an HMM with the following observation matrix: B(i , k) ~ Pr(Zt = kiLt = i; M) = L Pr(Zt = klZ: = j)8(M(i),j) j We can then learn B using EM, as in [8, 11, 12]. (We assume for now that the HMM transition matrix is independent of the map, and encodes the known topology of the grid, Le., the robot can move to any neighboring cell, no matter what its label is. We will lift this restriction in the 2D example. We can formulate an online version of EM as follows. We use fixed-lag s"moothing with a sliding window of length W, and compute the expected sufficient statistics (ESS) for the observation matrix within this window as follows: Ot(i, k) = 2:~=t-W : Z," =k LT1t(i) , where LTlt(i) = Pr(LT = iIYl:t)? We can compute L using the forwards-backwards algorithm, using Lt-W-Ilt-I as the prior. (The initial condition is L = 11", where 11" is the (known) prior for Lo.) Thus the cost per time step is O(2W Nl). In the M step, we normalize each row of Ot + d x Ot-l, where 0 < d < 1 is a decay constant, to get the new estimate of B . We need to downweight the previous ESS since they were computed using out-of-date parameters; in addition, exponential forgetting allows us to handle dynamic environments. [1] discuss some variations on this algorithm. 1018 K. P. Murphy _al_. .,. . . ., '. . __'. . ......,.'I!'. a ~Ir! ' . lei b c d Figure 2: (a) The full joint posterior on P(Mt !Yl:t). 0 and 255, on the axis into the page, represent the maps where every cell is off and every cell is on, respectively; the mode at t = 16 is for map 171, which corresponds to the correct pattern 01010101. (b-d) Estimated map. Light cells are more likely to contains Os, so the correct pattern should have light bars in the odd rows. (b) The marginals of the exact joint. (c) Online EM. (d) Omine EM. As the window length increases, past locations are allowed to look at more and more future data, and hence their estimates become more accurate; however, the space and time requirements increase. Nevertheless, there are occasions when even the maximum window size (i.e., looking all the way back to 'T = 0) will perform poorly, because of the greedy hill-climbing nature of EM. For a simple example of this, consider the environment shown in Figure 1 (b). Suppose the robot starts in cell 1, keeps going right until it comes to the end of the "corridor", and then heads back "home". Suppose further that there is a single slippage error at t = 4, so the actual path and observation sequence of the robot is as follows: t Lt Zt At 1 1 0 2 2 1 3 4 3 o --7 --7 4 1 5 6 7 456 101 8 7 0 --7 --7 --7 --7 --7 9 10 11 876 101 +- +- +- 12 13 14 15 16 54321 01010 +- +- +- +- +- To study the effect of this sequence, we computed Pr(Mt , L t !Yl:t) by applying the junction tree algorithm to the graphical model in Figure l(a). We then marginalized out L t to compute the posterior P(Mt ): see Figure 2(a). At t = 1, there are 27 modes, corresponding to all possible bit patterns on the unobserved cells. At each time step, the robot thinks it is moving one step to the right. Hence at t = 8, the robot thinks it is in cell 8, and observes O. When it tries to move rightf it knows it will remain in cell 8 (since the robot knows where the boundaries are). Hence at t = 9, it is almost 70% confident that it is in cell 8. At t = 9, it observes a 1, which contradicts its previous observation of O. There are two possible explanations: this is a sensor error, or there was a motor error. Which of these is more likely depends on the relative values of the sensor noise, Po, and the system noise, Pa. In our experiments, we found that the motor error hypothesis is much more likely; hence the mode of the posterior jumps from the wrong map (in which M(5) = 1) to the right map (in which M(5) = 0). Furthermore, as the robot returns to "familiar territory", it is able to better localize itself (see Figure 3(a)), and continues to learn the map even for far-away cells, because they are all correlated (in Figure 2(b), the entry for cell 8 becomes sharper even as the robot returns to cell 1) We now compare the Bayesian solution with EM. Online EM with no smoothing was not able to learn the correct map. Adding smoothing with the maximum window size of Wt = t did not improve matters: it is still unable to escape the local Bayesian Map Learning in Dynamic Environments a b 1019 I c Figure 3: Estimated location. Light cells are more likely to contain the robot. (a) Optimal Bayes solution which marginalizes out the map. (b) Dead-reckoning solution which ignores the map. Notice how "blurry" it is. (c) Online EM solution using fixed-lag smoothing with a maximal window length. minimum in which M(5) = 1, as shown in Figure 2(c). (We tried various values of the decay rate d, from 0.1 to 0.9, and found that it made little difference.) With the wrong map, the robot "gets lost" on the return journey: see Figure 3(c). Offline EM, on the other hand, does very well, as shown in Figure 2(d); although the initial estimate oflocation (see Figure 3(b)) is rather diffuse, as it updates the map it can use the benefit of hindsight to figure out where it must have been. 4 Rao-Blackwellised particle filtering Although the Bayesian solution exhibits some desirable properties, its running time is exponential in the size of the environment. In this section, we discuss a sequential Monte Carlo algorithm called particle filtering (also known as sm filtering, the bootstrap filter, the condensation algorithm, survival of the fittest, etc; see [10, 4] for recent reviews). Particle filtering (PF) has already been successfully applied to the problem of (global) robot localization [5]. However, in that case, the state space was only of dimension 3: the unknowns were the position of the robot, (x, y) E lR?, and its orientation, () E [0,211"]. In our case, the state space is discrete and of dimension 0(1 + NL), since we need to keep track of the map as well as the robot's location (we ignore orientation in this paper). Particle filtering can be very inefficient in high-dimensional spaces. The key observation which makes it tractable in this context is that, if Ll:t were known, then the posterior on M t would be factored; hence M t can be marginalized out analytically, and we only need to sample Lt. This idea is known in the statistics literature as RaoBlackwellisation [10, 41. In more detail, we will approximate the posterior at time t using a set of weighted particles, where each particle specifies a trajectory L 1 :t , and the corresponding conditionally factored representation of P(Mt ) = TIi P(Mt(i)); we will denote the j'th particle at time t as bF). Note that we do not need to actually store the complete trajectories Ll:t: we only need the most recent value of L. The approach we take is essentially the same as the one used in the conditional linear Gaussian models of [4, 3], except we replace the Kalman filter update with one which exploits the conditionally factored representation of P(Mt ). In particular, the algorithm is as follows: For each particle j = 1, ... , N s , we do the following: 1. Sample L~~l from a proposal distribution, which we discuss below. 2. Update each component of the map separately using L~~l and Zt+1 Pr(Mt~lIL~~l = i,bP),Zt+l) oc Pr(zt+1IMt~l(i)) rrPr(Mi~l(k)IMF)(k)) k K. P. Murphy 1020 IIK _ a b _ ... '~I ~ I. I I I d c Figure 4: (a-b) Results using 50 particles. (c-d) Results using BK. . 3. Update the weIghts: (j) Wt+l (j) (j) = u t + 1 w t ,where (j) Ut+l . IS defined below. We then res ample Ns particles from the normalised weights, using Liu's residual resampling algorithm [10], and set = 1/Ns for all j. We consider two proposal distributions. The first is a simple one which just uses the transition model to predict the new location: Pr(Lt+1lb~j), at+1) . In this case, the incremental weight is U~~l <X P(zt+1IL~~l,b~j)). The optimal proposal distribution (the one which minimizes the variance of the importance weights) takes the most recent evidence into account, and can be shown to have the form Pr(Lt+1lb~j), at+l, Zt+l) with incremental weight Ut+1 <X P(Zt+1lb~j)) . Computing this requires marginalizing out Mt+l and Lt+l' which can be done in O(NL) time (details omitted). WWl In Figure 4, we show the results of applying the above algorithm to the same problem as in Section 3; it can be seen that it approximates the exact solution- very closely, using only 50 particles. The results shown are for a particular random number seed; other seeds produce qualitatively very similar results, indicating that 50 particles is in fact sufficient in this case. Obviously, as we increase the number of particles, the error and variance decrease, but the running time increases (linearly). The question of how many particles to use is a difficult one: it depends both on the noise parameters and the structure of the environment (if every cell has a unique label, localization is easy). Since we are sampling trajectories, the number of hypotheses, and hence the number of particles needed, grows exponentially with time. In the above example, the robot was able to localize itself quite accurately when it reached the end of the corridor, where most hypotheses "died off". In general, the number of particles will depend on the length of the longest cycle in the environment, so we will need to use active learning to ensure tractability. In the dynamic two-dimensional grid world of Figure l(c), we chose actions so as to maximize expected discounted reward (using policy iteration), where the reward for visiting cell i is where H(?) is the normalized entropy. Hence, if the robot is "lost", so H(L t ) ~ 1, the robot will try to visit a cell which it is certain about (see [6] for a better approach); otherwise, it will try to explore uncertain cells. After learning the map, the robot spends its time visiting each of the doors, to keep its knowledge of their state (open or closed) up-to-date. We now briefly consider some alternative approximate inference algorithms. Examining the graphical structure of our model (see Figure l(a)) , we see that it is identical 1021 Bayesian Map Learning in Dy namic Environments to a Factorial HMM [7] (ignoring the inputs). Unfortunately, we cannot use their variational approximation, because they assume a conditional Gaussian observation model, whereas ours is almost deterministic. Another popular approximate inference algorithm for dynamic Bayes nets (DBNs) is the "BK algorithm" [2, 1]. This entails projecting the joint posterior at time t onto a product-of-marginals representation P(Lt, Mt (1) , . . . , Mt(NdIYl:t) = P(Lt IYl :t) II P(Mt(i)IYl :t) i and using this as a factored prior for Bayesian updating at time t + 1. Given a factored prior, we can compute a factored posterior in O(NL) time by conditioning on each L t +1, and then averaging. We found that the BK method does very poorly on this problem (see Figure 4), because it ignores correlation between the cells. Of course, it is possible to use pairwise or higher order marginals for tightly coupled sets of variables. Unfortunately, the running time is exponential in the size of the largest marginal , and in our case, all the Mt(i) variables are coupled. Acknowledgments I would like to thank Nando de Freitas for helping me get particle filtering to work, Sebastian Thrun for an interesting discussion at the conference, and Stuart Russell for encouraging me to compare to EM . This work was supported by grant number ONR N00014-97-1-0941. References [1) X. Boyen and D. Koller. Approximate learning of dynamic models. In NIPS, 1998 . [2) X. Boyen and D. Koller. Tractable inference for complex stochastic processes. In UAI, 1998. [3) R. Chen and S. Liu. Mixture Kalman filters . Submitted, 1999. [4) A. Doucet, S. Godsill, and C. Andrieu. On sequential Monte Carlo sampling methods for Bayesian filtering. Statistics and Computing, 1999. [5) D. Fox, W. Burgard, F. Dellaert, and S. Thrun. Monte carlo localization: Efficient position estimation for mobile robots. In AAAI, 1999. [6) D. Fox, W . Burgard, and S. Thrun. Active Markov localization for mobile robots. Robotics and Autonomous Systems, 1998. [7] Z. Ghahramani and M. Jordan. Factorial Hidden Markov Models. Machine Learning, 29:245- 273, 1997. [8) S. Koenig and R. Simmons. Unsupervised learning of probabilistic models for robot navigation. In ICRA, 1996. [9] D. Kortenkamp , R. Bonasso, and R. Murphy, editors. Artificial Intelligence and Mobile Robots: case studies of successful robot systems. MIT Press, 1998. [10] J . Liu and R. Chen. Sequential monte carlo methods for dynamic systems. JASA , 93:1032-1044, 1998. [11) H. Shatkay and L. P. Kaelbling. Learning topological maps with weak local odometric information. In IlCAI, 1997. [12) S. Thrun, W . Burgard, and D. Fox. A probabilistic approach to concurrent mapping and localization for mobile robots. Machine Learning, 31:29- 53, 1998.
1716 |@word briefly:1 version:1 bf:1 open:2 tried:1 initial:2 liu:3 contains:1 selecting:1 ours:1 rightmost:1 past:1 freitas:1 current:1 must:4 motor:3 update:5 resampling:1 alone:1 greedy:1 intelligence:1 es:2 lr:1 location:12 simpler:1 become:3 corridor:2 introduce:2 pairwise:1 forgetting:1 expected:2 indeed:1 nor:1 relying:1 discounted:1 actual:1 little:1 window:6 considering:1 pf:1 becomes:2 encouraging:1 estimating:2 mass:1 what:3 namic:1 minimizes:1 spends:1 unobserved:1 hindsight:1 iiyl:1 berkeley:2 blackwellised:2 every:3 act:1 exactly:1 wrong:2 murphyk:1 grant:1 before:1 local:3 treat:3 died:1 path:1 black:1 chose:1 studied:1 unique:1 acknowledgment:1 lost:3 bootstrap:1 integrating:1 get:5 cannot:1 wheel:1 onto:1 context:1 applying:2 writing:1 restriction:1 equivalent:1 map:37 deterministic:2 yt:1 pomdp:1 formulate:1 simplicity:1 factored:6 handle:1 autonomous:2 variation:1 analogous:1 simmons:1 dbns:1 suppose:2 exact:3 us:1 slip:1 hypothesis:4 pa:7 expensive:1 updating:1 continues:1 slippage:2 observed:2 cycle:1 autonomously:1 decrease:1 russell:1 observes:2 environment:11 reward:2 dynamic:9 depend:1 iyl:2 localization:6 division:1 po:8 joint:3 represented:1 various:1 fast:1 monte:4 artificial:1 lift:1 kevin:1 quite:1 lag:2 widely:1 valued:1 otherwise:4 statistic:3 succesful:1 think:2 noisy:3 itself:3 online:11 obviously:1 sequence:2 net:1 product:3 maximal:1 neighboring:2 date:2 poorly:2 fittest:1 normalize:1 getting:1 requirement:1 produce:2 incremental:2 help:1 odd:1 c:1 come:1 closely:1 correct:3 filter:3 stochastic:1 nando:1 wall:1 helping:1 sufficiently:1 seed:2 mapping:1 predict:1 omitted:1 estimation:2 label:9 largest:1 concurrent:1 vice:1 successfully:2 weighted:1 mit:1 sensor:4 gaussian:2 rather:1 kortenkamp:1 mobile:5 command:1 longest:1 contrast:1 inference:7 hidden:2 koller:2 going:1 classification:1 orientation:2 priori:1 smoothing:3 marginal:1 field:1 sampling:2 identical:1 represents:1 stuart:1 look:1 unsupervised:1 future:1 escape:1 tightly:1 murphy:5 familiar:1 mixture:1 navigation:1 nl:5 pc:5 light:3 accurate:2 fox:3 tree:1 circle:1 re:1 uncertain:1 rao:3 cost:1 tractability:1 kaelbling:1 entry:1 uniform:1 burgard:3 successful:2 examining:1 combined:1 confident:1 probabilistic:2 off:4 yl:2 na:1 aaai:1 marginalizes:1 dead:2 inefficient:1 return:3 multiplexer:2 account:1 tii:1 de:1 imt:1 matter:2 depends:2 later:2 try:3 closed:2 reached:1 start:2 bayes:2 il:1 ir:1 variance:2 reckoning:2 climbing:1 generalize:1 weak:1 bayesian:15 territory:1 accurately:1 carlo:4 trajectory:4 submitted:1 explain:1 sharing:1 sebastian:1 obvious:1 mi:5 static:2 popular:1 knowledge:1 ut:2 actually:1 back:2 higher:1 done:1 though:1 furthermore:1 just:5 correlation:2 until:1 hand:2 koenig:1 o:1 mode:3 lei:1 grows:1 effect:1 contain:1 true:1 normalized:1 andrieu:1 analytically:1 hence:9 illustrated:1 white:1 conditionally:2 ll:2 oc:1 occasion:1 hill:1 complete:1 variational:1 common:1 mt:13 conditioning:1 exponentially:1 approximates:1 marginals:3 versa:1 ai:1 grid:9 particle:19 moving:1 robot:36 entail:1 etc:1 posterior:7 recent:4 store:1 certain:1 n00014:1 binary:1 onr:1 seen:1 minimum:2 somewhat:1 maximize:1 ii:1 sliding:1 multiple:1 desirable:1 full:1 visit:1 basic:1 essentially:1 iteration:1 represent:1 robotics:1 cell:28 proposal:3 addition:1 condensation:1 separately:1 whereas:1 ot:3 ample:1 jordan:1 door:3 backwards:1 easy:1 independence:1 topology:1 idea:1 passed:1 dellaert:1 cause:2 action:5 useful:1 factorial:2 specifies:1 ilcai:1 bonasso:1 notice:1 dotted:1 delta:1 estimated:2 dummy:1 per:1 track:1 discrete:2 group:1 key:1 four:1 nevertheless:1 localize:3 compete:1 ilt:1 journey:1 almost:2 home:1 dy:1 bit:2 topological:1 precisely:1 bp:1 encodes:1 diffuse:1 performing:1 downweight:1 alternate:1 remain:1 em:17 contradicts:1 making:1 projecting:1 pr:11 computationally:2 equation:1 remains:1 discus:4 needed:1 know:4 flip:1 tractable:2 end:2 junction:1 operation:1 actuator:2 away:1 blurry:1 alternative:1 gate:1 running:3 ensure:1 graphical:4 maintaining:1 marginalized:2 exploit:1 ghahramani:1 approximating:1 icra:1 move:4 already:2 question:1 strategy:1 visiting:2 exhibit:1 unable:1 thank:1 thrun:4 hmm:3 me:2 length:4 kalman:2 useless:1 difficult:1 unfortunately:4 sharper:1 godsill:1 zt:15 lil:1 unknown:3 perform:2 policy:1 observation:10 markov:3 sm:1 looking:1 head:1 lb:3 drift:1 bk:3 specified:1 california:1 learned:2 nip:1 able:3 bar:1 below:2 pattern:3 boyen:2 explanation:1 belief:2 treated:1 difficulty:1 residual:1 occupancy:1 improve:1 axis:1 coupled:4 slippery:1 prior:7 review:1 literature:1 marginalizing:1 relative:1 fully:1 interesting:1 filtering:9 jasa:1 sufficient:2 editor:1 lo:1 row:2 course:2 supported:1 free:1 offline:1 normalised:1 benefit:1 boundary:1 dimension:3 transition:5 world:3 ignores:2 stuck:1 collection:1 forward:1 jump:1 made:1 qualitatively:1 far:1 cope:1 approximate:4 ignore:1 unreliable:1 keep:3 global:1 active:3 uai:1 doucet:1 assumed:1 channel:1 learn:4 robust:1 ca:1 nature:1 ignoring:1 complex:1 did:1 mnl:2 linearly:1 noise:4 child:1 allowed:1 n:2 position:2 exponential:3 decay:2 virtue:1 evidence:2 survival:1 intractable:1 adding:1 sequential:3 importance:1 chen:2 entropy:1 lt:11 appearance:1 likely:4 explore:1 iik:1 corresponds:1 shatkay:1 conditional:3 replace:1 change:1 determined:1 except:1 wt:2 averaging:1 called:3 pas:1 indicating:1 formally:1 correlated:2
783
1,717
An Improved Decomposition Algorithm for Regression Support Vector Machines Pavel Laskov Department of Computer and Information Sciences University of Delaware Newark, DE 19718 laskov@asel. udel. edu Abstract A new decomposition algorithm for training regression Support Vector Machines (SVM) is presented. The algorithm builds on the basic principles of decomposition proposed by Osuna et. al ., and addresses the issue of optimal working set selection. The new criteria for testing optimality of a working set are derived. Based on these criteria, the principle of "maximal inconsistency" is proposed to form (approximately) optimal working sets. Experimental results show superior performance of the new algorithm in comparison with traditional training of regression SVM without decomposition. Similar results have been previously reported on decomposition algorithms for pattern recognition SVM. The new algorithm is also applicable to advanced SVM formulations based on regression, such as density estimation and integral equation SVM. 1 Introd uction The increasing interest in applications of Support Vector Machines (SVM) to largescale problems ushers in new requirements for computational complexity of their training algorithms. Requests have been recently made for algorithms capable of handling problems containing 105 - 106 examples [1]. Training an SVM constitutes a quadratic programming problem, and a typical SVM package uses an off-the-shelf optimization software to obtain a solution to it. The number of variables in the optimization problem is equal to the number of training data points (for the pattern recognition SVM) or twice that number (for the regression SVM). The speed of general-purpose optimization methods is insufficient for problems containing more than a few thousand examples. This has motivated a quest for special-purpose training algorithms to take advantage of the particular structure of SVM training problems. The main avenue of research in SVM training algorithms is decomposition. The key idea of decomposition, due to Osuna et. al. [2], is to freeze all but a small number of optimization variables, and to solve a sequence of small fixed-size problems. The set of variables whose values are optimized at a current iteration is called the working set. Complexity of re-optimizing the working set is assumed to be constant-time. An Improved Decomposition Algorithm for Regression Support Vector Machines 485 In order for a decomposition algorithm to be successful) the working set must be selected in a smart way. The fastest known decomposition algorithm is due to Joachims [3]. It is based on Zoutendijk)s method of feasible directions proposed in the optimization community in the early 1960)s. However Joachims) algorithm is limited to pattern recognition SVM because it makes use of labels being ?l. The current article presents a similar algorithm for the regression SVM. The new algorithm utilizes a slightly different background from optimization theory. The Karush-Kuhn-Tucker Theorem is used to derive conditions for determining whether or not a given working set is optimal. These conditions become the algorithm)s termination criteria) as an alternative to Osuna)s criteria (also used by Joachims without modification) which used conditions for individual points. The advantage of the new conditions is that knowledge of the hyperplane)s constant factor b) which in some cases is difficult to compute) is not required. Further investigation of the new termination conditions allows to form the strategy for selecting an optimal working set. The new algorithm is applicable to the pattern recognition SVM) and is provably equivalent to Joachims) algorithm. One can also interpret the new algorithm in the sense of the method of feasible directions. Experimental results presented in the last section demonstrate superior performance of the new method in comparison with traditional training of regression SVM. 2 General Principles of Regression SVM Decomposition The original decomposition algorithm proposed for the pattern recognition SVM in [2] has been extended to the regression SVM in [4]. For the sake of completeness I will repeat the main steps of this extension with the aim of providing terse and streamlined notation to lay the ground for working set selection. Given the training data of size I) training of the regression SVM amounts to solving the following quadratic programming problem in 21 variables: Maximize subject to: W(a) 1y-T0: - -0: TD 0: 2 eTa 0 a-Ct < 0 a > 0 (1) where K -K The basic idea of decomposition is to split the variable vector a into the working set aB of fixed size q and the non-working set aN containing the rest ofthe variables. The corresponding parts of vectors e and y will also bear subscripts Nand B . The matrix D is partitioned into D BB ) DBN = D~B and D NN . A further requirement is that) for the i-th element of the training data) both 0i and 0; are either included in or omitted from the working set.l The values of the variables in the non-working set are frozen for the iteration) and optimization is only performed with respect to the variables in the working set. Optimization of the working set is also a quadratic program. This can be seen by re-arranging the terms of the objective function and the equality constraint in IThis rule facilitates formulation of sub-problems to be solved at each iteration. P. Laskov 486 (1) and dropping the terms independent of o.B from the objective. The resulting quadratic program (sub-problem) is formulated as follows: -T Maximize subject to: -T - (YB - QNDNB)QB - - a T TCBQB +cNQN o.B - 1 -T 'iQBDBBQB (2) < 0 > 0 Cl The basic decomposition algorithm chooses the first working set at random, and proceeds iteratively by selecting sub-optimal working sets and re-optimizing them, by solving quadratic program (2), until all subsets of size q are optimal. The precise formulation of termination conditions will be developed in the following section. 3 Optimality of a Working Set In order to maintain strict improvement of the objective function, the working set must be sub-optimal before re-optimization. The classical Karush-Kuhn-Tucker (KKT) conditions are necessary and sufficient for optimality of a quadratic program. I will use these conditions applied to the standard form of a quadratic program, as described in [5], p. 36. The standard form of a quadratic program requires that all constraints are of equality type except for non-negativity constraints. To cast the refression SVM quadratic program (1) into the standard form, the slack variables s = (81 , '" ,82l) corresponding to the box constraints, and the following matrices are introduced: I 0]I ' o E= [f], z= m' (3) where 1 is a vector of length I , C is a vector of length 21. The zero element in vector z reflects the fact that a slack variable for the equality constraint must be zero. In the matrix notation all constraints of problem (1) can be compactly expressed as: ETz f z > 0 (4) In this notation the Karush-Kuhn-Tucker Theorem can be stated as follows: Theorem 1 (Karush-Kuhn-Tucker Theorem) The primal vector z solves the quadratic problem (1) if and only if it satisfies (4) and there exists a dual vector u T = (ITT w T ) = (ITT (I' yT? such that: IT = DO. + Ew Y ~ 0 uT z =a Y> 0 (5) (6) (7) It follows from the Karush-Kuhn-Tucker Theorem that if for all u satisfying conditions (6) - (7) the system of inequalities (5) is inconsistent then the solution of problem (1) is not optimal. Since the objective function of sub-problem (2) was obtained by merely re-arranging terms in the objective function of the initial problem (1), the same conditions guarantee that the sub-problem (2) is not optimal. Thus, the main strategy for identifying sub-optimal working sets will be to enforce inconsistency of the system (5) while satisfying conditions (6) - (7) . 487 An Improved Decomposition Algorithm for Regression Support Vector Machines Let us further analyze inequalities in (5). Each inequality has one of the following forms: 7T'i -rPi + E + Vi + J.L > 0 rPi + E - v; - J.L > 0 (8) (9) where I rPi = Yi - 2:)aj - a;)Kij j=l Consider the values ai can possible take: 1. ai = O. In this case Si = C, and, by complementarity condition (7), Vi Then inequality (8) becomes: 7T'i = -rPi + E + J.L ~ 0 ~ J.L ~ = C. By complementarity condition (7), 7T'i becomes: 2. ai -rPi + E + J.L + Vi ~ = 0 rPi - = O. J.L::; = O. E Then inequality (8) rPi - E < ai < C. By complementarity condition (7), Vi = 0, 7T'i = O. Then inequality (8) becomes: 3. 0 - rPi + E + J.L Similar reasoning for 0 ~ J.L = rPi - E a; and inequality (9) yields the following results: 1. a; = O. Then 2. a; = C. 3. 0 < = Then a; < C . Then As one can see, the only free variable in system (5) is J.L. Each inequality restricts J.L to a certain interval on a real line. Such intervals will be denoted as J.L-sets in the rest of the exposition. Any subset of inequalities in (5) is inconsistent if the intersection of the corresponding J.L-sets is empty. This provides a lucid rule for determining optimality of any working set: it is sub-optimal if the intersection of J.L-sets of all its points is empty. A sub-optimal working set will also be denoted as "inconsistent". The following summarizes the rules for calculation of J.L-sets, taking into account that for regression SVM aia; = 0: [rPi - E, rPi + E], if ai = 0, a; = 0 [rPi - E, rPi - E], if 0 < ai < C, a; = 0 (-00, rPi - E], if ai = C, a; = 0 [rPi + E, rPi + E], if ai = 0, 0 < a; < C [rPi + E, +(0), (10) 488 4 P. Laskov Maximal Inconsistency Algorithm While inconsistency of the working set at each iteration guarantees convergence of decomposition, the rate of convergence is quite slow if arbitrary inconsistent working sets are chosen. A natural heuristic is to select "maximally inconsistent" working sets, in a hope that such choice would provide the greatest improvement of the objective function. The notion of "maximal inconsistency" is easy to define: let it be the gap between the smallest right boundary and the largest left boundary of p-sets of elements in the training set: G=L-R L = O<i<l max pL R = O<i<l min pr 1 where p!, pi are the left and the right boundaries respectively (possibly minus or plus infinity) of the p-set Mi. It is convenient to require that the largest possible inconsistency gap be maintained between all pairs of points comprising the working set. The obvious implementation of such strategy is to select q/2 elements with the largest values of pi and q/2 elements with the smallest values of pr. The maximal inconsistency strategy is summarized in Algorithm 1. Algorithm 1 Maximal inconsistency SVM decomposition algorithm. Let S be the list of all samples. while (L > R) ? compute Mi according to the rules (10) for all elements in S ? select q/2 elements with the largest values of pi ("left pass") ? select q /2 elements with the smallest values of p r ("right pass") ? re-optimize the working set Although the motivation provided for the maximal inconsistency algorithm is purely heuristic, the algorithm can be rigorously derived, in a similar fashion as Joachims' algorithm, from Zoutendijk's feasible direction problem. Details of such derivation cannot be presented here due to space constraints. Because of this relationship I will further refer to both algorithms as "feasible direction" algorithms. 5 Experimental Results Experimental evaluation of the new algorithm was performed on the modThe original data set is available under ified KDD Cup 1998 data set. http:j/www.ics.uci.edu/"-'kdd/databases/kddcup98/kddcup98.html. The following modifications were made to obtain a pure regression problem: ? All 75 character fields were eliminated. ? Numeric fields CONTROLN, ODATEDW, TCODE and DOB were elimitated. The remaining 400 features and the labels were scaled between 0 and 1. Initial subsets of the training database of different sizes were selected for evaluation of the scaling properties of the new algorithm. The training times of the algorithms, with and without decomposition, the numbers of support vectors, including bounded support vectors, and the experimental scaling factors, are displayed in Table 1. An Improved Decomposition Algorithm for Regression Support Vector Machines Table 1: Training time (sec) and number Examples no dcmp 500 39 226 1000 2000 1490 3000 5744 27052 5000 scaling factor: 2.84 SV-scaling factor: 3.06 489 of SVs for the KDD Cup problem dcmp total SV BSV 10 274 0 41 518 3 158 970 5 1429 397 7 1252 2349 15 2.08 2.24 Table 2: Training time (sec) and number of SVs for the KDD Cup problem, reduced feature space. Examples no dcmp dcmp total SV BSV 500 170 18 56 30 346 1000 62 44 374 2000 1768 198 144 510 4789 729 366 3000 222 22115 1139 5000 863 354 scaling factor: 2.55 1.72 SV-scaling factor: 3.55 2.35 The experimental scaling factors are obtained by fitting lines to log-log plots of the running times against sample sizes, in the number of examples and the number of unbounded support vectors respectively. Experiments were run on SGI Octane with 195MHz clock and 256M RAM. RBF kernel with, = 10, C = 1, termination accuracy 0.001, working set size of 20, and cache size of 5000 samples were used. A similar experiment was performed on a reduced feature set consisting of the first 50 features selected from the full-size data set. This experiment illustrates the behavior of the algorithms when the large number of support vectors are bounded. The results are presented in Table 2. 6 Discussion It comes at no surprise that the decomposition algorithm outperforms the conventional training algorithm by an order of magnitude. Similar results have been well established for pattern recognition SVM. Remarkable is the co-incidence of scaling factors of the maximal inconsistency algorithm and Joachims' algorithm: his scaling factors range from 1.7 to 2.1 [3]. I believe however, that a more important performance measure is SV -scaling factor, and the results above suggest that this factor is consistent even for problems with significantly different compositions of support vectors. Further experiments should investigate properties of this measure. Finally, I would like to mention other methods proposed in order to speed-up training of SVM, although no experimental results have been reported for these methods with regard to training of the regression SVM. Chunking [6], p. 366, iterates through 490 P. Laskov the training data accumulating support vectors and adding a "chunk" of new data until no more changes to a solution occur. The main problem with this method is that when the percentage of support vectors is high it essentially solves the problem of almost the same size more than once. Sequential Minimal Optimization (SMO), proposed by Platt [7] and easily extendable to the regression SVM [1], employs an idea similar to decomposition but always uses the working set of size 2. For such a working set, a solution can be calculated "by hand" without numerical optimization. A number of heuristics is applied in order to choose a good working set. It is difficult to draw a comparison between the working set selection mechanisms of SMO and the feasible direction algorithms but experimental results of Joachims [3] suggest that SMO is slower. Another advantage of feasible direction algorithms is that the size of the working set is not limited to 2, as in SMO. Practical experience shows that the optimal size of the working set is between 10 and 100. Lastly, traditional optimization methods, such as Newton's or conjugate gradient methods, can be modified to yield the complexity of 0(s3), where s is the number of detected support vectors [8]. This can be a considerable improvement over the methods that have complexity of 0(13), where 1 is the total number of training samples. The real challenge lies in attaining sub-0(s3) complexity. While the experimental results suggest that feasible direction algorithms might attain such complexity, their complexity is not fully understood from the theoretical point of view. More specifically, the convergence rate, and its dependence on the number of support vectors, needs to be analyzed. This will be the main direction of the future research in feasible direction SVM training algorithms. References [1] Smola, A., Sch61kopf, B. (1998) A Tutorial on Support Vector Regression. NeuroCOLT2 Technical Report NC2- TR-1998-030. [2] Osuna, E., Freund, R., Girosi, F. (1997) An Improved Training Algorithm for Support Vector Machines. Proceedings of IEEE NNSP'97. Amelia Island FL. [3] Joachims, T. (1998) Making Large-Scale SVM Learning Practical. Advances in Kernel Methods - Support Vector Learning. B. Sch61kopf, C. Burges, A. Smola, (eds.) MIT-Press. [4] Osuna, E. (1998) Support Vector Machines: Training and Applications. Ph. D. Dissertation. Operations Research Center, MIT. [5] Boot, J. (1964) Quadratic Programming. Algorithms - Anomalies - Applications. North Holland Publishing Company, Amsterdam. [6] Vapnik, V. (1982) Estimation of Dependencies Based on Empirical Data. Springer-Verlag. [7] Platt, J. (1998) Fast Training of Support Vector Machines Using Sequential Minimal Optimization. Advances in Kernel Methods - Support Vector Learning. B. Sch5lkopf, C. Burges, A. Smola, (eds.) MIT-Press. [8] Kaufman, L. (1998) Solving the Quadratic Programming Problem Arising in Support Vector Classification. Advances in Kernel Methods - Support Vector Learning. B. Sch5lkopf, C. Burges, A. Smola, (eds.) MIT-Press.
1717 |@word termination:4 decomposition:21 pavel:1 mention:1 minus:1 tr:1 initial:2 selecting:2 outperforms:1 current:2 incidence:1 rpi:17 si:1 must:3 numerical:1 kdd:4 girosi:1 plot:1 selected:3 nnsp:1 dissertation:1 completeness:1 provides:1 iterates:1 unbounded:1 become:1 fitting:1 behavior:1 td:1 company:1 cache:1 increasing:1 becomes:3 provided:1 notation:3 bounded:2 etz:1 kaufman:1 developed:1 guarantee:2 scaled:1 platt:2 before:1 understood:1 subscript:1 approximately:1 might:1 plus:1 twice:1 co:1 fastest:1 limited:2 range:1 practical:2 testing:1 sch61kopf:2 empirical:1 significantly:1 attain:1 convenient:1 suggest:3 cannot:1 selection:3 accumulating:1 www:1 optimize:1 equivalent:1 conventional:1 yt:1 center:1 identifying:1 pure:1 rule:4 his:1 notion:1 arranging:2 anomaly:1 programming:4 us:2 complementarity:3 element:8 recognition:6 satisfying:2 lay:1 database:2 solved:1 thousand:1 complexity:7 rigorously:1 solving:3 smart:1 purely:1 compactly:1 easily:1 derivation:1 fast:1 detected:1 whose:1 quite:1 heuristic:3 solve:1 terse:1 advantage:3 sequence:1 frozen:1 maximal:7 udel:1 uci:1 sgi:1 convergence:3 empty:2 requirement:2 derive:1 solves:2 come:1 direction:9 kuhn:5 require:1 karush:5 investigation:1 extension:1 pl:1 ground:1 ic:1 early:1 smallest:3 omitted:1 purpose:2 estimation:2 applicable:2 label:2 largest:4 reflects:1 hope:1 mit:4 always:1 aim:1 modified:1 shelf:1 derived:2 joachim:8 improvement:3 sense:1 nn:1 nand:1 comprising:1 provably:1 issue:1 dual:1 html:1 classification:1 denoted:2 special:1 equal:1 field:2 once:1 eliminated:1 constitutes:1 future:1 report:1 few:1 employ:1 individual:1 consisting:1 maintain:1 ab:1 interest:1 investigate:1 evaluation:2 analyzed:1 primal:1 integral:1 capable:1 necessary:1 experience:1 re:6 theoretical:1 minimal:2 lucid:1 kij:1 eta:1 mhz:1 subset:3 successful:1 reported:2 dependency:1 sv:5 chooses:1 chunk:1 extendable:1 density:1 off:1 containing:3 choose:1 possibly:1 account:1 de:1 attaining:1 summarized:1 sec:2 north:1 vi:4 performed:3 view:1 analyze:1 nc2:1 accuracy:1 yield:2 ofthe:1 ed:3 streamlined:1 against:1 tucker:5 obvious:1 mi:2 knowledge:1 ut:1 improved:5 maximally:1 yb:1 formulation:3 box:1 smola:4 lastly:1 until:2 clock:1 working:34 hand:1 aj:1 believe:1 equality:3 iteratively:1 maintained:1 criterion:4 demonstrate:1 reasoning:1 recently:1 superior:2 interpret:1 refer:1 composition:1 freeze:1 cup:3 ai:8 dbn:1 optimizing:2 certain:1 verlag:1 inequality:9 inconsistency:10 yi:1 seen:1 maximize:2 full:1 technical:1 calculation:1 regression:18 basic:3 essentially:1 iteration:4 kernel:4 background:1 interval:2 rest:2 usher:1 strict:1 subject:2 facilitates:1 inconsistent:5 split:1 easy:1 idea:3 delaware:1 avenue:1 t0:1 whether:1 motivated:1 introd:1 svs:2 amount:1 ph:1 reduced:2 http:1 percentage:1 restricts:1 tutorial:1 s3:2 arising:1 dropping:1 key:1 ram:1 merely:1 run:1 package:1 almost:1 utilizes:1 draw:1 summarizes:1 scaling:10 fl:1 ct:1 laskov:5 quadratic:12 occur:1 constraint:7 infinity:1 software:1 sake:1 speed:2 optimality:4 min:1 qb:1 department:1 according:1 request:1 conjugate:1 slightly:1 character:1 osuna:5 partitioned:1 island:1 modification:2 making:1 pr:2 chunking:1 equation:1 previously:1 slack:2 mechanism:1 available:1 operation:1 enforce:1 alternative:1 aia:1 slower:1 original:2 remaining:1 running:1 publishing:1 newton:1 build:1 classical:1 objective:6 neurocolt2:1 strategy:4 dependence:1 traditional:3 gradient:1 length:2 relationship:1 insufficient:1 providing:1 ified:1 difficult:2 stated:1 implementation:1 boot:1 displayed:1 extended:1 precise:1 arbitrary:1 community:1 introduced:1 cast:1 required:1 pair:1 amelia:1 optimized:1 smo:4 established:1 address:1 proceeds:1 pattern:6 challenge:1 program:7 max:1 including:1 greatest:1 natural:1 largescale:1 advanced:1 sch5lkopf:2 negativity:1 determining:2 freund:1 fully:1 bear:1 remarkable:1 sufficient:1 consistent:1 article:1 principle:3 pi:3 repeat:1 last:1 free:1 burges:3 taking:1 regard:1 boundary:3 calculated:1 numeric:1 made:2 bb:1 kkt:1 assumed:1 table:4 itt:2 cl:1 main:5 motivation:1 fashion:1 slow:1 sub:10 lie:1 theorem:5 list:1 svm:29 exists:1 uction:1 vapnik:1 adding:1 sequential:2 magnitude:1 illustrates:1 gap:2 surprise:1 intersection:2 expressed:1 amsterdam:1 holland:1 springer:1 satisfies:1 formulated:1 exposition:1 rbf:1 feasible:8 change:1 considerable:1 included:1 typical:1 except:1 specifically:1 hyperplane:1 ithis:1 called:1 total:3 pas:2 experimental:9 ew:1 select:4 support:23 quest:1 newark:1 handling:1
784
1,718
A Multi-class Linear Learning Algorithm Related to Winnow Chris Mesterhann* Rutgers Computer Science Department 110 Frelinghuysen Road Piscataway, NJ 08854 mesterha@paul.rutgers.edu Abstract In this paper, we present Committee, a new multi-class learning algorithm related to the Winnow family of algorithms. Committee is an algorithm for combining the predictions of a set of sub-experts in the online mistake-bounded model oflearning. A sub-expert is a special type of attribute that predicts with a distribution over a finite number of classes. Committee learns a linear function of sub-experts and uses this function to make class predictions. We provide bounds for Committee that show it performs well when the target can be represented by a few relevant sub-experts. We also show how Committee can be used to solve more traditional problems composed of attributes. This leads to a natural extension that learns on multi-class problems that contain both traditional attributes and sub-experts. 1 Introduction In this paper, we present a new multi-class learning algorithm called Committee. Committee learns a k class target function by combining information from a large set of sub-experts. A sub-expert is a special type of attribute that predicts with a distribution over the target classes. The target space of functions are linear-max functions. We define these as functions that take a linear combination of sub-expert predictions and return the class with maximum value. It may be useful to think of the sub-experts as individual classifying functions that are attempting to predict the target function. Even though the individual sub-experts may not be perfect, Committee attempts to learn a linear-max function that represents the target function. In truth, this picture is not quite accurate. The reason we call them sub-experts and not experts is because even though a individual sub-expert might be poor at prediction, it may be useful when used in a linear-max function. For example, some sub-experts might be used to add constant weights to the linear-max function. The algorithm is analyzed for the on-line mistake-bounded model oflearning [Lit89]. This is a useful model for a type of incremental learning where an algorithm can use feedback about its current hypothesis to improve its performance. In this model, the algorithm goes through a series of learning trials. A trial is composed of three steps. First, the algorithm ?Part of this work was supported by NEe Research Institute, Princeton, NJ. C. Mesterharm 520 receives an instance, in this case, the predictions of the sub-experts. Second, the algorithm predicts a label for the instance; this is the global prediction of Committee. And last, the algorithm receives the true label of the instance ; Committee uses this information to update its estimate of the target. The goal of the algorithm is to minimize the total number of prediction mistakes the algorithm makes while learning the target. The analysis and performance of Committee is similar to another learning algorithm, Winnow [Lit89] . Winnow is an algorithm for learning a linear-threshold function that maps attributes in [0, 1] to a binary target. It is an algorithm that is effective when the concept can be represented with a few relevant attributes, irrespective of the behavior of the other attributes. Committee is similar but deals with learning a target that contains only a few relevant sub-experts. While learning with sub-experts is interesting in it's own right, it turns out the distinction between the two tasks is not significant. We will show in section 5 how to transform attributes from [0, 1] into sub-experts. Using particular transformations, Committee is identical to the Winnow algorithms, Balanced and WMA [Lit89]. Furthermore, we can generalize these transformations to handle attribute problems with multi-class targets. These transformations naturally lead to a hybrid algorithm that allows a combination of sub-experts and attributes for multi-class learning problems. This opens up a range of new practical problems that did not easily fit into the previous framework of [0, 1J attributes and binary classification. 2 Previous work Many people have successfully tried the Winnow algorithms on real-world tasks. In the course of their work, they have made modifications to the algorithms to fit certain aspects of their problem. These modifications include multi-class extensions. For example, [DKR97] use Winnow algorithms on text classification problems. This multiclass problem has a special form ; a document can belong to more than one class. Because of this property, it makes sense to learn a different binary classifier for each class. The linear functions are allowed, even desired, to overlap. However, this paper is concerned with cases where this is not possible. For example, in [GR96] the correct spelling of a word must be selected from a set of many possibilities. In this setting, it is more desirable to have the algorithm select a single word. The work in [GR96] presents many interesting ideas and modifications of the Winnow algorithms. At a minimum, these modification are useful for improving the performance of Winnow on those particular problems. Part of that work also extends the Winnow algorithm to general multi-class problems. While the results are favorable, the contribution ofthis paper is to give a different algorithm that has a stronger theoretical foundation for customizing a particular multi-class problem. Blum also works with multi-class Winnow algorithms on the calendar scheduling problem of [MCF+94] . In [Blu95], a modified Winnow is given with theoretical arguments for good performance on certain types of multi-class disjunctions. In this paper, these results are extended, with the new algorithm Committee, to cover a wider range of multi-class linear functions. Other related theoretical work on multi-class problems includes the regression algorithm EG ?. In [KW97], Kivinen and Warmuth introduce EG ?, an algorithm related to Winnow but used on regression problems. In general, while regression is a useful framework for many multi-class problems, it is not straightforward how to extend regression to the concepts learned by Committee. A particular problem is the inability of current regression techniques to handle 0-1 loss. A Multi-class Linear Learning Algorithm Related to Winnow 521 3 Algorithm This section of the paper describes the details of Committee. Near the end of the section, we will give a formal statement of the algorithm. 3.1 Prediction scheme Assume there are n sub-experts. Each sub-expert has a positive weight that is used to vote for k different classes; let Wi be the weight of sub-expert i. A sub-expert can vote for several classes by spreading its weight with a prediction distribution. For example, if k = 3, a sub-expert may give 3/5 of its weight to class 1, 1/5 of its weight to class 2, and 1/5 of its weight to class 3. Let Xi represent this prediction distribution, where x~ is the fraction of the weight sub-expert i gives to class j . The vote for class j is L~=I WiX~. Committee predicts the class that has the highest vote. (On ties, the algorithm picks one of the classes involved in the tie.) We call the function computed by this prediction scheme a linear-max function, since it is the maximum class value taken from a linear combination of the SUb-expert predictions. 3.2 Target function The goal of Committee is to mInimIZe the number of mistakes by quickly learning sub-expert weights that correctly classify the target function. Assume there exists fL, a vector of nonnegative weights that correctly classifies the target. Notice that fL can be multiplied by any constant without changing the target. To remove this confusion, we will normalize the weights to sum to 1, i.e., L~=Il-1i = 1. Let ((j) be the target's vote for class j. n ((j) = L l-1iX~ t=I Part of the difficulty of the learning problem is hidden in the target weights. Intuitively, a target function will be more difficult to learn ifthere is a small difference between the (votes of the correct and incorrect classes. We measure this difficulty by looking at the minimum difference, over all trials, of the vote of the correct label and the vote of the other labels. Assume for trial t that Pt is the correct label. 8= min tETnals (min(((pt)-((j))) rlpt Because these are the weights of the target, and the target always makes the correct prediction, 8 > o. One problem with the above assumptions is that they do not allow noise (cases where 8 <::; 0). However, there are variations of the analysis that allow for limited amounts of noise [Lit89, Lit91]. Also experimental work [Lit95, LM] shows the family of Winnow algorithms to be much more robust to noise than the theory would predict. Based on the similarity of the algorithm and analysis, and some preliminary experiments, Committee should be able to tolerate some noise. 3.3 Updates Committee only updates on mistakes using multiplicative updates. The algorithm starts by initializing all weights to 1 In. During the trials, let P be the correct label and .x be the predicted label of Committee. When .x =1= P the weight of each sub-expert i is multiplied by aX; -x;. This corresponds to increasing the weights of the sub-experts who predicted the C. Mesterharm 522 correct label instead of the label Committee predicted. The value of 0' is initialized at the start of the algorithm. The optimal value of 0' for the bounds depends on 6. Often 6 is not known in advance, but experiments on Winnow algorithms suggest that these algorithms are more flexible, often performing well with a wider range of 0' values [LM). Last, the weights are renormalize to sum to 1. While this is not strictly necessary, normalizing has several advantages including reducing the likelyhood of underflow/overflow errors. 3.4 Committee code Initialization 'Vi E {l , . .. , n} W i:= lin. Set 0' > 1. Trials Instance sub-experts (Xl , . .. , xn) . Prediction >. is the first class c such that for all other classes J, "n c L...- i =1 W i X i > - "n j L...-i=1 W i X t ? Update Let p be the correct label. If mistake (>' fori:=l ton Wi := p O'X i >. -x, Wi . Normalize weights, L:~= l 3.5 # p) Wt =1 Mistake bound We do not have the space to give the proof for the mistake bound of Committee, but the technique is similar to the proof of the Winnow algorithm, Balanced, given in [Lit89). For the complete proof, the reader can refer to [Mes99). Theorem 1 Committee makes at most 2ln (n) 162mistakes when the target conditions in section 3.2 are satisfied and 0' is set to (1 - 6) - 1/ 2. Surprisingly, this bound does not refer to the number of classes. The effects of larger values of k show up indirectly in the 6 value. While it is not obvious, this bound shows that Committee performs well when the target can be represented by a small fraction of the sub-experts. Call the sub-experts in the target the relevant sub-experts. Since 6 is a function of the target, 6 only depends on the relevant sub-experts. On the other hand, the remaining sub-experts have a small effect on the bound since they are only represented in the In( n) factor. This means that the mistake bound of Committee is fairly stable even when adding a large number of additional sub-experts. In truth, this doesn ' t mean that the algorithm will have a good bound when there are few relevant sub-experts. In some cases, a small number of sub-experts can give an arbitrarily small 6 value. (This is a general problem with all the Winnow algorithms.) What it does mean is that, given any problem, increasing the number of irrelevant sub-experts will only have a logarithmic effect on the mistake bound. 4 Attributes to sub-experts Often there are no obvious sub-experts to use in solving a learning problem. Many times the only information available is a set of attributes. For attributes in [0,1]' we will show how to use Committee to learn a natural kind of k class target function , a linear machine. To learn this target, we will transform each attribute into k separate sub-experts. We will use some of the same notion as Committee to help understand the transformation. 523 A Multi-class Linear Learning Algorithm Related to Winnow 4.1 Attribute target (linear machine) A linear machine [DH73] is a prediction function that divides the feature space into disjoint convex regions where each class corresponds to one region. The predictions are made by a comparing the value of k different linear functions where each function corresponds to a class. More formally, assume there are m - 1 attributes and k classes. Let Zi E [0,1] be attribute i. Assume the target function is represented using k linear functions of the attributes. Let ((j) = L::II-L{ Zi be the linear function for class j where I-Li is the weight of attribute i in class j. Notice that we have added one extra attribute. This attribute is set to 1 and is needed for the constant portion of the linear functions. The target function labels an instance with the class of the largest ( function. (Ties are not defined.) Therefore, ((j) is similar to the voting function for class j used in Committee. 4.2 Transforming the target One difficulty with these linear functions is that they may have negative weights. Since Committee only allows targets with nonnegative weights, we need transform to an equivalent problem that has nonnegative weights. This is not difficult. Since we are only concerned with the relative difference between the ( functions, we are allowed to add any function to the (functions as long as we add it to all (functions. This gives us a simple procedure to remove negative weights. For example, if ((1) = 3Z1 - 2Z2 + 1z3 -4, we can add 2Z2 +4 to every ( function to remove the negative weights from ((1). It is straightforward to extend this and remove all negative weights. We also need to normalize the weights. Again, since only the relative difference between the ( functions matter, we can divide all the ( functions by any constant. We normalize the weights to sum to 1, i.e., L:~=1L:~11-L{ = 1. At this point, without loss of generality, assume that the original ( functions are nonnegative and normalized. The last step is to identify a 8 value. We use the same definition of 8 as Committee substituting the corresponding ( functions of the linear machine. Assume for trial t that Pt is the correct label. 8= 4.3 min tETrwls (min( ((Pt) - ((j))) ji-P, Transforming the attributes The transformation works as follows: convert attribute Zi into k sub-experts. Each sub-expert will always vote for one of the k classes with value Zi. The target weight for each of these sub-experts is the corresponding target weight of the attribute, label pair in the ( functions . Do this for every attribute. Notice that we are not using distributions for the sub-expert predictions. A sub-expert's prediction can be converted to a distribution by adding a constant amount to each class prediction. For example, a sub-expert that predicts Zl = .7, Z2 = 0, Z3 = can be changed to Zl = .8, Z2 = .1, Z3 = .1 by adding .1 to each class. This conversion does not affect the predicting or updating of Committee. ? 524 C. Mesterharm Theorem 2 Committee makes at most 2In(mk)/8 2 mistakes on a linear machine, as defined in this section, when 0 is set to (1 - 8)-1/2. Proof: The above target transformation creates mk normalized target sub-experts that vote with the same ( functions as the linear machine. Therefore, this set of sub-experts has the same 8 value. Plugging these values into the bound for Committee gives the result. This transformation provides a simple procedure for solving linear machine problems. While the details of the transformation may look cumbersome, the actual implementation of the algorithm is relatively simple. There is no need to explicitly keep track of the sub-experts. Instead, the algorithm can use a linear machine type representation. Each class keeps a vector of weights, one weight for each attribute. During an update, only the correct class weights and the predicted class weights are changed. The correct class weights are multiplied by O Zi; the predicted class weights are multiplied by o -z' . The above procedure is very similar to the Balanced algorithm from [Lit89] , in fact, for k = 2, it is identical. A similar transformation duplicates the behavior of the linear-threshold learning version ofWMA as given in [Lit89]. While this transformation shows some advantages for k = 2, more research is needed to determine the proper way to generalize to the multi-class case. For both of these transformations, the bounds given in this paper are equivalent (except for a superficial adjustment in the 8 notation of WMA) to the original bounds given in [Lit89] . 4.4 Combining attributes and sub-experts These transformations suggest the proper way to do a hybrid algorithm that combines sub-experts and attributes: use the transformations to create new sub-experts from the attributes and combine them with the original sub-experts when running Committee. It may even be desirable to break original sub-experts into attributes and use both in the algorithm because some sub-experts may perform better on certain classes. For example, if it is felt that a sub-expert is particularly good at class 1, we can perform the following transformation. Now, instead of using one weight for the whole sub-expert, Committee can also learn based on the sub-expert's performance for the first class. Even if a good target is representable only with the original sub-experts, these additional sub-experts will not have a large effect because of the logarithmic bound. In the same vein, it may be useful to add constant attributes to a set of sub-experts. These add only k extra SUb-experts, but allow the algorithm to represent a larger set of target functions . 5 Conclusion In this paper, we have introduced Committee, a multi-class learning algorithm. We feel that this algori thm will be important in practice, extending the range of problems that can be handled by the Winnow family of algorithms. With a solid theoretical foundation, researchers can customize Winnow algorithms to handle various multi-class problems. A Multi-class Linear Learning Algorithm Related to Winnow 525 Part of this customization includes feature transformations. We show how Committee can handle general linear machine problems by transforming attributes into sub-experts. This suggests a way to do a hybrid learning algorithm that allows a combination of sub-experts and attributes. This same techniques can also be used to add to the representational power on a standard sub-expert problem. In the future, we plan to empirically test Committee and the feature transformations on real world problems. Part of this testing will include modifying the algorithm to use extra information, that is related to the proof technique [Mes99), in an attempt to lower the number of mistakes. We speculate that adjusting the multiplier to increase the change in progress per trial will be useful for certain types of multi-class problems. Acknowledgments We thank Nick Littlestone for stimulating this work by suggesting techniques for converting the Balanced algorithm to multi-class targets. Also we thank Haym Hirsh, Nick Littlestone and Warren Smith for providing valuable comments and corrections. References [Blu95] Avrim Blum. Empirical support for winnow and weighted-majority algorithms: results on a calendar scheduling domain. In ML-95, pages 64-72, 1995. [DH73) R. O. Duda and P. Hart. Pattern Classification and Scene Analysis. Wiley, New York,1973 . [DKR97] I. Dagan, Y. Karov, and D. Roth. Mistake-driven learning in text categorization. In EMNLP-97, pages 55-63,1997. [GR96) A. R. Golding and D. Roth. Applying winnow to context-sensitive spelling correction . In ML-96, 1996. [KW97) Jyrki Kivinen and Manfred K. Warmuth. Additive versus exponentiated gradient updates for linear prediction. Information and Computation, 132(1): 1-64, 1997. [Lit89] Nick Littlestone. Mistake bounds and linear-threshold learning algorithms. PhD thesis, University of California, Santa Cruz, 1989. Technical Report UCSC-CRL-89-11. [Lit91) Nick Littlestone. Redundant noisy attributes, attribute errors, and linearthreshold learning using winnow. In COLT-91 , pages 147-156,1991. [Lit95] Nick Littlestone. Comparing several linear-threshold learning algorithms on tasks involving superfluous attributes. In ML-95, pages 353-361 , 1995. [LM) Nick Littlestone and Chris Mesterharm. A simulation study of winnow and related algorithms . Work in progress. [MCF+94) T. Mitchell, R. Caruana, D. Freitag, 1. McDermott, and D. Zabowski. Experience with a personal learning assistant. CACM, 37(7):81-91, 1994. [Mes99) Chris Mesterharm. A multi-class linear learning algorithm related to winnow with proof. Technical report, Rutgers University, 1999.
1718 |@word trial:8 version:1 stronger:1 duda:1 open:1 simulation:1 tried:1 pick:1 solid:1 series:1 contains:1 document:1 current:2 comparing:2 z2:4 must:1 cruz:1 additive:1 remove:4 update:7 selected:1 warmuth:2 smith:1 manfred:1 provides:1 ucsc:1 incorrect:1 freitag:1 combine:2 introduce:1 behavior:2 multi:23 actual:1 increasing:2 classifies:1 bounded:2 notation:1 what:1 kind:1 karov:1 transformation:16 nj:2 every:2 voting:1 tie:3 classifier:1 zl:2 positive:1 hirsh:1 mistake:15 might:2 initialization:1 suggests:1 limited:1 range:4 practical:1 acknowledgment:1 testing:1 practice:1 procedure:3 ifthere:1 empirical:1 word:2 road:1 suggest:2 scheduling:2 context:1 applying:1 equivalent:2 map:1 roth:2 go:1 straightforward:2 convex:1 handle:4 notion:1 variation:1 feel:1 target:38 pt:4 us:2 hypothesis:1 dkr97:2 particularly:1 updating:1 predicts:5 haym:1 vein:1 initializing:1 region:2 highest:1 valuable:1 balanced:4 transforming:3 personal:1 mcf:2 solving:2 creates:1 easily:1 represented:5 various:1 effective:1 cacm:1 disjunction:1 quite:1 larger:2 solve:1 calendar:2 think:1 transform:3 noisy:1 online:1 advantage:2 relevant:6 combining:3 representational:1 normalize:4 extending:1 categorization:1 perfect:1 incremental:1 wider:2 help:1 progress:2 predicted:5 correct:11 attribute:37 modifying:1 preliminary:1 extension:2 strictly:1 correction:2 predict:2 lm:3 substituting:1 favorable:1 assistant:1 label:13 spreading:1 sensitive:1 largest:1 create:1 successfully:1 weighted:1 always:2 modified:1 ax:1 sense:1 hidden:1 linearthreshold:1 classification:3 flexible:1 colt:1 plan:1 special:3 fairly:1 identical:2 represents:1 look:1 future:1 report:2 duplicate:1 few:4 composed:2 individual:3 attempt:2 possibility:1 analyzed:1 superfluous:1 accurate:1 necessary:1 experience:1 divide:2 initialized:1 desired:1 littlestone:6 renormalize:1 theoretical:4 mk:2 instance:5 classify:1 cover:1 caruana:1 oflearning:2 wix:1 lit95:2 quickly:1 again:1 thesis:1 satisfied:1 algori:1 emnlp:1 expert:67 return:1 li:1 suggesting:1 converted:1 speculate:1 includes:2 matter:1 lit91:2 explicitly:1 depends:2 vi:1 multiplicative:1 break:1 portion:1 start:2 contribution:1 minimize:2 il:1 who:1 identify:1 generalize:2 researcher:1 lit89:9 cumbersome:1 definition:1 involved:1 obvious:2 naturally:1 proof:6 adjusting:1 mitchell:1 dh73:2 tolerate:1 though:2 generality:1 furthermore:1 likelyhood:1 hand:1 receives:2 effect:4 contain:1 true:1 concept:2 normalized:2 multiplier:1 deal:1 eg:2 during:2 customize:1 complete:1 confusion:1 performs:2 mesterharm:5 ji:1 empirically:1 belong:1 extend:2 significant:1 refer:2 stable:1 similarity:1 add:7 own:1 winnow:27 irrelevant:1 driven:1 certain:4 binary:3 arbitrarily:1 mcdermott:1 minimum:2 additional:2 converting:1 determine:1 redundant:1 ii:1 desirable:2 technical:2 long:1 lin:1 hart:1 plugging:1 prediction:20 involving:1 regression:5 rutgers:3 represent:2 extra:3 comment:1 call:3 near:1 concerned:2 fori:1 fit:2 zi:5 affect:1 idea:1 multiclass:1 golding:1 handled:1 york:1 useful:7 customizing:1 santa:1 amount:2 notice:3 disjoint:1 correctly:2 track:1 per:1 threshold:4 blum:2 changing:1 fraction:2 sum:3 convert:1 extends:1 family:3 reader:1 bound:15 fl:2 nonnegative:4 scene:1 felt:1 aspect:1 argument:1 min:4 attempting:1 performing:1 relatively:1 department:1 piscataway:1 combination:4 poor:1 kw97:2 representable:1 describes:1 wi:3 modification:4 intuitively:1 taken:1 ln:1 turn:1 committee:40 needed:2 end:1 available:1 multiplied:4 indirectly:1 original:5 remaining:1 include:2 running:1 overflow:1 added:1 spelling:2 traditional:2 gradient:1 separate:1 thank:2 majority:1 chris:3 reason:1 code:1 z3:3 providing:1 zabowski:1 difficult:2 statement:1 negative:4 implementation:1 proper:2 perform:2 conversion:1 finite:1 extended:1 looking:1 thm:1 introduced:1 pair:1 z1:1 nick:6 california:1 distinction:1 learned:1 nee:1 able:1 pattern:1 max:5 including:1 power:1 overlap:1 natural:2 hybrid:3 difficulty:3 predicting:1 kivinen:2 scheme:2 improve:1 picture:1 irrespective:1 text:2 relative:2 loss:2 interesting:2 versus:1 foundation:2 classifying:1 course:1 changed:2 supported:1 last:3 surprisingly:1 formal:1 allow:3 understand:1 warren:1 institute:1 exponentiated:1 dagan:1 feedback:1 xn:1 world:2 doesn:1 made:2 keep:2 ml:3 global:1 xi:1 learn:6 robust:1 superficial:1 improving:1 domain:1 did:1 whole:1 noise:4 paul:1 allowed:2 wiley:1 sub:66 xl:1 learns:3 ix:1 theorem:2 ton:1 normalizing:1 ofthis:1 exists:1 avrim:1 adding:3 phd:1 customization:1 logarithmic:2 adjustment:1 gr96:3 corresponds:3 truth:2 underflow:1 stimulating:1 goal:2 jyrki:1 crl:1 change:1 except:1 reducing:1 wt:1 called:1 total:1 experimental:1 vote:10 select:1 formally:1 people:1 support:1 inability:1 princeton:1
785
1,719
The Relevance Vector Machine Michael E. Tipping Microsoft Research St George House, 1 Guildhall Street Cambridge CB2 3NH, U.K. mtipping~microsoft.com Abstract The support vector machine (SVM) is a state-of-the-art technique for regression and classification, combining excellent generalisation properties with a sparse kernel representation. However, it does suffer from a number of disadvantages, notably the absence of probabilistic outputs, the requirement to estimate a trade-off parameter and the need to utilise 'Mercer' kernel functions. In this paper we introduce the Relevance Vector Machine (RVM), a Bayesian treatment of a generalised linear model of identical functional form to the SVM. The RVM suffers from none of the above disadvantages, and examples demonstrate that for comparable generalisation performance, the RVM requires dramatically fewer kernel functions. 1 Introd uction In supervised learning we are given a set of examples of input vectors {Xn}~=l along with corresponding targets {tn}~=l' the latter of which might be real values (in regression) or class labels (classification). From this 'training' set we wish to learn a model of the dependency of the targets on the inputs with the objective of making accurate predictions of t for previously unseen values of x. In real-world data, the presence of noise (in regression) and class overlap (in classification) implies that the principal modelling challenge is to avoid 'over-fitting' of the training set. A very successful approach to supervised learning is the support vector machine (SVM) [8]. It makes predictions based on a function of the form N 2:: (1) wnK(x, x n ) + Wo, n=l where {w n } are the model 'weights' and K(?,?) is a kernel function. The key feature of the SVM is that, in the classification case, its target function attempts to minimise the number of errors made on the training set while simultaneously maximising the 'margin' between the two classes (in the feature space implicitly defined by the kernel). This is an effective 'prior' for avoiding over-fitting, which leads to good generalisation, and which furthermore results in a sparse model dependent only on a subset of kernel functions: those associated with training examples Xn that lie either on the margin or on the 'wrong' side of it. State-of-the-art results have been reported on many tasks where SVMs have been applied. y(x) = 653 The Relevance Vector Machine However, the support vector methodology does exhibit significant disadvantages: ? Predictions are not probabilistic. In regression the SVM outputs a point estimate, and in classification, a 'hard' binary decision. Ideally, we desire to estimate the conditional distribution p(tlx) in order to capture uncertainty in our prediction. In regression this may take the form of 'error-bars', but it is particularly crucial in classification where posterior probabilities of class membership are necessary to adapt to varying class priors and asymmetric misclassification costs. ? Although relatively sparse, SVMs make liberal use of kernel functions, the requisite number of which grows steeply with the size of the training set. ? It is necessary to estimate the error/margin trade-off parameter 'e' (and in regression, the insensitivity parameter I f' too). This generally entails a cross-validation procedure, which is wasteful both of data and computation. ? The kernel function K(?,?) must satisfy Mercer's condition. In this paper, we introduce the 'relevance vector machine' (RVM), a probabilistic sparse kernel model identical in functional form to the SVM. Here we adopt a Bayesian approach to learning, where we introduce a prior over the weights governed by a set of hyperparameters, one associated with each weight, whose most probable values are iteratively estimated from the data. Sparsity is achieved because in practice we find that the posterior distributions of many of the weights are sharply peaked around zero. Furthermore, unlike the support vector classifier, the nonzero weights in the RVM are not associated with examples close to the decision boundary, but rather appear to represent 'prototypical' examples of classes. We term these examples 'relevance' vectors, in deference to the principle of automatic relevance determination (ARD) which motivates the presented approach [4, 6J. The most compelling feature of the RVM is that, while capable of generalisation performance comparable to an equivalent SVM, it typically utilises dramatically fewer kernel functions. Furthermore, the RVM suffers from none of the other limitations of the SVM outlined above. In the next section, we introduce the Bayesian model, initially for regression, and define the procedure for obtaining hyperparameter values, and thus weights. In Section 3, we give brief examples of application of the RVM in the regression case, before developing the theory for the classification case in Section 4. Examples of RVM classification are then given in Section 5, concluding with a discussion. 2 Relevance Vector Regression Given a dataset of input-target pairs {xn, tn}~=l' we follow the standard formulation and assume p(tlx) is Gaussian N(tIY(x), a 2 ). The mean ofthis distribution for a given x is modelled by y(x) as defined in (1) for the SVM. The likelihood of the dataset can then be written as - ~w)1I2 } , N x (N + 1) p(tlw, a 2 ) = (27ra 2 )-N/2 exp { - 2: 2 lit (2) where t = (tl ... tN), W = (wo .. . WN) and ~ is the 'design' matrix with ~nm = K(x n , Xm - l) and ~nl = 1. Maximum-likelihood estimation of wand a 2 from (2) will generally lead to severe overfitting, so we encode a preference for smoother functions by defining an ARD Gaussian prior [4, 6J over the weights: N p(wla) = II N(wiIO,ai i=O 1 ), (3) ME. Tipping 654 with 0 a vector of N + 1 hyperparameters. This introduction of an individual hyperparameter for every weight is the key feature of the model, and is ultimately responsible for its sparsity properties. The posterior over the weights is then obtained from Bayes' rule: p(wlt, 0,0'2) = (21r)-(N+l)/21:E1- 1 / 2 exp { -~(w - J.lY:E-1(w - JL)}, (4) with :E = (q,TBq, JL = + A)-I, (5) :Eq, TBt, (6) where we have defined A = diag(ao,al, ... ,aN) and B = 0'-2IN. Note that 0'2 is also treated as a hyperparameter, which may be estimated from the data. By integrating out the weights, we obtain the marginal likelihood, or evidence [2], for the hyperparameters: p(tIO,0'2) = (21r)-N/2IB- 1 + q,A -1q, TI- 1 / 2 exp { -~e(B-l + q,A -lq,T)-lt} . (7) For ideal Bayesian inference, we should define hyperpriors over 0 and 0'2, and integrate out the hyperparameters too. However, such marginalisation cannot be performed in closed-form here, so we adopt a pragmatic procedure, based on that of MacKay [2], and optimise the marginal likelihood (7) with respect to 0 and 0'2, which is essentially the type II maximum likelihood method [1] . This is equivalent to finding the maximum of p(o, 0'2It), assuming a uniform (and thus improper) hyperprior. We then make predictions, based on (4), using these maximising values. 2.1 Optimising the hyperparameters Values of 0 and 0'2 which maximise (7) cannot be obtained in closed form, and we consider two alternative formulae for iterative re-estimation of o . First, by considering the weights as 'hidden' variables, an EM approach gives: new 1 1 (8) ai = -( 2) 2' Wi p(wlt,Q,u2) Eii + J-Li Second, direct differentiation of (7) and rearranging gives: 'Yi ' ainew = 2 (9) J-Li where we have defined the quantities 'Yi = 1 - aiEii, which can be interpreted as a measure of how 'well-determined' each parameter Wi is by the data [2]. Generally, this latter update was observed to exhibit faster convergence. For the noise variance, both methods lead to the same re-estimate: (10) In practice, during re-estimation, we find that many of the ai approach infinity, and from (4), p(wilt,0,0'2) becomes infinitely peaked at zero - implying that the corresponding kernel functions can be 'pruned'. While space here precludes a detailed explanation, this occurs because there is an 'Occam' penalty to be paid for smaller values of ai, due to their appearance in the determinant in the marginal likelihood (7). For some ai, a lesser penalty can be paid by explaining the data with increased noise 0'2, in which case those ai -+ 00. The Relevance Vector Machine 3 3.1 655 Examples of Relevance Vector Regression Synthetic example: the 'sine' function The function sinc(x) = Ixl- 1 sin Ixl is commonly used to illustrate support vector regression [8], where in place of the classification margin, the f.-insensitive region is introduced, a 'tube' of ?f. around the function within which errors are not penalised. In this case, the support vectors lie on the edge of, or outside, this region. For example, using linear spline kernels and with f. = 0.01, the approximation ofsinc(x) based on 100 uniformly-spaced noise-free samples in [-10, 10J utilises 39 support vectors [8]. By comparison, we approximate the same function with a relevance vector model utilising the same kernel. In this case the noise variance is fixed at 0.012 and 0 alone re-estimated. The approximating function is plotted in Figure 1 (left), and requires only 9 relevance vectors. The largest error is 0.0087, compared to 0.01 in the SV case. Figure 1 (right) illustrates the case where Gaussian noise of standard deviation 0.2 is added to the targets. The approximation uses 6 relevance vectors, and the noise is automatically estimated, using (10), as (7 = 0.189. 1.2 0.8 0.6 . 0.4 '. .- . 0.2 , 0 -0.2 -0.4 5 10 -10 -5 10 Figure 1: Relevance vector approximation to sinc(x): noise-free data (left), and with added Gaussian noise of (]" = 0.2 (right) . The estimated functions are drawn as solid lines with relevance vectors shown circled, and in the added-noise case (right) the true function is shown dashed. 3.2 Some benchmarks The table below illustrates regression performance on some popular benchmark datasets - Friedman's three synthetic functions (results averaged over 100 randomly generated training sets of size 240 with a lOOO-example test set) and the 'Boston housing' dataset (averaged over 100 randomised 481/25 train/test splits). The prediction error obtained and the number of kernel functions required for both support vector regression (SVR) and relevance vector regression (RVR) are given. Dataset Friedman #1 Friedman #2 Friedman #3 Boston Housing _ errors_ SVR RVR _ kernels _ SVR RVR 2.92 4140 0.0202 8.04 116.6 110.3 106.5 142.8 2.80 3505 0.0164 7.46 59.4 6.9 11.5 39.0 656 4 M E. TIpping Relevance Vector Classification We now extend the relevance vector approach to the case of classification - Le. where it is desired to predict the posterior probability of class membership given the input x. We generalise the linear model by applying the logistic sigmoid function a(y) = 1/(1 + e- Y ) to y(x) and writing the likelihood as N P(tlw) = II a{y(xn)}tn [1 - a{Y(Xn)}]l-tn . (11) n==l However, we cannot integrate out the weights to obtain the marginal likelihood analytically, and so utilise an iterative procedure based on that of MacKay [3]: 1. For the current, fixed, values of a we find the most probable weights WMP (the location of the posterior mode). This is equivalent to a standard optimisation of a regularised logistic model, and we use the efficient iterativelyreweighted least-squares algorithm [5] to find the maximum. 2. We compute the Hessian at WMP: \7\7logp(t, wla)1 WMP = _(<)TB<) + A), (12) where Bnn = a{y(x n )} [1 - a{y(x n )}], and this is negated and inverted to give the covariance I: for a Gaussian approximation to the posterior over weights, and from that the hyperparameters a are updated using (9). Note that there is no 'noise' variance a 2 here. This procedure is repeated until some suitable convergence criteria are satisfied. Note that in the Bayesian treatment of multilayer neural networks, the Gaussian approximation is considered a weakness of the method if the posterior mode is unrepresentative of the overall probability mass. However, for the RVM, we note that p(t, wla) is log-concave (i.e. the Hessian is negative-definite everywhere), which gives us considerably more confidence in the Gaussian approximation. 5 5.1 Examples of RVM Classification Synthetic example: Gaussian mixture data We first utilise artificially generated data in two dimensions in order to illustrate graphically the selection of relevance vectors. Class 1 (denoted by 'x ') was sampled from a single Gaussian, and overlaps to a small degree class 2 ('.'), sampled from a mixture of two Gaussians. A relevance vector classifier was compared to its support vector counterpart, using the same Gaussian kernel. A value of C for the SVM was selected using 5-fold crossvalidation on the training set. The results for a typical dataset of 200 examples are given in Figure 2. The test errors for the RVM (9.32%) and SVM (9.48%) are comparable, but the remarkable feature of contrast is the complexity of the classifiers. The support vector machine utilises 44 kernel functions compared to just 3 for the relevance vector method. It is also notable that the relevance vectors are some distance from the decision boundary (in x-space). Given further analysis, this observation can be seen to be consistent with the hyperparameter update equations. A more qualitative explanation is that the output of a basis function lying on or near the decision boundary is a poor indicator of class membership, and such basis functions are naturally 'penalised' under the Bayesian framework. 657 The Relevance Vector Machine SVM: error=9.48% vectors=44 , \ \ @. ~ " , x x \ X x x X x x x ~x ., .... ? . ? II ? 'C' ? "I. x ?\ ? x x ~ ?? _~ x x~ .. x.":X~x x ~ x :xx xx'f. RVM: error=9.32% vectors=3 ? ? xxx x x x ,, ,, ? I I , \ Figure 2: Results of training functionally identical SVM (left) and RVM (right) classifiers on a typical synthetic dataset. The decision boundary is shown dashed, and relevance/support vectors are shown circled to emphasise the dramatic reduction in complexity of the RVM model. 5.2 Real examples In the table below we give error and complexity results for the 'Pima Indian diabetes' and the 'U.S.P.S. handwritten digit' datasets. The former task has been recently used to illustrate Bayesian classification with the related Gaussian Process (GP) technique [9], and we utilised those authors' split of the data into 200 training and 332 test examples and quote their result for the GP case. The latter dataset is a popular support vector benchmark, comprising 7291 training examples along with a 2007-example test set, and the SVM result is quoted from [7]. Dataset Pima Indians U.S.P.S. ___ errors ___ SVM GP RVM 67 68 65 4.4% 5.1% _ _ kernels _ _ SVM GP RVM 109 2540 200 4 316 In terms of prediction accuracy, the RVM is marginally superior on the Pima set, but outperformed by the SVM on the digit data. However, consistent with other examples in this paper, the RVM classifiers utilise many fewer kernel functions. Most strikingly, the RVM achieves state-of-the-art performance on the diabetes dataset with only 4 kernels. It should be noted that reduced set methods exist for subsequently pruning support vector models to reduce the required number of kernels at the expense of some increase in error (e.g. see [7] for some example results on the U.S.P.S. data). 6 Discussion Examples in this paper have effectively demonstrated that the relevance vector machine can attain a comparable (and for regression, apparently superior) level of generalisation accuracy as the well-established support vector approach, while at the same time utilising dramatically fewer kernel functions - implying a considerable 658 ME. Tipping saving in memory and computation in a practical implementation. Importantly, we also benefit from the absence of any additional nuisance parameters to set, apart from the need to choose the type of kernel and any associated parameters. In fact, for the case of kernel parameters, we have obtained improved (both in terms of accuracy and sparsity) results for all the benchmarks given in Section 3.2 when optimising the marginal likelihood with respect to multiple input scale parameters in Gaussian kernels (q. v. [9]). Furthermore, we may also exploit the Bayesian formalism to guide the choice of kernel itself [2], and it should be noted that the presented methodology is applicable to arbitrary basis functions, so we are not limited, for example, to the use of 'Mercer' kernels as in the SVM. A further advantage of the RVM classifier is its standard formulation as a probabilistic generalised linear model. This implies that it can be extended to the multiple-class case in a straightforward and principled manner, without the need to train and heuristically combine multiple dichotomous classifiers as is standard practice for the SVM. Furthermore, the estimation of posterior probabilities of class membership is a major benefit, as these convey a principled measure of uncertainty of prediction, and are essential if we wish to allow adaptation for varying class priors, along with incorporation of asymmetric misclassification costs. However, it must be noted that the principal disadvantage of relevance vector methods is in the complexity of the training phase, as it is necessary to repeatedly compute and invert the Hessian matrix, requiring O(N2) storage and O(N3) computation. For large datasets, this makes training considerably slower than for the SVM. Currently, memory constraints limit us to training on no more than 5,000 examples, but we have developed approximation methods for handling larger datasets which were employed on the U.S.P.S. handwritten digit database'. We note that while the case for Bayesian methods is generally strongest when data is scarce, the sparseness of the resulting classifier induced by the Bayesian framework presented here is a compelling motivation to apply relevance vector techniques to larger datasets. Acknowledgements The author wishes to thank Chris Bishop, John Platt and Bernhard Sch5lkopf for helpful discussions, and JP again for his Sequential Minimal Optimisation code. References [1) J. O. Berger. Statistical decision theory and Bayesian analysis. Springer, New York, second edition , 1985. [2) D. J. C. Mackay. Bayesian interpolation. Neural Computation, 4(3):415-447, 1992. [3) D. J . C. Mackay. The evidence framework applied to classification networks. Neural Computation, 4(5):720-736, 1992. (4) D . J . C. Mackay. Bayesian non-linear modelling for the prediction competition. In ASHRAE Transactions, vol. 100, pages 1053- 1062. ASHRAE, Atlanta, Georgia, 1994. (5) 1. T . Nabney. Efficient training of RBF networks for classification. In Proceedings of ICANN99, pages 210-215, London, 1999. lEE. [6) R. M. Neal. Bayesian Learning for Neural Networks. Springer, New York, 1996. (7) B. Sch6lkopf, S. Mika, C. J. C. Burges, P. Knirsch, K.-R. Miiller, G . Ratsch , and A. J. Smola. Input space versus feature space in kernel-based methods. IEEE Transactions on Neural Networks, 10(5) :1000- 1017, 1999. [8) V. N. Vapnik. Statistical Learning Theory. Wiley, New York, 1998. [9) C. K. 1. Williams and D. Barber. Bayesian classification with Gaussian processes. IEEE Trans. Pattern Analysis and Machine Intelligence, 20(12) :1342-1351, 1998.
1719 |@word determinant:1 wla:3 heuristically:1 covariance:1 paid:2 dramatic:1 solid:1 reduction:1 current:1 com:1 must:2 written:1 john:1 update:2 implying:2 alone:1 fewer:4 selected:1 intelligence:1 location:1 preference:1 liberal:1 along:3 direct:1 qualitative:1 fitting:2 combine:1 manner:1 introduce:4 notably:1 ra:1 automatically:1 considering:1 becomes:1 xx:2 mass:1 interpreted:1 developed:1 finding:1 differentiation:1 every:1 ti:1 concave:1 wrong:1 classifier:8 platt:1 ly:1 appear:1 generalised:2 before:1 maximise:1 limit:1 interpolation:1 might:1 mika:1 limited:1 averaged:2 practical:1 responsible:1 practice:3 definite:1 cb2:1 digit:3 procedure:5 attain:1 confidence:1 integrating:1 svr:3 cannot:3 close:1 selection:1 storage:1 applying:1 writing:1 equivalent:3 demonstrated:1 graphically:1 straightforward:1 williams:1 rule:1 mtipping:1 importantly:1 his:1 updated:1 target:5 us:1 regularised:1 diabetes:2 particularly:1 asymmetric:2 database:1 observed:1 capture:1 region:2 improper:1 trade:2 principled:2 complexity:4 ideally:1 ultimately:1 basis:3 strikingly:1 train:2 effective:1 london:1 outside:1 whose:1 larger:2 precludes:1 tlw:2 unseen:1 gp:4 itself:1 housing:2 advantage:1 adaptation:1 combining:1 insensitivity:1 competition:1 crossvalidation:1 convergence:2 requirement:1 illustrate:3 ard:2 eq:1 implies:2 subsequently:1 ao:1 probable:2 lying:1 around:2 considered:1 guildhall:1 exp:3 predict:1 major:1 achieves:1 adopt:2 deference:1 estimation:4 outperformed:1 applicable:1 label:1 currently:1 quote:1 rvm:21 largest:1 gaussian:13 rather:1 avoid:1 varying:2 encode:1 modelling:2 likelihood:9 steeply:1 contrast:1 helpful:1 inference:1 dependent:1 membership:4 typically:1 initially:1 hidden:1 comprising:1 overall:1 classification:16 denoted:1 art:3 mackay:5 marginal:5 saving:1 identical:3 optimising:2 lit:1 peaked:2 wlt:2 spline:1 tiy:1 randomly:1 simultaneously:1 individual:1 phase:1 microsoft:2 attempt:1 tlx:2 friedman:4 atlanta:1 severe:1 weakness:1 mixture:2 nl:1 accurate:1 edge:1 capable:1 necessary:3 hyperprior:1 re:4 plotted:1 desired:1 minimal:1 ashrae:2 increased:1 formalism:1 compelling:2 disadvantage:4 logp:1 cost:2 deviation:1 subset:1 uniform:1 successful:1 too:2 reported:1 dependency:1 sv:1 synthetic:4 considerably:2 st:1 probabilistic:4 off:2 lee:1 michael:1 again:1 nm:1 tube:1 satisfied:1 choose:1 knirsch:1 li:2 satisfy:1 wilt:1 notable:1 performed:1 sine:1 utilised:1 closed:2 apparently:1 bayes:1 square:1 accuracy:3 variance:3 spaced:1 modelled:1 bayesian:15 handwritten:2 tbt:1 none:2 marginally:1 strongest:1 penalised:2 suffers:2 naturally:1 associated:4 sampled:2 dataset:9 treatment:2 popular:2 tipping:4 supervised:2 follow:1 methodology:2 xxx:1 improved:1 formulation:2 furthermore:5 wmp:3 just:1 smola:1 until:1 logistic:2 mode:2 grows:1 requiring:1 true:1 counterpart:1 former:1 analytically:1 iteratively:1 nonzero:1 i2:1 bnn:1 neal:1 sin:1 during:1 nuisance:1 noted:3 criterion:1 eii:1 demonstrate:1 tn:5 recently:1 sigmoid:1 superior:2 functional:2 insensitive:1 nh:1 jl:2 extend:1 jp:1 functionally:1 significant:1 cambridge:1 ai:6 automatic:1 outlined:1 entail:1 posterior:8 apart:1 binary:1 yi:2 inverted:1 seen:1 george:1 additional:1 utilises:3 employed:1 dashed:2 ii:4 smoother:1 multiple:3 faster:1 adapt:1 determination:1 cross:1 e1:1 prediction:9 regression:15 multilayer:1 optimisation:2 essentially:1 represent:1 kernel:28 invert:1 achieved:1 ratsch:1 crucial:1 marginalisation:1 unlike:1 induced:1 near:1 presence:1 ideal:1 split:2 wn:1 reduce:1 lesser:1 minimise:1 introd:1 wo:2 penalty:2 suffer:1 miiller:1 hessian:3 york:3 repeatedly:1 dramatically:3 generally:4 detailed:1 svms:2 reduced:1 exist:1 estimated:5 hyperparameter:4 vol:1 key:2 drawn:1 wasteful:1 wand:1 utilising:2 everywhere:1 uncertainty:2 place:1 decision:6 comparable:4 nabney:1 fold:1 infinity:1 sharply:1 incorporation:1 constraint:1 n3:1 dichotomous:1 concluding:1 pruned:1 relatively:1 developing:1 poor:1 smaller:1 em:1 wi:2 making:1 equation:1 previously:1 randomised:1 gaussians:1 apply:1 hyperpriors:1 alternative:1 slower:1 exploit:1 wnk:1 approximating:1 objective:1 added:3 quantity:1 occurs:1 exhibit:2 distance:1 thank:1 street:1 me:2 chris:1 barber:1 maximising:2 assuming:1 code:1 berger:1 pima:3 expense:1 negative:1 design:1 implementation:1 motivates:1 negated:1 sch6lkopf:1 observation:1 datasets:5 benchmark:4 defining:1 extended:1 arbitrary:1 introduced:1 pair:1 required:2 established:1 trans:1 bar:1 below:2 pattern:1 xm:1 sparsity:3 challenge:1 tb:1 optimise:1 memory:2 explanation:2 overlap:2 misclassification:2 treated:1 suitable:1 indicator:1 scarce:1 sch5lkopf:1 brief:1 prior:5 circled:2 acknowledgement:1 prototypical:1 limitation:1 versus:1 ixl:2 remarkable:1 validation:1 integrate:2 degree:1 consistent:2 mercer:3 principle:1 occam:1 looo:1 free:2 side:1 guide:1 allow:1 generalise:1 burges:1 explaining:1 sparse:4 emphasise:1 benefit:2 boundary:4 dimension:1 xn:5 world:1 author:2 made:1 commonly:1 transaction:2 approximate:1 pruning:1 implicitly:1 bernhard:1 overfitting:1 quoted:1 iterative:2 table:2 learn:1 rearranging:1 obtaining:1 excellent:1 artificially:1 diag:1 motivation:1 noise:11 hyperparameters:6 edition:1 n2:1 repeated:1 convey:1 tl:1 georgia:1 wiley:1 wish:3 lq:1 lie:2 house:1 governed:1 ib:1 formula:1 bishop:1 svm:20 sinc:2 evidence:2 ofthis:1 essential:1 uction:1 vapnik:1 sequential:1 effectively:1 tio:1 illustrates:2 sparseness:1 margin:4 boston:2 lt:1 appearance:1 infinitely:1 desire:1 u2:1 springer:2 utilise:4 conditional:1 rbf:1 unrepresentative:1 absence:2 considerable:1 hard:1 generalisation:5 determined:1 uniformly:1 typical:2 principal:2 pragmatic:1 support:14 latter:3 relevance:26 indian:2 icann99:1 requisite:1 avoiding:1 handling:1
786
172
678 ALOW-POWER CMOS CIRCUIT WHICH EMULATES TEMWORALELECTIDCALPROPERTIES OF NEURONS Jack L. Meador and Clint S. Cole Electrical and Computer Engineering Dept. Washington State University Pullman WA. 99164-2752 ABSTRACf This paper describes a CMOS artificial neuron. The circuit is directly derived from the voltage-gated channel model of neural membrane, has low power dissipation, and small layout geometry. The principal motivations behind this work include a desire for high performance, more accurate neuron emulation, and the need for higher density in practical neural network implementations. INTRODUCTION Popular neuron models are based upon some statistical measure of known natural behavior. Whether that measure is expressed in terms of average firing rate or a firing probability, the instantaneous neuron activation is only represented in an abstract sense. Artificial electronic neurons derived from these models represent this excitation level as a binary code or a continuous voltage at the output of a summing amplifier. While such models have been shown to perform well for many applications, and form an integral part of much current work, they only partially emulate the manner in which natural neural networks operate. They ignore, for example, differences in relative arrival times of neighboring action potentials -- an important characteristic known to exist in natural auditory and visual networks {Sejnowski, 1986}. They are also less adaptable to fme-grained, neuron-centered learning, like the post-tetanic facilitation observed in natural neurons. We are investigating the implementation and application of neuron circuits which better approximate natural neuron function. BACKGROUND The major temporal artifacts associated with natural neuron function include the spacio-temporal integration of synaptic activity, the generation of an action potential (AP), and the post-AP hyperpolarization (refractory) period (Figure 1). Integration, manifested as a gradual membrane depolarization, occurs when the neuron accumulates sodium ions which migrate through pores in its cellular membrane. The rate of ion migration is related to the level of presynaptic AP bombardment, and is also known to be a non-linear function of transmembrane potential. Efferent AP generation occurs when the voltage-sensitive membrane of the axosomal hillock reaches some threshold potential whereupon a rapid increase in sodium permeability leads to A Low-Power CMOS Circuit Which Emulates Neurons complete depolarization. Immediately thereafter, sodium pores "close" simultaneously with increased potassium permeability, thereby repolarizing the membrane toward its resting potential. The high potassium permeability during AP generation leads to the transient post-AP hyperpolarization state known as the refractory period. v Actl vat Ion Threshold ( 1) ( 3) Figure 1. Temporal artifacts associated with neuron function. (1) gradual depolarization, (2) AP generation, (3) refractory period. Several analytic and electronic neural models have been proposed which embody these characteristics at varying levels of detail. These neuromimes have been used to good advantage in studying neuron behavior. However, with the advent of artificial neural networks (ANN) for computing, emphasis has switched from modeling neurons for physiologic studies to developing practical neural network implementations. As the desire for high performance ANNs grows, models amenable to hardware implementation become more attractive. The general idea behind electronic neuromimes is not new. Beginning in 1937 with work by Harmon {Harmon, 1937},lIectronic circuits have been used to model and study neuronal behavior. In the late 196(Ys, Lewis {Lewis, 1968} developed a circuit which simulated the Hodgkin-Huxley model for a single neuron, followed by MacQregor's circuit {MacGregor, 1973} in the early 1970's which modelled a group of 50 neurons. With the availability of VLSI in the 1980's, electronic neural implementations have largely moved to the realm of integrated circuits. Two different strategies have been documented: analog implementations employing operational amplifiers {Graf, et at, 1987,1988; Sivilotti, et at, 1986; Raffel, 1988; Schwartz, et al, 1988}; and digital implementations such as systolic arrays {Kung, 1988}. More recently, impulse neural implementations are receiving increased attention. like other models, these neuromimes generate outputs based on some non-linear function of the weighted net inputs. However, interneuron communication is realized through impulse streams rather than continuous voltages or binary numbers {Murray, 1988; N. El-Leithy, 1987}. Impulse networks communicate neuron activation as variable pulse repetition rates. The impulse neuron circuits which shall be discussed offer both small geometry and low power dissipation as well as a closer approximation to natural neuron function. 679 680 Meador and Cole - A CMOS IMPULSE NEURON An impulse neuron circuit developed for use in CMOS neural networks is shown in Figure 2. In this circuit, membrane ion current is modeled by charge flowing to and from Ca. Potassium and sodium influx is represented by current flow from Vdd to the capacitor, and ion efflux by flow from the capacitor to ground. The Field EffectTransistors (FETs) connected between Vdd, Vsr , and the capacitor emulate voltageand chemically-gated ion channels found in natural neural membrane. In the Figure, PET 1 corresponds to the post-synaptic chemicaIly-gated ion channels associated with one synapse. PETs 2, 3, and 4 emulate the voltage-gated channels distributed throughout a neuron membrane. The following equations summarize circuit operation: Ca dVa/dt=/31E (Vr,Va)+/3:uF(Va)-/34G (Va) (1) E(Vr,Va) = (Vr-Va-V",)(Vdd-Va)-(Vdd-Va)2 /2 (2) F(Y. ) ={(Vdd-Vtp) (Va-Vdd)-(Va-Vdd)2 /2 if g(t) ~O (3) 0 a otherwlSe G(V)={(Vdd-V",)Va-Va2/2 if h(t~)=O 0 a otherwlSe g(t) =h (t)(l-h (t-C)) (4) (5) 0 if Va(t) > Vth; 1 h()Vtl<Va(t)<Vth and h (Va(t-e))=O t - 1 if Va(t)<Vtl ; Vtl<Va(t)<Vth and h (~(t-e))=l (6) Vdd {52 Exci tator y Synapse 10 {53 {51 + Axon Vs + Va Ca 1 /5. Figure 2. A CMOS impulse neuron with one excitatory synapse-PET. A Low-Power CMOS Circuit Which Emulates Neurons Equation (1) expresses how changes in Va (which emulates instantaneous neuron excitation) depend upon the sum of three current components controlled by these PETs. E, F, and G in equations (2) through (4) express PET drain-source currents as functions terminal voltages. Equations (3) and (5) rely upon the assumption that PET 2 and PET 3 are implemented as a single dual-gate device where the transconductance f3n=fJ2f33/f.P2+/33). Non-saturated PET operation is assumed for these equations even though the PETs will momentarily pass through saturation at the onset of conduction in the actual circuit. The Schmitt trigger circuit establishes a nonlinear positive feedback path responsible for action potential initiation. The upper threshold of the trigger (VIII) emulates the natural neuron activation threshold while the lower threshold (Va) emulates the maximum hyperpolarization voltage. Equation (6) expresses the hysterisis present in the Schmitt trigger transfer characteristic. When Vs reaches the upper Schmitt threshold, PET 3 turns on, creating a current path from Vtid to Cs , and emulating the upswing of a natural action potential spike. A moment later, PET 2 turns off, starting the action potential downswing. Simultaneously, PET 4 turns on, begining the absolute refractory period where Cs is discharged toward the maximum hyperpolarization potential. When that potential is reached, the Schmitt trigger turns off PET 4 and the impulse firing cycle is complete. The capacitor terminal voltage Va emulates all gross temporal artifacts associated with membrane potential, including spacio-temporal integration, the action potential spike, and a refractory period. The instantaneous net excitation to the neuron is represented by the total current flowing into the summing node on the floating plate of the capacitor. Charge packets are transferred from Vtid to the capacitor by the excitatory synapse PET. Excitatory packet magnitude is dependent upon the transconductance PI. Inhibitory synapses (not shown) operate similarly, but instead reduce capacitor voltage by drawing charge to Va. A buffered action potential signal useful for driving many synapse PETs is available at the axon output. The membrane potential components (E,F,and G) of the circuit equations describe nonlinear relationships between post-synaptic excitation (E), membrane potential (F and G), and membrane ion currents. The functional forms of these components are equivalent to those found between terminal voltages and currents in non-saturated PETs. It is notable that natural voltage-gated channels do not necessarily follow the same current-voltage relationship of a PET. Even though more accurate models and emulations of natural membrane conductance exist, it seems unlikely at this time that they would help further improve neural network implementation. There is little doubt that more complex circuitry would be required to better approximate the true nonlinear relationship found in the biochemistry of natural neural membrane. That need conflicts directly with the goal of high-density integration. IMPULSE NEURAL NE1WORKS ~ Organizing a collection of neuron circuits into a useful network confIguration requires some weight specification method. Weight values can be either directly specified by the designer or learned by the network. A method particularly suited for use with the fIXed PET-synapses of the foregoing circuit is to fust learn weights using an "off-line" 681 682 Meador and Cole simulation, then translate the numerical results to physical FET transconductances. To do this, the activation function of an impulse neuron is derived and used in a modified back-propagation learning procedure. IMPULSE NEURON ACfIVATION FUNCTION Learning algorithms typically require some expression of the neuron activation function. Neuron activation can be expressed as a numerical value, a binary pattern, or a circuit voltage. In an impulse neuron, activation is expressed in terms of firing rate. The more frequently an impulse neuron circuit fues, the greater its activation. Impulse neuron activation is a nonlinear function of the excitation imparted through its synapse connections. An analytical expression of this nonlinear function can be derived using a rectangular approximation of neuron impulse waveforms. It is fust necessary to defme a unit-impulse as one impulse conducted by a synapse FET having some pre-determined reference transconductance (f3~). In Figure 3, To represents an invariant activation impulse width which is assumed to be identical for all neurons. T 1 represents the variable time period required for the neuron to accumulate the equivalent of K unit-impulses input excitation prior to firing. It can be assumed that net input comes from a single excitatory synapse with no other excitation. It shall also be assumed that impulses arrive at a constant rate, so (7) T 1 =K /W;jR; where R/ is the firing rate of the source neuron and connecting neuron; and neuron j. The firing rate of the receiving neuron will be Rj this becomes: W;j is the weight of the synapse =1/(To + T 1). Substituting for T1 (8) F"tgure 3 compares this function with the logistic activation function. The impulse activation function approaches zero at the rate of 1/K when R; approaches zero. The function also approaches an asymptote of Rj =l/T0 as R; increases without bound. Any non-synaptic source which causes current flow from Vtit to Co will shift the curve to the left, and reflect a spontaneous firing rate at zero input excitation. A similar current source to VoU will shift the function to the right, reflecting a positive firingonset threshold. Circuit-level simulations show a clear correspondence to these analytical results. This functional form is also evident in activation curves experimentally observed with natural neurons {Guyton, 1986}. Various natural neurons are known to exhibit both spontaneous firing and fuing-onset thresholds as well. The impulse activation function constant, K, is determined by several factors including fJ~, Co, and To. Assuming that To? T h no leakage current exists, and that a FET conducting in its non-saturated region can be approximated by a resistor, the following expression for K is obtained: (9) A Low-Power CMOS Circuit Which Emulates Neurons where Rchon =l/pteJ<Vdd -V".), Ca is the summing capacitance, Va Vth are the low and high threshold voltages of the Schmitt trigger, and V". is the gate threshold voltage for an excitatory PET-synapse. A more accurate K value can be obtained by using the non-saturated PET current equation and solving a nonlinear differential equation. 1.0 rj~ ~ tTo~ 1, 1 ri j rectangUlar Impulse Train J 0 .5 Logistic Activation Impulse Activation o ~--~----------------------------------~---------------------------------------------ri o Figure 3. Rectangular impulse train approximation for impulse activation function derivation. Unlike the logistic function which asymptotically approaches zero, impulse activation is equal to zero over a range of net excitation. BACK?PROPAGATION IN IMPULSE NE1WORKS A back-propagation algorithm has been used to learn connection weights for impulse neural networks. At this time, weight values are non-adaptive (they are fixed at circuit fabrication) because they are implemented as invariant PET transconductances. Adaptive synapses compatible with impulse neuron circuits are in the early stages of development, but are not available at this time. Much can be learned about these networks using non-adaptive prototypes, however. As a result, weight learning is performed offline as part of the network design process. The back-propagation procedure used to learn weights for impulse networks differs from the generalized delta rule {Rumelhart, 1986} in two ways. The fust difference is the use of the impulse activation function instead of the logistic function. Any activation nonlinearity is a viable candidate for use with the generalized delta rule as long as it is differentiable. This is where difficulties mount with the impulse activation function. First of all, it is not differentiable at zero. What seems to be more important, however, is that its first derivative equals zero over a range of 683 684 Meador and Cole inputs. Examination of the generalized delta rule (which performs gradient-descent) reveals that when the fust derivative of neuron activation becomes zero, connections associated with that-neuron will cease to adapt. Once this happens, the procedure will most probably never arrive at a problem solution. To work around this problem, a second deviation from the generalized delta rule was implemented. This involves a departure from using the true first derivative when the impulse activation becomes zero. A small constant can be used to guarantee that learning continues even though the associated neuron activation is zero: Act =l/(To +K /Net) A ,_{(l/(To+K/Net)' if Net >0 ct - e otherwise (10) (11) The use of these equations yields a back-propagation algorithm for impulse networks which does not perform true gradient descent, yet which so far has been observed to learn solutions to logic problems such as XOR and the 4-2-4 encoder. Investigation of other offline learning algorithms for impulse networks continues. Currently, this algorithm fulfills the immediate need for an offline procedure which can be used in the design of multi-layer impulse neural networks. IMPLEMENTATION Two requirements for high density integration are low power dissipation and small circuit geometry. CMOS impulse neurons use switching circuits having no continuous power dissipation. A conventional op-amp circuit must draw constant current to achieve linear bias. An op-amp also requires larger circuit geometries for gain accuracy over typical fabrication process variations. Such is not the case for nonlinear switching circuits. As a result, these neurons and others like them are expected to help improve analog neural network integration density. An impulse neuron circuit has been designed which eliminates FETs 2 and 3 of Figure 2 in exchange for reduced layout area. In this circuit, Va no longer exhibits an activation potential spike. This spike seems irrelevant given the buffered impulse available at the axon output. The modified neuron circuit occupies 200 X 25 lambda chip area. A fIXed PET-synapse occupies a 16 by 18 lambda rectangle. With these dimensions a full-interconnect layout containing 40 neurons and 1600 fIXed connections will fit on a MOSIS 2-micron tiny chip. XOR and 4-2-4 networks of these circuits are being developed for 2-micron CMOS. CONCLUSION The motivation of this work is to improve neural network implementation technology by designing CMOS circuits derived from the temporal characteristics of natural neurons. The results obtained thus far include: Two CMOS circuits which closely correspond to the voltage-gat ed-channel model of natural neural membrane. A Low-Power CMOS Circuit Which Emulates Neurons Simulations which show that these impulse neurons emulate gross artifacts of natural neuron function. Initial work on a back-propagation algorithm which learns logic solutions using the impulse neuron activation function. The development of prototype impulse network I.Cs. Future goals involve extending this investigation to plastic synapse and neuron circuits, alternate algorithms for both offline and online learning, and practical implementations. Rererences H. P. Graf W. Hubbard L. D. Jackel P. G. N. deVegvar. A CMOS associative Memory Chip. IEEE ICNN Con. Proc., pp. 461-468, (1987). H. P. Graf L. D. Jackel W. E. Hubbard. VLSI Implementation of a Neural Network Model. IEEE Computer, pp. 41-49, (1988). A.C. Guyton. Chapt. 10. Organization of the Nervous System: Basic Functions of Synapses. Textbook o[ Physiology, p.l36. (1986) N. EI-Liethy, R.W. Newcomb, M. Zaghlou. A Basic MOS Neural-Type Junction A Perspective on Neural-Type Microsystems. IEEE ICNN Con. Proc., pp. 469-477, (1987). E. R. Lewis. Using Electronic Circuits to Model Simple Neuroelectric Interactions. Proc. IEEE 56, pp. 931-949, (1968). R. J. MacGregor R. M. Oliver. A General-Purpose Electronic Model for Arbitrary Configurations of Neurons. 1. Theor. Bioi. 38, pp. 527-538 (1973). S. Y. Kung. and J. N. Hwang. Parallel Achitectures for Artificial Neural Nets. IEEE ICNN Con. Proc., pp. 11-165 to 11-172, (1988). J. Mann R. Lippman B. Berger J. Raffel. A Self-Organizing Neural Net Chip. IEEE Cust.Integr. Ckts. Conf" pp. 10.3.1-10.3.5 (1988). A. F. Murray A. V. W. Smith. Asynchronous VLSI Neural Networks Using PulseStream Arithmetic. IEEE lnl. of Sol. St. Phys. 23, pp. 688-697, (1988). J. I. Raffel. Electronic Implementation of Neuromorphic Systems. IEEE Cust. Integr. Ckts. Conf" pp. 10.1.1-10.1.7, (1988). D. Rumelhart, G.E. Hinton, and RJ. Williams. Learning Internal Representations by Error Propagation. Parallel Distributed Processing, Vol 1: Foundations, pp. 318-364, (1986). O. H. Schmitt. Mechanical Solution of the Equations of Nerve Impulse Propagation. Am. 1. Physiol. 119, pp. 399-400, (1937). D. B. Schwartz R. E. Howard. A Programmable Analog Neural Network Chip. IEEE Cust. Integr. Ckts. Conf., pp. 10.2.1-1.2.4, (1988). 685 686 Meador and Cole T J. Sejnowski. Open Questions About Computation in Cerebral Cortex. Parallel Distributed Processing Vol. 2:Psychological and Biological Models, pp. 378-385, (1986). M. A. Sivilotti M. R. Emerling C. A. Mead. VISI Architectures for Implementation of Neural Networks. Am. Ins. of Phys., 408-413, (1986).
172 |@word pulsestream:1 seems:3 open:1 pulse:1 gradual:2 simulation:3 thereby:1 moment:1 initial:1 configuration:2 amp:2 current:15 activation:26 yet:1 must:1 physiol:1 numerical:2 analytic:1 asymptote:1 designed:1 v:2 device:1 nervous:1 vtp:1 beginning:1 smith:1 node:1 neuromimes:3 become:1 differential:1 viable:1 manner:1 expected:1 rapid:1 embody:1 behavior:3 frequently:1 multi:1 terminal:3 actual:1 little:1 becomes:3 circuit:38 advent:1 what:1 sivilotti:2 depolarization:3 textbook:1 developed:3 guarantee:1 temporal:6 act:1 charge:3 schwartz:2 unit:2 positive:2 t1:1 engineering:1 switching:2 accumulates:1 mount:1 mead:1 clint:1 firing:9 ap:7 path:2 emphasis:1 co:2 range:2 devegvar:1 practical:3 responsible:1 differs:1 lippman:1 procedure:4 area:2 physiology:1 pre:1 close:1 whereupon:1 equivalent:2 raffel:3 conventional:1 layout:3 attention:1 starting:1 vtl:3 williams:1 rectangular:3 immediately:1 rule:4 array:1 facilitation:1 variation:1 spontaneous:2 trigger:5 designing:1 rumelhart:2 approximated:1 particularly:1 continues:2 observed:3 electrical:1 region:1 connected:1 momentarily:1 cycle:1 sol:1 transmembrane:1 gross:2 efflux:1 vdd:10 depend:1 solving:1 fuing:1 upon:4 chip:5 represented:3 emulate:4 various:1 derivation:1 train:2 describe:1 sejnowski:2 artificial:4 larger:1 foregoing:1 drawing:1 otherwise:1 encoder:1 online:1 associative:1 advantage:1 differentiable:2 net:9 analytical:2 interaction:1 neighboring:1 organizing:2 translate:1 achieve:1 moved:1 potassium:3 abstracf:1 requirement:1 extending:1 cmos:14 help:2 op:2 p2:1 implemented:3 c:3 involves:1 come:1 waveform:1 emulation:2 closely:1 newcomb:1 centered:1 packet:2 occupies:2 transient:1 mann:1 require:1 exchange:1 icnn:3 investigation:2 biological:1 theor:1 around:1 ground:1 mo:1 circuitry:1 substituting:1 driving:1 major:1 biochemistry:1 early:2 purpose:1 proc:4 currently:1 jackel:2 cole:5 sensitive:1 hubbard:2 repetition:1 establishes:1 weighted:1 emerling:1 modified:2 rather:1 varying:1 voltage:16 derived:5 tetanic:1 sense:1 am:2 dependent:1 el:1 interconnect:1 integrated:1 unlikely:1 typically:1 fust:4 vlsi:3 dual:1 development:2 integration:6 field:1 equal:2 f3:1 having:2 washington:1 once:1 never:1 identical:1 represents:2 future:1 others:1 simultaneously:2 floating:1 geometry:4 alow:1 amplifier:2 conductance:1 organization:1 vsr:1 saturated:4 behind:2 amenable:1 accurate:3 oliver:1 integral:1 closer:1 necessary:1 harmon:2 permeability:3 psychological:1 increased:2 modeling:1 neuromorphic:1 deviation:1 bombardment:1 fabrication:2 conducted:1 hillock:1 conduction:1 migration:1 density:4 st:1 off:3 receiving:2 pore:2 connecting:1 reflect:1 containing:1 lambda:2 conf:3 creating:1 derivative:3 doubt:1 potential:16 availability:1 notable:1 onset:2 stream:1 later:1 performed:1 reached:1 parallel:3 otherwlse:2 accuracy:1 fme:1 xor:2 emulates:9 characteristic:4 largely:1 conducting:1 yield:1 discharged:1 correspond:1 modelled:1 plastic:1 anns:1 synapsis:4 reach:2 phys:2 synaptic:4 ed:1 pp:13 associated:6 con:3 efferent:1 gain:1 auditory:1 popular:1 realm:1 back:6 adaptable:1 reflecting:1 nerve:1 higher:1 dt:1 follow:1 flowing:2 synapse:12 though:3 stage:1 ei:1 nonlinear:7 propagation:8 logistic:4 artifact:4 impulse:44 hwang:1 grows:1 true:3 attractive:1 during:1 width:1 self:1 excitation:9 fet:5 macgregor:2 visi:1 generalized:4 plate:1 evident:1 complete:2 performs:1 dissipation:4 fj:1 jack:1 instantaneous:3 recently:1 functional:2 hyperpolarization:4 physical:1 leithy:1 refractory:5 cerebral:1 analog:3 discussed:1 resting:1 accumulate:1 buffered:2 similarly:1 nonlinearity:1 specification:1 vou:1 longer:1 cortex:1 imparted:1 perspective:1 irrelevant:1 meador:5 initiation:1 manifested:1 binary:3 greater:1 period:6 signal:1 arithmetic:1 full:1 rj:4 adapt:1 offer:1 long:1 post:5 y:1 vat:1 va:22 controlled:1 basic:2 represent:1 ion:8 background:1 source:4 operate:2 unlike:1 eliminates:1 probably:1 flow:3 capacitor:7 neuroelectric:1 fit:1 architecture:1 reduce:1 idea:1 prototype:2 shift:2 t0:1 whether:1 expression:3 cause:1 action:7 programmable:1 migrate:1 physiologic:1 useful:2 clear:1 involve:1 hardware:1 documented:1 generate:1 reduced:1 exist:2 inhibitory:1 designer:1 delta:4 shall:2 vol:2 express:3 group:1 thereafter:1 begining:1 threshold:10 rectangle:1 asymptotically:1 upswing:1 mosis:1 sum:1 micron:2 communicate:1 hodgkin:1 chemically:1 throughout:1 arrive:2 electronic:7 draw:1 dva:1 bound:1 ct:1 layer:1 followed:1 correspondence:1 tto:1 activity:1 huxley:1 ri:2 influx:1 transconductance:3 uf:1 transferred:1 developing:1 alternate:1 membrane:15 describes:1 jr:1 happens:1 invariant:2 equation:11 turn:4 integr:3 studying:1 available:3 operation:2 junction:1 gate:2 include:3 transconductances:2 repolarizing:1 murray:2 leakage:1 capacitance:1 ne1works:2 realized:1 occurs:2 spike:4 strategy:1 question:1 exhibit:2 gradient:2 simulated:1 presynaptic:1 cellular:1 toward:2 pet:22 viii:1 systolic:1 assuming:1 code:1 modeled:1 relationship:3 berger:1 implementation:15 design:2 gated:5 perform:2 upper:2 neuron:63 howard:1 descent:2 immediate:1 emulating:1 communication:1 hinton:1 arbitrary:1 defme:1 required:2 specified:1 mechanical:1 connection:4 conflict:1 learned:2 microsystems:1 pattern:1 departure:1 summarize:1 saturation:1 including:2 memory:1 power:9 natural:18 rely:1 difficulty:1 examination:1 sodium:4 improve:3 technology:1 prior:1 drain:1 relative:1 graf:3 generation:4 digital:1 foundation:1 switched:1 schmitt:6 tiny:1 pi:1 excitatory:5 compatible:1 asynchronous:1 offline:4 bias:1 absolute:1 cust:3 distributed:3 feedback:1 curve:2 dimension:1 lnl:1 collection:1 adaptive:3 employing:1 far:2 approximate:2 ignore:1 logic:2 investigating:1 reveals:1 summing:3 assumed:4 continuous:3 channel:6 transfer:1 learn:4 ca:4 operational:1 necessarily:1 complex:1 motivation:2 arrival:1 neuronal:1 vr:3 axon:3 resistor:1 candidate:1 late:1 learns:1 grained:1 cease:1 exists:1 gat:1 magnitude:1 interneuron:1 suited:1 exci:1 visual:1 vth:4 desire:2 expressed:3 partially:1 corresponds:1 lewis:3 bioi:1 goal:2 ann:1 change:1 experimentally:1 determined:2 typical:1 principal:1 f3n:1 total:1 pas:1 rererences:1 internal:1 fulfills:1 kung:2 dept:1
787
1,720
A Neuromorphic VLSI System for Modeling the Neural Control of Axial Locomotion Girish N. Patel girish@ece.gatech.edu Edgar A. Brown ebrown@ece.gatech.edu Stephen P. DeWeerth steved@ece.gatech.edu School of Electrical and Computer Engineering Georgia Institute of Technology Atlanta, Ga. 30332-0250 Abstract We have developed and tested an analog/digital VLSI system that models the coordination of biological segmental oscillators underlying axial locomotion in animals such as leeches and lampreys. In its current form the system consists of a chain of twelve pattern generating circuits that are capable of arbitrary contralateral inhibitory synaptic coupling. Each pattern generating circuit is implemented with two independent silicon Morris-Lecar neurons with a total of 32 programmable (floating-gate based) inhibitory synapses, and an asynchronous address-event interconnection element that provides synaptic connectivity and implements axonal delay. We describe and analyze the data from a set of experiments exploring the system behavior in terms of synaptic coupling. 1 Introduction In recent years, neuroscientists and modelers have made great strides towards illuminating structure and computational properties in biological motor systems. For example, much progress has been made toward understanding the neural networks that elicit rhythmic motor behaviors, including leech heartbeat, crustacean stomatogastric mill and lamprey swimming (a good review on these is in [1] and [2]). It is thought that these same mechanisms form the basis for more complex motor behaviors. The neural substrate for these control mechanisms are called central pattern generators (CPG). In the case of locomotion these circuits are distributed along the body (in the spinal cord of vertebrates or in the ganglia of invertebrates) and are richly interactive with sensory input and descending connections from the brain, giving rise to a highly distributed system as shown in Figure 1. In cases in which axial locomotion is involved, such as leech and lamprey swimming, synaptic interconnection patterns among autonomous segmental oscillators along the animal's axis produce coordinated motor patterns. These intersegmental coordination architectures have been well studied through both physiological experimentation and mathematical modeling. In addition, undulatory gaits in snakes have also been studied from a robotics perspective [3]. However, a thorough understanding of the computational principles in these systems is still lacking. A Neuromorphic System for Modeling Axial Locomotion 725 Figure 1: Neuroanatomy of segmented animals. In order to better understand the computational paradigms that mediate intersegmental coordination and the resulting neural control of axial locomotion (and other motor patterns), we are using neuromorphic very large-scale integrated (VLSI) circuits to develop models of these biological systems. The goals in our research are (i) to study how the properties of individual neurons in a network affect the overall system behavior; (ii) to facilitate the validation of the principles underlying intersegmental coordination; and (iii) to develop a real-time, low power, motion control system. We want to exploit these principles and architectures both to improve our understanding of the biology and to design artificial systems that perform autonomously in various environments. Parameter Input Embedded Controller Event Output Address-Event Communication Network " / 12 segments Figure 2: Block-level diagram of the implemented system. The intersegmental communications network facilitates communication among the intrasegmental units with pipelined stages. In this paper, we present a VLSI model of intersegmental coordination as shown in Figure 2. Each segment in our system is implemented with a custom Ie containing a CPG consisting of two silicon model neurons, each one with 16 inhibitory synapses whose values are stored on chip and are continuously variable; an asynchronous address event communications IC that implements the queuing and delaying of events providing synaptic connectivity and thus simulating axonal properties; and a microcontroller (with internal AID converter and timer) that facilitates the modification of individual parameters through a serial bus. The entire system consists of twelve such segments linked to a computer on which a graphical user interface (GUI) is implemented. By using the GUI, we are able to control all of the synaptic connections in the system and to measure the result- G. N Patel, E. A. Brown and S. P. DeWeerth 726 ing neural outputs. We present the system model, and we investigate the role of synaptic coupling in the establishment of phase lags along this chain of neural oscillators. 2 Pattern generating circuits The smallest neural system capable of generating the basic alternating activity that characterizes the swimming ePGs is the half-center oscillator, essentially two bursting neurons with reciprocally inhibitory connections [1] as shown in Figure 3a. In biological systems, the associated neurons have both slow and fast time constants to facilitate the fast spiking (action potentials) and the slower bursting oscillations that control the elicited movements as shown in Figure 3b. To simplify the parameter space of our system, we use reduced two-state silicon neurons [4]. The output of each silicon neuron is an oscillation that represents the envelope of the bursting activity (i.e. the spiking activity and corresponding fast time constants are eliminated) as shown in Figure 3c. Each neuron also has 16 analog synapses that receive off-chip input. The synaptic parameters are stored in an array of floating-gate transistors [5] that provide nonvolatile analog memory. CPG B 1 JW~WID~,J,U~l.v 2 111 11111 1111111 11111 11111111111 11 111 111111 1111 1111 III 1 C 2 A Figure 3: Half-center oscillator and the generation of events in spiking and nonspiking silicon neurons. Events are generated by detecting rapid rises in the membrane potential of spiking neurons or by detecting rapid rises and falls in nonspiking neurons. 3 Intersegmental communication Our segmented system consists of an array of ePG circuits interconnected via an communication network that implements an asynchronous, address-event protocol [6][7]. Each ePG is connected to one node of this address-event intersegmental communication system as illustrated in Figure 2. This application-specific architecture uses a pipelined broadcast scheme that is based upon its biological counterpart. The principal advantage of using this custom scheme is that requisite addresses and delays are generated implicitly based upon the system architecture. In particular the system implements distance-dependent delays and relative addressing. The delays, which are thought to be integral to the network computation, replicate the axonal delays that result as action potentials propagate down an animal's body [2]. The relative addressing greatly simplifies the implementation of synaptic spread [8], the hypothesized translational invariance in the intersegmental connectivity in biological axial locomotion systems. Thus , we can set the synaptic parameters identically at every segment, greatly reducing system complexity. In this architecture (which is described in more depth in [4]), each event is passed from segment to neighboring segment bidirectionally down the length of the one-dimensional 727 A Neuromorphic System for Modeling Axial Locomotion communications network. By delaying each event at every segment, the pipeline architecture facilitates the creation of distance-dependent delays. The other primary advantage of this architecture is that it can easily generate a relative addressing scheme. Figure 4 illustrates the event-passing architecture with respect to the relative addressing and distancedependent delays. Each event, generated at a particular node (the center node, in this example), is transmitted bidirectionally down the length of the network. It is delayed by time /). T at each segment, not including the initiating segment. t=to+2/).T t=to+/).T t=to+/).T t=t o+2/).T = = V '8"'O"'8~8~c;( A -2 A = -1 A= 0 A 1 Figure 4: Relative addressing and distance-dependent delays. The events are generated by the neurons in each segment. Because these are not spiking neurons, we could not use the typical scheme of generating one event per action potential. Instead, we generate one event at the beginning and end of each burst (as illustrated in Figure 3) and designate the individual events as rising or falling . In each segment the events are stored in a queue (Figure 5), which implements delay based upon uniform conduction velocities. As an event arrives at each new segment, it is time stamped, its relative address is incremented (or decremented), and then it is stored in the queue for the /). T interval. As the event exits the queue, its data is decoded by the intrasegmental units, and synaptic inputs are applied to the appropriate intrasegmental neurons. events from rostral segment (closer to head) ..... .. 9: ev ents from events from intrasegmental unit -.. QUEUE (event storage and processing) l... 9: ... .. ... to ro stral and cau dal segments and intrasegmental unit caudal segment (closer to tail) Figure 5: Block-level diagram of a communications node illustrating how events enter and exit each stage of the pipeline. 4 Experiments and Discussion We have implemented the complete system shown in Figure 2, and have performed a number of experiments on the system. In Figure 6, we show the behaviors the system exhibits when it is configured with asymmetrical nearest-neighbor connections. The system displays traveling waves whose directions depend on the direction of the dominant coupling. Note that the intersegmental phase lags vary for different swim frequencies. One important set of experiments focussed on the role of long-distance connections on the system behaviors. In these experiments, we configured the system with strong descending (towards the tail) connections such that robust rearward traveling waves (forward swimming) are observed. The long-distance connections are weak enough to avoid any bifurcations in behavior (different type of behavior). Thus , the traveling wave solution resulting from the nearest-neighbor connections persists as we progressively add long-distance connections. In Figure 7 we show the dependency of the swim frequency and the total phase lag (summation of the normalized intersegmental phase lags, where 1 == 360 0 ) on the extent of the connections. The results show a clear difference in behav- G. N. Patel, E. A. Brown and S. P. DeWeerth 728 stronger descending coupling stronger ascending coupling . . . . . . . . . . .. , c: . ..... 1O&-" 8 1 6 4 0.5 o -0.2 -0.1 2 o 0.1 0.2 time (sec) ~ (l) CI) 1 o.~ ~t=~~~~~ -0.2 -0.1 o 0.1 0.2 time (sec) B 1 0.5 o time (sec) 0.05 o .L.fI==::1.L:;:::1l:::=:,:..:-.J o 0.05 -0.05 time (sec) Figure 6: Traveling waves in the system with asymmetrical, nearest-neighbor connections. Plots are cross-correlations between rising edge events generated by a neuron in segment six and events generated by homolog neurons in each segment. Stronger ascending connections (A & B) produce forward traveling waves (backward swimming) and stronger descending connections (C & D) produce rearward traveling waves (forward swimming) . An externally applied current (lext) controls the swim frequency. At small values of lext (6.7 nA) the periods of the swim cycles are approximately 0.180 ms and 0.150 ms for A & C, respectively; for large values of lext (32.8 nA), the periods of the swim cycles are approximately 36 ms and 33 ms for B & D, respectively. iors between the lowest tonic drive (lext = 21.9 nA) and the two higher tonic drives. (By tonic drive, we mean a constant dc current is applied to all neurons.) In the former, the sensitivity of long-distance connections on frequency and intersegmental phase lags is considerably greater than in the latter. The demarcation in behavior may be attributed to different behaviors at different tonic drives. For lower tonic drive, the long-distance connections tend to synchronize the system (decrease the intersegmental phase lags). At the higher tonic drives, long-distance connections do not affect the system considerably. For lext = 32.8 nA, connections that span up to four segments aid in producing uniformity in the intersegmental phase lags. Although this does not hold for lext = 48.1 nA, long-distance connections playa more significant role in preserving the total phase difference. At lext = 32.8 nA and lext = 48.1 nA, the system with short-distance connections produces a total phase difference of 1.19 and 1.33, respectively. In contrast, for lext = 32.8 nA and lext = 48.1 nA, the system with long-distance connections that span up to seven segments produces a total phase difference of 1.20 and 1.25, respectively. In the above experiments, we have demonstrated that, in a specific parameter regime, weak long-distance connections can affect the intersegmental phase lags. However, these weight profiles should not be construed as a possible explanation on what the weight profiles in a biological system might be. The parameter regime in which we observed this behavior is small; at moderate strengths of coupling, the traveling wave solutions disappear and move towards synchronous behavior. Recent experiments done on spinalized lampreys reveal that long-distance connections are moderately strong [10]. Thus, our CUTrent model is unable to replicate this aspect of intersegmental coordination. There are several explanations that may account for this discrepancy. 729 A Neuromorphic System for Modeling Axial Locomotion 40 r---~-~-~----' B 1.5 .---~-~-~--, A 30 m i)' ~ c .c ~ 20 c. c- -.... 1 ~ O> :? 0.5 10 o~--~----~--~--~ o 2 4 extent 6 8 o'----~----~--~----' o 2 4 6 8 extent Figure 7: Effects of weak long-distance connections on swimming frequency (A), on the total phase difference (summation of the normalized intersegmental phase lags) (B), and on the standard deviation of the intersegmental phase lags (C). 5 < = denote Iext = 48.1 nA, 32.8 nA, and 21 nA, respectively. In the segmental CPG network of the animal, there are many classes of neurons that send projections to many other classes of neurons. The phase a connection imposes is determined by which neuron class connects with which other neuron class. In our system, the segmental CPG network has only a single class of neurons upon which the long-distance connections can impose their phase. Depending on where in parameter space we operate our system, the long-distance connections have too little or too great an effect on the behavior of the system. At high tonic drives, the sensitivity of the weak long-distance connections on the intersegmental phase lags is small, whereas for small tonic drives, the long-distance connections have a great effect on the intersegmental phase lags. It has been shown that if the waveform of the oscillators is sinusoidal (i.e., the time scales of the two state variables are not too different), traveling wave solutions exist and have a large basin of attraction [11]. However, as the disparity between the two time scales is made larger (i.e., the neurons are stiff and the waveform of the oscillations appears square-wave like), the system will move towards synchrony. In our implementation, to facilitate accurate communication of events, we bias the neurons with relatively large differences in the time scales. Thus, this restriction reduces the parameter regime in which we can observe stable traveling waves. Another factor that determines the range of parameters in which stable traveling waves are observed is the slope of our synaptic coupling function. When the slope of the coupling function is steep, the total synaptic current over a cycle can increase significantly, causing weak connections to appear strong. This has an overall effect of synchronizing the network [11]. For coupling functions whose slopes are shallow, the total synaptic current over a cycle is reduced; therefore, the connections appear weak and larger intersegmental phase lags are possible. Thus, the sharp synaptic coupling function in our implementation, which is necessary for communication, is another factor that diminishes the parameter regime in which we can observe stable traveling waves. The above factors limit the parameter range in which we observe traveling waves. However, all of these issues can be addressed by improving our CPG network. The first issue can be addressed by increasing the number of neuron classes or adding more segments. The second and third issues can be addressed by adding spiking neurons in our CPG network so that the form of the oscillations can be coded in the spike train and the synaptic coupling functions can be implemented on the receiving side of the CPG chip. The fourth G. N. Patel, E. A. Brown and S. P. DeWeerth 730 issue can be addressed by designing self-adapting neurons that tune their internal parameters so that their waveforms and intrinsic frequencies are matched. Although weak coupling may not be biologically plausible, producing traveling waves based on phase oscillators would be an interesting research direction. 5 Conclusions and Future Work In this paper, we described a functional, neuromorphic VLSI system that implements an array of neural oscillators interconnected by an address-event communication network. This system represents our most ambitious neuromorphic VLSI effort to date, combining 24 custom ICs, a special-purpose asynchronous communication architecture designed analogously to its biological counterpart, large-scale synaptic interconnectivity with parameters stored using floating-gate devices, and a computer interface for setting the parameters and for measuring the neural activity. The working system represents the culmination of a four-year effort, and now provides a testbed for exploring a variety of biological hypotheses and theoretical predictions. Our future directions in the development of this system are threefold. First, we will continue to explore, in depth, the operation of the present system, comparing it to theoretical predictions and biological hypotheses. Second, we are implementing a segmented mechanical system that will provide a moving output and will facilitate the implementation of sensory feedback. Third, we are developing new CPG model centered around sensory feedback and motor learning. The modular design of the system, which puts all of the neural and synaptic specificity on the CPG IC, allows us to design a completely new CPG and to replace it in the system without changing the communication architecture. References [1] E. Marder & R.L. Calabrese. Principles of rhythmic motor pattern generation. Physiological Reviews 76 (3): 687-717,1996. [2] A.H. Cohen, G.B. Ermentrout, T Kiemel, N. Kopell, K.A. Sigvardt, & TL. Williams. Modeling of intersegmental coordination in the lamprey central pattern generator for locomotion. TINS 15:434-438, 1992. [3] S. Hirose. Biologically Inspired Robots: Snake-like Locomotors and Manipulators. Oxford University Press, 1993. [4] S. DeWeerth, G. Patel, D. Schimmel, M. Simoni, & R.L. Calabrese. A VLSI Architecture for Modeling Intersegmental Coordination. In Proceedings of the Seventeenth Conference on Advanced Research in VLSI, R.B. Brown and A.T Ishii (eds), Los Alamitos, CA: IEEE Computer Society, 182-200, 1997. [5] P. Hasler, B.A. Minch, and C. Diorio. Adaptive circuits using pFet floating-gate devices. In Scott Wills and Stephen DeWeerth editors, 20th Conference of Advanced Research in VLSI, pages 215-230, Los Alamitos, California, CA: IEEE Computer Society, 1999. [6] M.A. Mahowald. VLSI Analogs of Neuronal Visual Processing: A Synthesis of Form and Function. Ph.D. Thesis, California Institute of Technology, Pasadena, CA, 1992. [7] K.A. Boahen. Communicating Neuronal Ensembles between Neuromorphic Chips. Analog Integrated Circuits and Signal Processing, 1997. [8] T Wiliams. Phase Coupling and Synaptic Spread in Chains of Coupled Neuronal Oscillators. Science, vol. 258, 1992. [9] G. Patel. A Neuromorphic Architecture for Modeling Intersegmental Coordination. Ph.D. Thesis, Georgia Institute of Technology, Atlanta, GA, 1999. [10] A. H. Cohen. Personal communication. [11] D. Somers & N. Kopell. Waves and synchrony in networks of oscillators of relaxation and nonrelaxation type. Phyica D, 89:169-183,1995. [12] N. Kopell & G.B. Ermentrout. Coupled oscillators and the design of central pattern generators. Mathematical Biosciences, 90:87-109, 1988.
1720 |@word illustrating:1 rising:2 stronger:4 replicate:2 propagate:1 disparity:1 current:5 timer:1 comparing:1 motor:7 plot:1 designed:1 progressively:1 half:2 device:2 beginning:1 short:1 detecting:2 provides:2 node:4 cpg:11 mathematical:2 along:3 burst:1 consists:3 rostral:1 rapid:2 behavior:13 brain:1 inspired:1 initiating:1 little:1 vertebrate:1 increasing:1 underlying:2 matched:1 circuit:8 lowest:1 what:1 developed:1 thorough:1 every:2 interactive:1 ro:1 control:7 unit:4 appear:2 producing:2 engineering:1 persists:1 limit:1 oxford:1 approximately:2 might:1 studied:2 bursting:3 range:2 seventeenth:1 cau:1 block:2 implement:6 elicit:1 epg:2 thought:2 significantly:1 projection:1 adapting:1 specificity:1 pipelined:2 ga:2 schimmel:1 storage:1 put:1 descending:4 restriction:1 demonstrated:1 center:3 send:1 williams:1 stomatogastric:1 communicating:1 attraction:1 array:3 autonomous:1 user:1 substrate:1 us:1 designing:1 hypothesis:2 locomotion:10 element:1 velocity:1 observed:3 role:3 electrical:1 cord:1 connected:1 cycle:4 autonomously:1 diorio:1 movement:1 incremented:1 decrease:1 leech:3 boahen:1 environment:1 complexity:1 moderately:1 ermentrout:2 personal:1 depend:1 uniformity:1 segment:20 creation:1 upon:4 heartbeat:1 exit:2 basis:1 completely:1 easily:1 chip:4 various:1 train:1 fast:3 describe:1 artificial:1 whose:3 lag:13 larger:2 plausible:1 modular:1 interconnection:2 advantage:2 transistor:1 gait:1 interconnected:2 neighboring:1 causing:1 combining:1 date:1 los:2 produce:5 generating:5 coupling:14 develop:2 depending:1 axial:8 nearest:3 school:1 progress:1 strong:3 implemented:6 direction:4 waveform:3 centered:1 wid:1 implementing:1 biological:10 summation:2 designate:1 exploring:2 hold:1 around:1 ic:3 great:3 vary:1 smallest:1 purpose:1 diminishes:1 intersegmental:23 coordination:9 interconnectivity:1 caudal:1 establishment:1 avoid:1 gatech:3 greatly:2 contrast:1 ishii:1 demarcation:1 dependent:3 snake:2 integrated:2 entire:1 pasadena:1 vlsi:10 overall:2 among:2 translational:1 issue:4 development:1 animal:5 special:1 bifurcation:1 eliminated:1 biology:1 represents:3 synchronizing:1 discrepancy:1 future:2 decremented:1 simplify:1 individual:3 delayed:1 floating:4 phase:21 consisting:1 ents:1 connects:1 kiemel:1 gui:2 atlanta:2 neuroscientist:1 highly:1 investigate:1 custom:3 arrives:1 chain:3 accurate:1 integral:1 capable:2 closer:2 edge:1 necessary:1 dal:1 theoretical:2 modeling:8 measuring:1 neuromorphic:9 mahowald:1 addressing:5 deviation:1 contralateral:1 uniform:1 delay:9 too:3 stored:5 conduction:1 dependency:1 minch:1 considerably:2 twelve:2 sensitivity:2 ie:1 off:1 receiving:1 analogously:1 continuously:1 synthesis:1 na:12 connectivity:3 thesis:2 central:3 containing:1 broadcast:1 account:1 potential:4 sinusoidal:1 stride:1 sec:4 configured:2 coordinated:1 queuing:1 performed:1 microcontroller:1 analyze:1 linked:1 characterizes:1 wave:15 elicited:1 slope:3 synchrony:2 construed:1 square:1 ensemble:1 weak:7 calabrese:2 edgar:1 drive:8 kopell:3 synapsis:3 synaptic:19 ed:1 frequency:6 involved:1 associated:1 attributed:1 modeler:1 bioscience:1 richly:1 crustacean:1 appears:1 rearward:2 higher:2 jw:1 done:1 stage:2 deweerth:6 traveling:13 correlation:1 working:1 reveal:1 manipulator:1 facilitate:4 effect:4 hypothesized:1 brown:5 asymmetrical:2 normalized:2 counterpart:2 former:1 alternating:1 illustrated:2 self:1 m:4 complete:1 motion:1 interface:2 fi:1 functional:1 spiking:6 spinal:1 cohen:2 lext:10 analog:5 tail:2 silicon:5 significant:1 enter:1 moving:1 stable:3 robot:1 add:1 segmental:4 dominant:1 playa:1 recent:2 perspective:1 stiff:1 moderate:1 continue:1 transmitted:1 preserving:1 greater:1 impose:1 neuroanatomy:1 paradigm:1 period:2 signal:1 stephen:2 ii:1 reduces:1 segmented:3 ing:1 cross:1 long:15 serial:1 coded:1 prediction:2 basic:1 controller:1 essentially:1 girish:2 robotics:1 receive:1 addition:1 want:1 whereas:1 interval:1 addressed:4 diagram:2 envelope:1 operate:1 tend:1 facilitates:3 axonal:3 iii:2 identically:1 enough:1 variety:1 affect:3 architecture:12 converter:1 simplifies:1 synchronous:1 six:1 passed:1 swim:5 effort:2 queue:4 passing:1 nonspiking:2 behav:1 action:3 programmable:1 clear:1 tune:1 morris:1 ph:2 reduced:2 generate:2 exist:1 inhibitory:4 per:1 threefold:1 sigvardt:1 vol:1 four:2 falling:1 changing:1 stamped:1 undulatory:1 hasler:1 backward:1 relaxation:1 swimming:7 year:2 fourth:1 somers:1 oscillation:4 culmination:1 display:1 activity:4 strength:1 marder:1 invertebrate:1 aspect:1 span:2 relatively:1 developing:1 pfet:1 membrane:1 shallow:1 modification:1 biologically:2 pipeline:2 bus:1 mechanism:2 ascending:2 end:1 operation:1 experimentation:1 lecar:1 observe:3 appropriate:1 simulating:1 gate:4 slower:1 graphical:1 lamprey:5 exploit:1 giving:1 disappear:1 society:2 move:2 alamitos:2 spike:1 primary:1 exhibit:1 distance:19 unable:1 seven:1 extent:3 toward:1 length:2 providing:1 simoni:1 steep:1 rise:3 design:4 implementation:4 ambitious:1 perform:1 neuron:27 communication:15 delaying:2 head:1 tonic:8 dc:1 arbitrary:1 sharp:1 mechanical:1 connection:30 california:2 testbed:1 address:8 able:1 pattern:10 ev:1 scott:1 regime:4 including:2 memory:1 reciprocally:1 explanation:2 power:1 event:28 synchronize:1 advanced:2 scheme:4 improve:1 technology:3 axis:1 coupled:2 review:2 understanding:3 relative:6 lacking:1 embedded:1 generation:2 interesting:1 generator:3 digital:1 validation:1 illuminating:1 basin:1 imposes:1 principle:4 editor:1 asynchronous:4 bias:1 side:1 understand:1 institute:3 fall:1 neighbor:3 focussed:1 rhythmic:2 distributed:2 feedback:2 depth:2 sensory:3 forward:3 made:3 hirose:1 adaptive:1 iext:1 patel:6 implicitly:1 robust:1 ca:3 improving:1 complex:1 protocol:1 spread:2 mediate:1 profile:2 body:2 neuronal:3 tl:1 georgia:2 slow:1 aid:2 nonvolatile:1 decoded:1 third:2 tin:1 externally:1 down:3 specific:2 physiological:2 intrinsic:1 adding:2 ci:1 illustrates:1 mill:1 explore:1 ganglion:1 bidirectionally:2 visual:1 wiliams:1 determines:1 goal:1 towards:4 oscillator:11 replace:1 typical:1 determined:1 reducing:1 principal:1 total:8 called:1 ece:3 invariance:1 internal:2 iors:1 latter:1 requisite:1 tested:1
788
1,721
v-Arc: Ensemble Learning in the Presence of Outliers t G. Ratsch t , B. Scholkopf1, A. Smola", K.-R. Miillert, T. Onodatt , and S. Mikat t GMD FIRST, Rudower Chaussee 5,12489 Berlin, Germany Microsoft Research, 1 Guildhall Street, Cambridge CB2 3NH, UK * Dep. of Engineering, ANU, Canberra ACT 0200, Australia tt CRIEPI, 2-11-1, Iwado Kita, Komae-shi, Tokyo, Japan {raetsch, klaus, mika}~first.gmd.de,bsc~microsoft.com, Alex.Smola~anu.edu.au,onoda~criepi.denken.or.jp Abstract AdaBoost and other ensemble methods have successfully been applied to a number of classification tasks, seemingly defying problems of overfitting. AdaBoost performs gradient descent in an error function with respect to the margin, asymptotically concentrating on the patterns which are hardest to learn. For very noisy problems, however, this can be disadvantageous. Indeed, theoretical analysis has shown that the margin distribution, as opposed to just the minimal margin, plays a crucial role in understanding this phenomenon. Loosely speaking, some outliers should be tolerated if this has the benefit of substantially increasing the margin on the remaining points. We propose a new boosting algorithm which allows for the possibility of a pre-specified fraction of points to lie in the margin area Or even on the wrong side of the decision boundary. 1 Introduction Boosting and related Ensemble learning methods have been recently used with great success in applications such as Optical Character Recognition (e.g. [8, 16]). The idea of a large minimum margin [17] explains the good generalization performance of AdaBoost in the low noise regime. However, AdaBoost performs worse on noisy tasks [10, 11], such as the iris and the breast cancer benchmark data sets [1]. On the latter tasks, a large margin on all training points cannot be achieved without adverse effects on the generalization error. This experimental observation was supported by the study of [13] where the generalization error of ensemble methods was bounded by the sum of the fraction of training points which have a margin smaller than some value p, say, plus a complexity term depending on the base hypotheses and p. While this bound can only capture part of what is going on in practice, it nevertheless already conveys the message that in some cases it pays to allow for some points which have a small margin, or are misclassified, if this leads to a larger overall margin on the remaining points. To cope with this problem, it was mandatory to construct regularized variants of AdaBoost, which traded off the number of margin errors and the size of the margin 562 G. Riitsch, B. Sch6lkopf, A. J. Smola, K.-R. Muller, T. Onoda and S. Mika [9, 11]. This goal, however, had so far been achieved in a heuristic way by introducing regularization parameters which have no immediate interpretation and which cannot be adjusted easily. The present paper addresses this problem in two ways. Primarily, it makes an algorithmic contribution to the problem of constructing regularized boosting algorithms. However, compared to the previous efforts, it parameterizes the above trade-off in a much more intuitive way: its only free parameter directly determines the fraction of margin errors. This, in turn, is also appealing from a theoretical point of view, since it involves a parameter which controls a quantity that plays a crucial role in the generalization error bounds (cf. also [9, 13]). Furthermore, it allows the user to roughly specify this parameter once a reasonable estimate of the expected error (possibly from other studies) can be obtained, thus reducing the training time. 2 Boosting and the Linear Programming Solution Before deriving a new algorithm, we briefly discuss the properties of the solution generated by standard AdaBoost and, closely related, Arc-GV (2], and show the relation to a linear programming (LP) solution over the class of base hypotheses G. Let {gt(x) : t = 1, ... ,T} be a sequence of hypotheses and a = [al ... aT] their weights satisfying at ~ O. The hypotheses gt are elements of a hypotheses class G = {g: x 14 [-1, In, which is defined by a base learning algorithm. The ensemble generates the label which is the weighted majority of the votes by sign(f(x)) In order to express that of notation we define where f(x) = ~ lI:ill gt(x). (1) f and therefore also the margin p depend on a and for ease p(z, a) := yf(x) where z := (x, y) and f is defined as in (1). (2) Likewise we use the normalized margin: p(a):= min P(Zi, a) , l~t~m (3) Ensemble learning methods have to find both, the hypotheses gt E G used for the combination and their weights a. In the following we will consider only AdaBoost algorithms (including Arcing). For more details see e.g. (4, 2]. The main idea of AdaBoost is to introduce weights Wt(Zi) on the training patterns. They are used to control the importance of each single pattern in learning a new hypothesis (Le. while repeatedly running the base algorithm). Training patterns that are difficult to learn (which are misclassified repeatedly) become more important. The minimization objective of AdaBoost can be expressed in terms of margins as m (4) i=1 In every iteration, AdaBoost tries to minimize this error by a stepwise maximization of the margin. It is widely believed that AdaBoost tries to maximize the smallest margin on the training set [2, 5, 3, 13, 11]. Strictly speaking, however, a general proof is missing. It would imply that AdaBoost asymptotically approximates (up to scaling) the solution of the following linear programming problem over the complete hypothesis set G (cf. [7], assuming a finite number of basis hypotheses): maximize subject to p p( Zi, a) ~ p for all 1 < i < m at, P ~ 0 for all 1 ~ t ~ IGI lIalil = 1 (5) 563 v-Arc: Ensemble Learning in the Presence o/Outliers Since such a linear program cannot be solved exactly for a infinite hypothesis set in general, it is interesting to analyze approximation algorithms for this kind of problems. Breiman [2] proposed a modification of AdaBoost - Arc-GV - making it possible to show the asymptotic convergence of p(a t ) to the global solution pIP: Theorem 1 (Breiman [2]). Choose at in each iteration as at := argmin Lexp [-llatlll (p(Zi' at) - p(a t - I ))], aE[O,I] (6) i and assume that the base learner always finds the hypothesis 9 E G which minimizes the weighted training error with respect to the weights. Then lim p( at) = pIp. t-HX> Note that the algorithm above can be derived from the modified error function 9 g v(a t ):= Lexp [-llatlll (p(Zi' at) - p(a t - I ))]. (7) The question one might ask now is whether to use AdaBoost or rather Arc-GV in practice. Does Arc-GV converge fast enough to benefit from its asymptotic properties? In [12] we conducted experiments to investigate this question. We empirically found that (a) AdaBoost has problems finding the optimal combination if pIp < 0, (b) Arc-GV's convergence does not depend on pIp, and (c) for pIp> 0, AdaBoost usually converges to the maximum margin solution slightly faster than Arc-GV. Observation (a) becomes clear from (4): 9(a) will not converge to and lIal11 can be bounded by some value. Thus the asymptotic case cannot be reached, whereas for Arc-GV the optimum is always found. Moreover, the number of iterations necessary to converge to a good solution seems to be reasonable, but for a near optimal solution the number of iterations is rather high. This implies that for real world hypothesis sets, the number of iterations needed to find an almost optimal solution can become prohibitive, but we conjecture that in practice a reasonably good approximation to the optimum is provided by both AdaBoost and Arc-GV. 3 v- Algorithms For the LP-AdaBoost [7] approach it has been shown for noisy problems that the generalization performance is usually not as good as the one of AdaBoost [7, 2, 11]. From Theorem 5 in [13] (cf. Theorem 3 on page 5) this fact becomes clear, as the minimum of the right hand side of inequality (cf. (13)) need not necessarily be achieved with a maximum margin. We now propose an algorithm to directly control the number of margin errors and therefore also the contribution of both terms in the inequality separately (cf. Theorem 3). We first consider a small hypothesis class and end up with a linear program - v-LP-AdaBoost. In subsection 3.2 we then combine this algorithm with the ideas from section 2 and get a new algorithm v-Arc - which approximates the v-LP solution. ? 3.1 v-LP-AdaBoost Let us consider the case where we are given a (finite) set G = {g: x I-t [-1, 1n ofT hypotheses. To find the coefficients a for the combined hypothesis f(x) we extend the LP-AdaBoost algorithm [7, 11] by incorporating the parameter v [15] and solve the following linear optimization problem: maximize pE::'I ~i subject to P(Zi' a) ::::: p - ~i for all 1 ~ i ~ m (8) ~i' at, P ::::: for all 1 ~ t ~ T and 1 ~ i ~ m lIalh = 1 v!n ? G. Riitsch, B. SchOlkopf, A. J. Smola, K.-R. Muller, T. Onoda and S. Mika 564 This algorithm does not force all margins to be beyond zero and we get a soft margin classification (cf. SVMs) with a regularization constant The following proposition shows that v has an immediate interpretation: Proposition 2 (Ratsch et al. [12]). Suppose we run the algorithm given in (8) on some data with the resulting optimal P > o. Then v!n. 1. v upper bounds the fraction of margin errors. 2. 1 - v upper bounds the fraction of patterns with margin larger than p. Since the slack variables ~i only enter the cost function linearly, their absolute size is not important. Loosely speaking, this is due to the fact that for the optimum of the primal objective function, only derivatives wrt. the primal variables matter, and the derivative of a linear function is constant. In the case of SVMs [14], where the hypotheses can be thought of as vectors in some feature space, this statement can be translated into a precise rule for distorting training patterns without changing the solution: we can move them locally orthogonal to a separating hyperplane. This yields a desirable robustness property. Note that the algorithm essentially depends on the number of outliers, not on the size of the error [15]. 3.2 The v-Arc Algorithm Suppose we have a very large (but finite) base hypothesis class G. Then it is difficult to solve (8) as (5) directly. To this end, we propose a new algorithm - v-Arc - that approximates the solution of (8). The optimal p for fixed margins P(Zi' a) in (8) can be written as a?+) . (9) P(Zi' a?+ and subtracting pv(a) := argmax (p - _1 f)p - p(Zi' pE[O,I] vm i=1 where (~)+ := max(~, 0). Setting ~i := (pv(a) I:~l ~i from the resulting inequality on both sides yields (for all 1 ~ i ~ m) v!n P(Zi' a) 1 + ~i - - vm m L~i i=1 ~ 1 pv(a) - vm m L~i . i=1 (10) Two more substitutions are needed to transform the problem into one which can be solved by the AdaBoost algorithm. In particular we have to get rid of the slack variables ~i again by absorbing them into quantities similar to P(Zi' a) and p(a). This works as follows: on the right hand side of (10) we have the objective function (cf. (8? and on the left hand side a term that depends nonlinearly on a. Defining _ pv(a) 1 := pv(a) - - m L ~i vm. _ and Pv(Zi' a) := P(Zi' a) + ~i - 1 - m '"' vm~ ,=1 ~i, (11) i=l which we substitute for p(a) and p(z,a) in (5), respectively, we obtain a new optimization problem. Note that ,ov (a) and ,ov (Zi' a) play the role of a corrected or virtual margin. We obtain a nonlinear min-max problem maximize subject to ,o( Zi, a) ~ ,o(a) ,o(a) at lIallt for all 1 ~ i > 0 for all 1 ~ t ~1 ~ ~ m T ' (12) which Arc-GV can solve approximately (cf. section 2). Hence, by replacing the margin p(Z, a) by ,o(z,a) in equation (4) and the other formulas for Arc-GV (cf. [2]), 565 v-Arc: Ensemble Learning in the Presence o/Outliers we obtain a new algorithm which we refer to as v-Arc. We can now state interesting properties for v-Arc by using Theorem 5 of [13] that bounds the generalization error R(f) for ensemble methods. In our case Rp(f) ~ v by construction (i.e. the number of patterns with a margin smaller than p, cf. Proposition 2), thus we get the following simple reformulation of this bound: Theorem 3. Let p(x, y) be a distribution over X x [-1,1]' and let X be a sample of m examples chosen iid according to p. Suppose the base-hypothesis space G has VC dimension h, and let [) > 0. Then with probability at least 1 - [) over the random choice of the training set X, Y, every function f generated by v-Arc, where v E (0,1) and pv > 0, satisfies the following bound: R(f) ~ v + ~ l h) (hIO g2 (m 2 Pv m + Iog (!)) . ~ u (13) So, for minimizing the right hand side we can tradeoff between the first and the second term by controlling an easily interpretable regularization parameter v. 4 Experiments We show a set of toy experiments to illustrate the general behavior of v-Arc. As base hypothesis class G we use the RBF networks of [11], and as data a two-class problem generated from several 2D Gauss blobs (cf. Banana shape dataset from http://www.first.gmd.derdata/banana .html.). We obtain the following results: ? v-Arc leads to approximately vm patterns that are effectively used in the training of the base learner: Figure 1 (left) shows the fraction of patterns that have high average weights during the learning process (i.e. Ei=l Wt(Zi) > 112m). We find that the number of the latter increases (almost) linearly with v. This follows from (11) as the (soft) margin of patterns with p(z, a) < Pv is set to pv and the weight of those patterns will be the same. ? The (estimated) test error, averaged over 10 training sets, exhibits a rather flat minimum in v (Figure 1 (lower)). This indicates that just as for vSVMs, where corresponding results have been obtained, v is a well-behaved parameter in the sense that a slight misadjustment it is not harmful. ? v-Arc leads to the fraction v of margin errors (cf. dashed line in Figure 1) exactly as predicted in Proposition 2. ? Finally, a good value of v can already be inferred from prior knowledge of the expected error. Setting it to a value similar to the latter provides a good starting point for further optimization (cf. Theorem 3). Note that for v = 1, we recover the Bagging algorithm (if we used bootstrap samples), as the weights of all patterns will be the same (Wt(Zi) = 11m for all i = 1, . . . ,m) and also the hypothesis weights will be constant (at'" liT for all t = 1, .. . ,T) . Finally, we present a small comparison on ten benchmark data sets obtained from the VCI [1] benchmark repository (cf. http://ida.first.gmd.de/-raetsch/data/benchmarks.html). We analyze the performance of single RBF networks, AdaBoost, v-Arc and RBF-SVMs. For AdaBoost and v-Arc we use RBF networks [11] as base hypothesis. The model parameters of RBF (number of centers etc.), v-Arc (v) and SVMs (0', C) are optimized using 5-fold cross-validation. More details on the experimental setup can 566 0.8 ~ "- 0.16 number of important panems ~ ~ G. Riitsch, B. SchO/kopf, A. J. Smola, K.-R. Muller, T. Onoda and S. Mika 0. 15 Arc-GV ~ 0 .6 w 0. '4 Bagging '0 c ~ 0 .4 training error ~ 0.12 0.11 o o. I 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 o 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Figure 1: Toy experim~nt (0' = 0): the left graph shows the average Yfraction of important patterns, the avo fraction of margin errors and the avo training error for different values of the regularization constant v for v-Arc. The right graph shows the corresponding generalization error. In both cases, the parameter v allows us to reduce the test errors to values much lower than for the hard margin algorithm (for v = 0 we recover ArcGV / AdaBoost, and for v = 1 we get Bagging.) be found in [11]. Fig. 1 shows the generalization error estimates (after averaging over 100 realizations of the data sets) and the confidence interval. The results of the best classifier and the classifiers that are not significantly worse are set in bold face . To test the significance, we used at-test (p = 80%). On eight out of the ten data sets, v-Arc performs significantly better than AdaBoost. This clearly shows the superior performance of v-Arc for noisy data sets and supports this soft margin approach for AdaBoost. Furthermore, we find comparable performances for v-Arc and SVMs. In three cases the SVM performs better and in two cases v-Arc performs best. Summarizing, AdaBoost is useful for low noise cases, where the classes are separable. v-Arc extends the applicability of boosting to problems that are difficult to separate and should be applied if the data are noisy. 5 Conclusion We analyzed the AdaBoost algorithm and found that Arc-GV and AdaBoost are efficient for approximating the solution of non-linear min-max problems over huge hypothesis classes. We re-parameterized the LP Reg-AdaBoost algorithm (cf. [7, 11]) and introduced a new regularization constant v that controls the fraction of patterns inside the margin area. The new parameter is highly intuitive and has to be optimized only on a fixed interval [0,1] . Using the fact that Arc-GV can approximately solve min-max problems, we found a formulation of Arc-G V - v-Arc - that implements the v-idea for Boosting by defining an appropriate soft margin. The present paper extends previous work on regularizing boosting (DOOM [9], AdaBoostReg [11]) and shows the utility and flexibility of the soft margin approach for AdaBoost. Banana B.Cancer Diabetes German Heart Ringnorm F .Sonar Thyroid Titanic Waveform RBF 10.8 ? 0.06 27.6 ? 0.47 24.3 ? 0.19 24.7 ? 0.24 17.6 ? 0.33 1.7 ? 0.02 34.4 ? 0.20 4.5 ? 0.21 23.3 ? 0.13 10.7 ? 0.11 AB 12.3 ? 0.07 30.4 ? 0.47 26.5 ? 0.23 27.5 ? 0.25 20.3 ? 0.34 1.9 ? 0.03 35.7 ? 0.18 4.4 ? 0.22 22.6 ? 0.12 10.8 ? 0.06 v-Arc 10.6 ? 0.05 25.8 ? 0.46 23.7 ? 0.20 24.4 ? 0.22 16.5 ? 0.36 1.7 ? 0.02 34.4 ? 0.19 4.4 ? 0.22 23.0 ? 0.14 10.0 ? 0.07 SVM 11.5 ? 0.07 26.0 ? 0.47 23.5 ? 0.17 23.6 ? 0.21 16.0 ? 0.33 1.7 ? 0.01 32.4 ? 0.18 4.8 ? 0.22 22.4 ? 0.10 9.9 ? 0.04 Table 1: Generalization error estimates and confidence intervals. The best classifiers for a particular data set are marked in bold face (see text). v-Arc: Ensemble Learning in the Presence of Outliers 567 We found empirically that the generalization performance in v-Arc depends only slightly on the choice of the regularization constant. This makes model selection (e.g. via cross-validation) easier and faster. Future work will study the detailed regularization properties of the regularized versions of AdaBoost, in particular in comparison to v-LP Support Vector Machines . Acknowledgments: Partial funding from DFG grant (Ja 379/52) is gratefully acknowledged. This work was done while AS and BS were at GMD FIRST. References [1] C. Blake, E. Keogh, and C. J. Merz. UCI repository of machine learning databases, 1998. http://www.ics.uci.edu/ ",mlearn/MLRepository.html. [2] L. Breiman. Prediction games and arcing algorithms. Technical Report 504, Statistics Department, University of California, December 1997. [3] M. Frean and T. Downs. A simple cost function for boosting. Technical report, Dept. of Computer Science and Electrical Eng., University of Queensland, 1998. [4] Y. Freund and R. E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. In Computational Learning Theory: Eurocolt '95, pages 23-37. Springer-Verlag, 1995. [5] Y. Freund and R. E . Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. J. of Compo fj Syst. Sc. , 55(1):119- 139, 1997. [6] J . Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: a statistical view of boosting. Technical report, Stanford University, 1998. [7) A. Grove and D. Schuurmans. Boosting in the limit: Maximizing the margin of learned ensembles. In Proc. of the 15th Nat. Conf. on AI, pages 692- 699 , 1998. [8] Y. LeCun, L. D . Jackel, L. Bottou, C. Cortes, J . S. Denker, H. Drucker, I. Guyon, U. A. Muller, E. Sackinger, P. Simard, and V. Vapnik. Learning algorithms for classification: A comparison on handwritten digit recognition. Neural Networks, pages 261-276, 1995. [9) L. Mason, P. L. Bartlett, and J. Baxter. Improved generalization through explicit optimization of margins. Machine Learning, 1999. to appear. (10) J. R. Quinlan. Boosting first-order learning (invited lecture). Lecture Notes in Computer Science, 1160:143, 1996. (11) G. Ratsch , T. Onoda, and K.-R. Muller. Soft margins for AdaBoost. Technical Report NC-TR-1998-021, Department of Computer Science, Royal Holloway, University of London, Egham, UK , 1998. To appear in Machine Learning. (12) G. Ratsch, B. Schokopf, A. Smola, S. Mika, T. Onoda, and K.-R. Muller. Robust ensemble learning. In A.J . Smola, P.L. Bartlett, B. Scholkopf, and D. Schuurmans, editors, Advances in LMC, pages 207-219. MIT Press, Cambridge, MA , 1999. [13] R. Schapire, Y. Freund, P. L. Bartlett, and W . Sun Lee. Boosting the margin: A new explanation for the effectiveness of voting methods. Annals of Statistics, 1998. (Earlier appeared in: D. H. Fisher, Jr. (ed.), Proc. ICML97, M. Kaufmann). [14] B. Scholkopf, C. J. C. Burges, and A. J. Smola. Advances in Kernel Methods Support Vector Learning. MIT Press, Cambridge, MA, 1999. (15) B. Scholkopf, A. Smola, R. C. Williamson, and P. L. Bartlett. New support vector algorithms. Neural Computation, 12:1083 - 1121, 2000. (16) H. Schwenk and Y. Bengio. Training methods for adaptive boosting of neural networks. In Michael I. Jordan, Michael J. Kearns, and Sara A. Solla, editors, Advances in Neural Inf. Processing Systems, volume 10. The MIT Press, 1998. [17) V. Vapnik. The Nature of Statistical Learning Theory. Springer Verlag, New York, 1995.
1721 |@word repository:2 version:1 briefly:1 seems:1 queensland:1 eng:1 tr:1 substitution:1 riitsch:3 com:1 ida:1 nt:1 written:1 additive:1 shape:1 gv:13 interpretable:1 prohibitive:1 compo:1 provides:1 boosting:15 become:2 scholkopf:4 combine:1 inside:1 introduce:1 expected:2 indeed:1 behavior:1 roughly:1 eurocolt:1 increasing:1 becomes:2 provided:1 bounded:2 notation:1 moreover:1 what:1 kind:1 argmin:1 substantially:1 minimizes:1 finding:1 every:2 act:1 voting:1 exactly:2 wrong:1 classifier:3 uk:2 control:4 grant:1 appear:2 before:1 engineering:1 limit:1 approximately:3 mika:5 plus:1 might:1 au:1 sara:1 ringnorm:1 ease:1 averaged:1 acknowledgment:1 lecun:1 practice:3 implement:1 cb2:1 bootstrap:1 digit:1 area:2 thought:1 significantly:2 pre:1 confidence:2 get:5 cannot:4 selection:1 doom:1 www:2 shi:1 missing:1 center:1 maximizing:1 starting:1 rule:1 deriving:1 annals:1 construction:1 play:3 suppose:3 user:1 controlling:1 programming:3 hypothesis:22 diabetes:1 element:1 recognition:2 satisfying:1 database:1 role:3 solved:2 capture:1 electrical:1 sun:1 solla:1 trade:1 complexity:1 rudower:1 depend:2 ov:2 learner:2 basis:1 translated:1 easily:2 schwenk:1 fast:1 london:1 sc:1 klaus:1 heuristic:1 larger:2 widely:1 solve:4 say:1 stanford:1 statistic:2 transform:1 noisy:5 seemingly:1 sequence:1 blob:1 propose:3 subtracting:1 uci:2 realization:1 flexibility:1 intuitive:2 convergence:2 optimum:3 converges:1 depending:1 illustrate:1 frean:1 dep:1 predicted:1 involves:1 implies:1 waveform:1 closely:1 tokyo:1 vc:1 australia:1 virtual:1 explains:1 ja:1 hx:1 generalization:13 proposition:4 keogh:1 adjusted:1 strictly:1 blake:1 ic:1 guildhall:1 great:1 algorithmic:1 traded:1 smallest:1 proc:2 label:1 jackel:1 successfully:1 weighted:2 minimization:1 bsc:1 clearly:1 mit:3 always:2 modified:1 rather:3 breiman:3 pip:5 arcing:2 derived:1 indicates:1 sense:1 summarizing:1 relation:1 going:1 misclassified:2 germany:1 overall:1 classification:3 ill:1 html:3 construct:1 once:1 lit:1 hardest:1 future:1 report:4 primarily:1 dfg:1 argmax:1 microsoft:2 ab:1 friedman:1 huge:1 message:1 possibility:1 investigate:1 highly:1 analyzed:1 primal:2 grove:1 partial:1 necessary:1 orthogonal:1 loosely:2 harmful:1 re:1 theoretical:2 minimal:1 soft:6 earlier:1 maximization:1 cost:2 introducing:1 applicability:1 conducted:1 tolerated:1 combined:1 lee:1 off:2 vm:6 michael:2 again:1 opposed:1 choose:1 possibly:1 worse:2 conf:1 derivative:2 simard:1 criepi:2 li:1 japan:1 toy:2 syst:1 de:2 bold:2 coefficient:1 matter:1 igi:1 depends:3 view:2 try:2 analyze:2 reached:1 disadvantageous:1 recover:2 contribution:2 minimize:1 kaufmann:1 likewise:1 ensemble:12 yield:2 handwritten:1 iid:1 mlearn:1 ed:1 conveys:1 proof:1 dataset:1 concentrating:1 ask:1 lim:1 subsection:1 knowledge:1 adaboost:35 specify:1 improved:1 formulation:1 done:1 furthermore:2 just:2 smola:9 hand:4 replacing:1 ei:1 nonlinear:1 vci:1 sackinger:1 logistic:1 yf:1 behaved:1 effect:1 normalized:1 regularization:7 hence:1 avo:2 during:1 game:1 lmc:1 mlrepository:1 iris:1 tt:1 complete:1 theoretic:2 performs:5 kopf:1 fj:1 recently:1 scho:1 funding:1 superior:1 absorbing:1 empirically:2 jp:1 nh:1 volume:1 extend:1 interpretation:2 approximates:3 slight:1 raetsch:2 refer:1 cambridge:3 enter:1 ai:1 gratefully:1 had:1 gt:4 base:10 etc:1 inf:1 mandatory:1 verlag:2 inequality:3 success:1 muller:6 misadjustment:1 minimum:3 converge:3 maximize:4 dashed:1 desirable:1 technical:4 faster:2 believed:1 cross:2 iog:1 prediction:1 variant:1 regression:1 breast:1 ae:1 essentially:1 iteration:5 kernel:1 achieved:3 whereas:1 chaussee:1 separately:1 interval:3 ratsch:4 crucial:2 invited:1 subject:3 december:1 effectiveness:1 jordan:1 near:1 presence:4 bengio:1 enough:1 baxter:1 zi:17 hastie:1 reduce:1 idea:4 parameterizes:1 tradeoff:1 drucker:1 whether:1 distorting:1 utility:1 bartlett:4 effort:1 speaking:3 york:1 repeatedly:2 useful:1 clear:2 detailed:1 locally:1 ten:2 svms:5 gmd:5 http:3 schapire:3 sign:1 estimated:1 tibshirani:1 express:1 reformulation:1 nevertheless:1 acknowledged:1 changing:1 asymptotically:2 graph:2 fraction:9 sum:1 run:1 parameterized:1 extends:2 almost:2 reasonable:2 guyon:1 kita:1 decision:3 scaling:1 comparable:1 bound:7 pay:1 fold:1 alex:1 flat:1 generates:1 thyroid:1 min:4 separable:1 optical:1 conjecture:1 department:2 according:1 combination:2 jr:1 smaller:2 slightly:2 character:1 appealing:1 lp:8 modification:1 making:1 b:1 hio:1 outlier:6 heart:1 equation:1 turn:1 discus:1 lexp:2 slack:2 needed:2 wrt:1 german:1 end:2 experim:1 eight:1 denker:1 appropriate:1 egham:1 robustness:1 rp:1 substitute:1 bagging:3 remaining:2 cf:15 running:1 quinlan:1 approximating:1 objective:3 move:1 already:2 quantity:2 question:2 exhibit:1 gradient:1 separate:1 berlin:1 separating:1 street:1 majority:1 assuming:1 minimizing:1 nc:1 difficult:3 setup:1 statement:1 sch6lkopf:1 upper:2 observation:2 arc:39 benchmark:4 finite:3 descent:1 immediate:2 defining:2 precise:1 banana:3 inferred:1 introduced:1 nonlinearly:1 specified:1 optimized:2 california:1 learned:1 address:1 beyond:1 usually:2 pattern:14 regime:1 oft:1 appeared:1 program:2 including:1 max:4 royal:1 explanation:1 force:1 regularized:3 imply:1 titanic:1 text:1 prior:1 understanding:1 asymptotic:3 freund:3 lecture:2 interesting:2 validation:2 editor:2 cancer:2 supported:1 free:1 adaboostreg:1 side:6 allow:1 burges:1 face:2 absolute:1 benefit:2 boundary:1 dimension:1 world:1 adaptive:1 far:1 cope:1 global:1 overfitting:1 rid:1 sonar:1 table:1 nature:1 learn:2 onoda:6 reasonably:1 robust:1 schuurmans:2 williamson:1 bottou:1 necessarily:1 constructing:1 significance:1 main:1 linearly:2 noise:2 fig:1 canberra:1 pv:10 explicit:1 lie:1 pe:2 theorem:7 formula:1 down:1 mason:1 svm:2 cortes:1 incorporating:1 stepwise:1 vapnik:2 effectively:1 importance:1 nat:1 anu:2 margin:41 easier:1 expressed:1 g2:1 springer:2 determines:1 satisfies:1 ma:2 goal:1 marked:1 rbf:6 fisher:1 adverse:1 hard:1 infinite:1 reducing:1 corrected:1 wt:3 hyperplane:1 averaging:1 kearns:1 experimental:2 gauss:1 merz:1 vote:1 holloway:1 support:4 latter:3 dept:1 reg:1 regularizing:1 phenomenon:1
789
1,722
Bayesian model selection for Support Vector machines, Gaussian processes and other kernel classifiers Matthias Seeger Institute for Adaptive and Neural Computation University of Edinburgh 5 Forrest Hill, Edinburgh EHI 2QL seeger@dai.ed.ac.uk Abstract We present a variational Bayesian method for model selection over families of kernels classifiers like Support Vector machines or Gaussian processes. The algorithm needs no user interaction and is able to adapt a large number of kernel parameters to given data without having to sacrifice training cases for validation. This opens the possibility to use sophisticated families of kernels in situations where the small "standard kernel" classes are clearly inappropriate. We relate the method to other work done on Gaussian processes and clarify the relation between Support Vector machines and certain Gaussian process models. 1 Introduction Bayesian techniques have been widely and successfully used in the neural networks and statistics community and are appealing because of their conceptual simplicity, generality and consistency with which they solve learning problems. In this paper we present a new method for applying the Bayesian methodology to Support Vector machines. We will briefly review Gaussian Process and Support Vector classification in this section and clarify their relationship by pointing out the common roots. Although we focus on classification here, it is straightforward to apply the methods to regression problems as well. In section 2 we introduce our algorithm and show relations to existing methods. Finally, we present experimental results in section 3 and close with a discussion in section 4. Let X be a measure space (e.g. X = ~d) and D = (X,t) = {(Xl,tt), ... , (Xn,t n )), Xi E X, ti E {-l,+l} a noisy LLd. sample from a latent function y : X -+ lR, where P(tly) denotes the noise distribution. Given further points X. we wish to predict t. so as to minimize the error probability P(tlx., D), or (more difficult) to estimate this probability. Generative Bayesian methods attack this problem by placing a stochastic process prior P(y(?)) over the space of latent functions and 604 M Seeger then compute posterior and predictive distributions P(yID), P(y.lx., D) as P(yID) = P(Dly)P(y) P(D) , P(y.ID,x.) = f (1) P(y.ly)P(yID) dy where y = (y(Xi?i, y. = y(x.), the likelihood P(Dly) = TIi P(tiIYi) and P(D) is a normalization constant. P(tlx., D) can then be obtained by averaging P(tly.) over P(y.lx., D). Gaussian process (GP) or spline smoothing models use a Gaussian process prior on y(.) which can be seen as function of X into a set of random variables such that for each finite XI C X the corresponding variables are jointly Gaussian (see [15] for an introduction). A GP is determined by a mean function 1 x 1-4 E[y(x)] and a positive definite covariance kernel K(x,x). Gaussian process classification (GPC) amounts to specifying available prior knowledge by choosing a class of kernels K(x, xIO), 0 E e, where 0 is a vector of hyperparameters, and a hyperprior P(O). Usually, these choices are guided by simple attributes of y(.) such as smoothness, trends, differentiability, but more general approaches to kernel design have also been considered [5]. For 2-class classification the most common noise (1 + exp( _U?-1 the distribution is the binomial one where P(tly) (j(ty), (j(u) logistic function, and y is the logit 10g(P( +llx)/ P( -llx? of the target distribution. For this noise model the integral in (1) is not analytically tractable, but a range of approximative techniques based on Laplace approximations [16], Markov chain Monte Carlo [7], variational methods [2] or mean field algorithms [8] are known. = = We follow [16]. The Laplace approach to GPC is to approximate the posterior P(yID,O) by the Gaussian distribution N(y, 1i- 1) where y = argmaxP(yID, 0) is the posterior mode and 1i = \7~\7y(-logP(YID,O?, evaluated at y. Then it is easy to show that the predictive distribution is Gaussian with mean k(x.),K-l y and variance k. - k(x.)'K- 1k(x.) where K is the covariance matrix (K(Xi,Xj?ij, k(?) = (K(Xi, '?i, k. = K(x., x.) and the prime denotes transposition. The final discriminant is therefore a linear combination of the K (Xi, .). The discriminative approach to the prediction problem is to choose a loss function get, y), being an approximation to the misclassification loss2 I{tY:5o} and then to search for a discriminant y(.) which minimizes E [get, y(x.?] for the points x. of interest (see [14]). Support Vector classification (SVC) uses the c-insensitive loss (SVC loss) get, y) = [1 - ty]+, [u]+ = uI{u~o} which is an upper bound on the misclassification loss, and a reproducing kernel Hilbert space (RKHS) with kernel K(x,xIO) as hypothesis space for y(.). Indeed, Support Vector models and the Laplace method for Gaussian processes are special cases of spline smoothing models in RKHS where the aim is to minimize the functional n ~9(ti'Yi) + AllyOIl~ (2) i=l where II . 11K denotes the norm of the RKHS. It can be shown that the minimizer of (2) can be written as k(?), K-1y where y maximizes n - ~9(ti'Yi) - Ay'K-ly. (3) i=l All these facts can be found in [13]. Now (3) is, up to terms not depending on y, the log posterior in the above GP framework if we choose g(t,y) = -logP(tly) and lW.l.O.g. we only consider GPs with mean function 0 in what follows. denotes the indicator function of the set A c lR.. 2IA 605 Bayesian Model Selection for Support Vector Machines absorb A into O. For the SVC loss, (3) can be transformed into a dual problem via y = Ka, where a is a vector of dual variables, which can be efficiently solved using quadratic programming techniques. [12] is an excellent reference. Note that the SVC loss cannot be written as the negative log of a noise distribution, so we cannot reduce SVC to a special case of a Gaussian process classification model. Although a generative model for SVC is given in [11], it is easier and less problematic to regard SVC as efficient approximation to a proper Gaussian process model. Various such models have been proposed (see [8],[4]). In this work, we simply normalize the SVC loss pointwise, i.e. use a Gaussian process model with the normalized BVe loss g(t, y) = [1 - ty]+ + log Z(y), Z(y) = exp( -[1 - y]+) + exp( -[1 + y]+). Note that g(t, y) is a close approximation of the (unnormalized) SVC loss. The reader might miss the SVM bias parameter which we dropped here for clarity, but it is straightforward to apply this semiparametric extension to GP models to0 3 . 2 A variational method for kernel classification The real Bayesian way to deal with the hyperparameters 0 is to average P(y.lx., D, 0) over the posterior P( OlD) in order to obtain the predictive distribution P(y.lx., D). This can be approximated by Markov chain Monte Carlo methods [7], [16] or simply by P(y.lx.,D,9), 9 = argmaxP(OID). The latter approach, called maximum a-posteriori (MAP), can be justified in the limit of large n and often works well in practice. The basic challenge of MAP is to calculate the evidence P(DI9) = ! P(D,yI9)dy = ! exp (- t.9(ti,Yi?) N(yIO,K(9))dy. (4) Our plan is to attack (4) by a variational approach. Let P be a density from a model class r chosen to approximate the posterior P(yID, 0). Then: -logP(DIO) = -JP(Y)lOg (P(D'YIO)~(Y)) P(y ID, O)P(y) - !- = F(P, 0) - dy (5) P (ID, y) P(y) log ( P(y 0)) dy where we call F(P, 0) = Ep[-log P(D, yIO)] +Ep[logP(y)] the variational free energy. The second term in (5) is the well-known Kullback-Leibler divergence between P and the posterior which is nonnegative and equals zero iff P(y) = P(yID,O) almost everywhere with respect to the distribution P. Thus, F is an upper bound on - log P (D I0), and changing (P, 0) to decrease F enlarges the evidence or decreases the divergence between the posterior and its approximation, both being favourable. This idea has been introduced in [3] as ensemble learning4 and has been successfully applied to MLPs [1]. The latter work also introduced the model class r we use here, namely the class of Gaussians with mean IL and factor-analyzed covariance ~ = V + L,~1 Cjcj, V diagonal with positive elements 5 . Hinton and 3This is the "random effects model with improper prior" of [13], p.19, and works by placing a flat improper prior on the bias parameter. 4We average different discriminants (given by y) over the ensemble P. 5 Although there is no danger of overfitting, the use of full covariances would render the optimization more difficult, time and memory consuming. M Seeger 606 van Camp [3] used diagonal covariances which would be M = 0 in our setting. By choosing a small M, we are able to track the most important correlations between the components in the posterior using O( M n) parameters to represent P. Having agreed on r, the criterion F and its gradients with respect to (J and the parameters of P can easily and efficiently be computed except for the generic term (6) a sum of one-dimensional Gaussian expectations which are, depending on the actual g, either analytically tractable or can be approximated using a quadrature algorithm. For example, the expectation for the normalized SVC loss can be decomposed into expectations over the (unnormalized) SVC loss and over log Z(y) (see end of section 1). While the former can be computed analytically, the latter expectation can be handled by replacing log Z (y) by a piecewise defined tight bound such that the integral can be solved analytically. For the GPC loss (6) cannot be solved analytically and was in our experiments approximated by Gaussian quadrature. We can optimize F using a nested loop algorithm as follows. In the inner loop we run an optimizer to minimize F w.r.t. P for fixed (J. We used a conjugate gradients optimizer since the number of parameters of P is rather large. The outer loop is an optimizer minimizing F w.r.t. (J, and we chose a Quasi-Newton method here since the dimension of e is usually rather small and gradients w.r.t. (J are costly to evaluate. We can use the resulting minimizer (P,O) of F in two different ways. The most natural is to discard P, plug 0 into the original architecture and predict using the mode of P(y ID, 0) as an approximation to the true posterior mode, benefitting from a kernel now adapted to the given data. This is particularly interesting for Support Vector machines due to the sparseness of the final kernel expansion (typically only a small fraction of the components in the weight vector K-1iJ is non-zero, the corresponding datapoints are termed Support Vectors) which allows very efficient predictions for a large number of test points. However, we can also retain P and use it as a Gaussian approximation of the posterior P(yID, 0). Doing so, we can use the variance of the approximative predictive distribution P(y.lx., D) to derive error bars for our predictions, although the interpretation of these figures is somewhat complicated in the case of kernel discriminants like SVM whose loss function does not correspond to a noise distribution. 2.1 Relations to other methods Let us have a look at alternative ways to maximize (4). If the loss get, y) is twice differentiable everywhere, progress can be made by replacing g by its second order Taylor expansion around the mode of the integrand. This is known as Laplace approximation and is used in [16] to maximize (4) approximately. However, this technique cannot be used for nondifferentiable losses of the c-insensitive type 6 ? Nevertheless, for the SVC loss the evidence (4) can be approximated in a Laplacelike fashion [11], and it will be interesting to compare the results of this work with ours. This approximation can be evaluated very efficiently, but is not continuous 7 6The nondifferentiabilities cannot be ignored since with probability one a nonzero number of the ih sit exactly at these margin locations. 7 Although continuity can be accomplished by a further modification, see [11]. 607 Bayesian Model Selection for Support Vector Machines w.r .t. (J and difficult to optimize if the dimension of e is not small. Opper and Winther [8] use mean field ideas to derive an approximate leave-one-out test error estimator which can be quickly evaluated, but suffers from the typical noisiness of cross-validation scores. Kwok [6] applies the evidence framework to Support Vector machines, but the technique seems to be restricted to kernels with a finite eigenfunction expansion (see [13] for details) . It is interesting to compare our variational method to the Laplace method of [16] and the variational technique of [2]. Let g(t, y) be differentiable and suppose that for given (J we restrict ourselves to approximate (6) by replacing g(ti' Yi) by the expansion 2g g A)(Yi - JLi )2 , g(ti' JLi) + 8 By (ti, JLi) (Yi - JLi ) + 2.18 8y2 (ti, Yi (7) where fj is the posterior mean . This will change the criterion F to Fapproz, say. Then it is easy to show that the Gaussian approximation to the posterior employed by the Laplace method, namely N(fj, (K- 1 + W)-l), W = diag(u(Yi)(1-u(Yi?) , minimizes Fapproz w.r.t. P if full covariances ~ are used, and plugging this minimizer into Fapproz we end up with the evidence approximation which is maximized by the Laplace method. The latter is not a variational technique since the approximation (7) to the loss function is not an upper bound, and works only for differentiable loss functions . If we upper bound the loss function g(t,y) by a quadratic polynomial and add the variational parameters of this bound to the parameters of P, our method becomes broadly similar to the lower bound algorithm of [2]. Indeed, since for fixed variational parameters of the polynomials we can easily solve for the mean and covariance of P, the former parameters are the only essential ones. However, the quadratic upper bound is poor for functions like the SVC loss , and in these cases our bound is expected to be tighter. 3 Experiments We tested our variational algorithm on a number of datasets from the UCI machine learning repository and the DELVE archive of the University of Toront08 : Leptograpsus crabs, Pima Indian diabetes, Wisconsin Breast Cancer, Ringnorm, Twonorm and Waveform (class 1 against 2). Descriptions may be found on the web. In each case we normalized the whole set to zero mean , unit variance in all input columns, picked a training set at random and used the rest for testing. We chose (for X = JRd) the well-known squared-exponential kernel (see [15]): K(x,xI9) = C (exp ( - 2~ t Wi (Xi - Xi)') + v), 9 = ?Wi):'C,V)'. (8) All parameters are constrained to be positive, so we chose the representation ()i = v'f. We did not use a prior on (J (see comment at end of this section). For comparison we trained a Gaussian Process classifier with the Laplace method (also without hyperprior) and a Support Vector machine using lO-fold cross-validation to select the free parameters. In the latter case we constrained the scale parameters Wi to be equal (it is infeasible to adapt d + 2 hyperparameters to the data using crossvalidation) and dropped the v parameter while allowing for a bias parameter. As mentioned above, within the variational method we can use the posterior mode fj 8See http://vvv . cs. utoronto. cal ...... del ve and http://vvv.ics.uci.edu/ ......mlearn/MLRepository.html . M Seeger 608 Name crabs pima wdbc twonorm ringnorm waveform train size 80 200 300 300 400 800 test size 120 332 269 7100 7000 2504 Var.GP IL 4 3 66 66 11 11 233 224 119 124 206 204 I y GP Lapl. 4 68 8 297 184 221 Var. SVM y IL 4 4 64 66 10 10 260 223 129 126 211 206 SVM 10-CV 4 67 9 163 160 197 Lin. discr. 3 67 19 207 1763 220 Table 1: Number of test errors for various methods. as well as the mean IL of P for prediction, and we tested both methods. Error bars were not computed. The baseline method was a linear discriminant trained to minimize the squared error. Table 1 shows the test errors the different methods attained. These results show that the new algorithm performs equally well as the other methods we considered. They have of course to be regarded in combination with how much effort was necessary to produce them. It took us almost a whole day and a lot of user interactions to do the cross-validation model selection. The rule-of-thumb that a lot of Support Vectors at the upper bound indicate too large a parameter C in (8) failed for at least two of these sets, so we had to start with very coarse grids and sweep through several stages of refinement. An effect known as automatic relevance determination (ARD) (see [7]) can be nicely observed on some of the datasets, by monitoring the length scale parameters Wi in (8). Indeed, our variational SVC algorithm almost completely ignored (by driving their length scales to very small values) 3 of the 5 dimensions in "crabs", 2 of 7 in "pima" and 3 of 21 in "waveform". On "wdbc", it detected dimension 24 as particularly important with regard to separation, all this in harmony with the GP Laplace method. Thus, a sensible parameterized kernel family together with a method of the Bayesian kind allows us to gain additional important information from a dataset which might be used to improve the experimental design. Results of experiments with the methods tested above and hyperpriors as well as a more detailed analysis of the experiments can be found in [9]. 4 Discussion We have shown how to perform model selection for Support Vector machines using approximative Bayesian variational techniques. Our method is applicable to a wide range of loss functions and is able to adapt a large number of hyperparameters to given data. This allows for the use of sophisticated kernels and Bayesian techniques like automatic relevance determination (see [7]) which is not possible using other common model selection criteria like cross-validation. Since our method is fully automatic, it is easy for non-experts to use 9 , and as the evidence is computed on the training set, no training data has to be sacrificed for validation. We refer to [9] where the topics of this paper are investigated in much greater detail. A pressing issue is the unfortunate scaling of the method with the training set 9 As an aside, this opens the possibility of comparing SVMs against other fullyautomatic methods within the DELVE project (see section 3) . Bayesian Model Selection for Support Vector Machines 609 size n which is currently O(n 3)1O. We are currently explori~g the applicability of the powerful approximations of [10] which might bring us very much closer to the desired O(n2) scaling (see also [2]). Another interesting issue would be to connect our method with the work of [5] who use generative models to derive kernels in situations where the "standard kernels" are not applicable or not reasonable. Acknowledgments We thank Chris Williams, Amos Storkey, Peter Sollich and Carl Rasmussen for helpful and inspiring discussions. This work was partially funded by a scholarship of the Dr. Erich Muller foundation. We are grateful to the Division of Informatics for supporting our visit in Edinburgh, and to Chris Williams for making it possible. References [1] David Barber and Christopher Bishop. Ensemble learning for multi-layer networks. In Advances in NIPS, number 10, pages 395-401. MIT Press, 1997. [2] Mark N. Gibbs. Bayesian Gaussian Processes for Regression and Classification. PhD thesis, University of Cambridge, 1997. [3] Geoffrey E. Hinton and D. Van Camp. Keeping neural networks simple by minimizing the description length of the weights. In Proceedings of the 6th annual conference on computational learning theory, pages 5- 13, 1993. [4] Tommi Jaakkola, Marina Meila, and Tony Jebara. Maximum entropy discrimination. In Advances in NIPS, number 13. MIT Press, 1999. [5] Tommi S. Jaakkola and David Haussler. Exploiting generative models in discriminative classifiers. In Advances in NIPS, number 11, 1998. [6] James Tin-Tau Kwok. Integrating the evidence framework and the Support Vector machine. Submitted to ESANN 99, 1999. [7] Radford M. Neal. Monte Carlo implementation of Gaussian process models for Bayesian classification and regression. Technical Report 9702, Department of Statistics, University of Toronto, January 1997. [8] Manfred Opper and Ole Winther. GP classification and SVM: Mean field results and leave-one-out estimator. In Advances in Large Margin Classifiers. MIT Press, 1999. [9] Matthias Seeger. Bayesian methods for Support Vector machines and Gaussian processes. Master's thesis, University of Karlsruhe, Germany, 1999. Available at http://vvw.dai.ed.ac.uk/-seeger. [10] John Skilling. Maximum entropy and Bayesian methods. Cambridge University Press, 1988. [11] Peter Sollich. Probabilistic methods for Support Vector machines. In Advances in NIPS, number 13. MIT Press, 1999. [12] Vladimir N. Vapnik. Statistical Learning Theory. Wiley, 1998. [13] Grace Wahba. Spline Models for Observational Data. CBMS-NSF Regional Conference Series. SIAM, 1990. [14] Grace Wahba. Support Vector machines, reproducing kernel Hilbert spaces and the randomized GACV. Technical Report 984, University of Wisconsin, 1997. [15] Christopher K. 1. Williams. Prediction with Gaussian processes: From linear regression to linear prediction and beyond. In M. 1. Jordan, editor, Learning in Graphical Models. Kluwer, 1997. [16] Christopher K.I. Williams and David Barber. Bayesian classification with Gaussian processes. IEEE 7rans. PAMI, 20(12):1342-1351, 1998. laThe running time is essentially the same as that of the Laplace method, thus being comparable to the fastest known Bayesian GP algorithm.
1722 |@word repository:1 briefly:1 polynomial:2 norm:1 seems:1 logit:1 open:2 covariance:7 series:1 score:1 rkhs:3 ours:1 existing:1 ka:1 comparing:1 written:2 john:1 aside:1 discrimination:1 generative:4 lr:2 manfred:1 transposition:1 coarse:1 location:1 lx:6 attack:2 toronto:1 introduce:1 sacrifice:1 expected:1 indeed:3 multi:1 decomposed:1 actual:1 inappropriate:1 becomes:1 project:1 maximizes:1 what:1 kind:1 minimizes:2 ti:8 exactly:1 classifier:5 uk:2 yi9:1 unit:1 ly:2 positive:3 dropped:2 limit:1 id:4 approximately:1 pami:1 might:3 chose:3 twice:1 specifying:1 ringnorm:2 delve:2 fastest:1 range:2 acknowledgment:1 testing:1 practice:1 definite:1 danger:1 yio:3 integrating:1 get:4 cannot:5 close:2 selection:8 cal:1 applying:1 optimize:2 map:2 straightforward:2 williams:4 simplicity:1 dly:2 estimator:2 rule:1 haussler:1 regarded:1 datapoints:1 jli:4 laplace:10 target:1 suppose:1 user:2 programming:1 gps:1 approximative:3 us:1 hypothesis:1 carl:1 diabetes:1 trend:1 element:1 approximated:4 particularly:2 storkey:1 xi9:1 ep:2 observed:1 solved:3 calculate:1 improper:2 decrease:2 mentioned:1 ui:1 xio:2 trained:2 grateful:1 tight:1 predictive:4 division:1 completely:1 gacv:1 easily:2 various:2 sacrificed:1 train:1 monte:3 ole:1 detected:1 choosing:2 whose:1 widely:1 solve:2 say:1 enlarges:1 statistic:2 gp:9 jointly:1 noisy:1 final:2 differentiable:3 pressing:1 matthias:2 took:1 interaction:2 uci:2 loop:3 iff:1 description:2 normalize:1 crossvalidation:1 exploiting:1 argmaxp:2 produce:1 bve:1 leave:2 depending:2 derive:3 ac:2 ard:1 ij:2 progress:1 esann:1 c:1 indicate:1 tommi:2 guided:1 waveform:3 attribute:1 stochastic:1 observational:1 dio:1 tighter:1 extension:1 clarify:2 around:1 considered:2 crab:3 ic:1 exp:5 predict:2 pointing:1 driving:1 optimizer:3 applicable:2 harmony:1 currently:2 successfully:2 amos:1 mit:4 clearly:1 gaussian:25 aim:1 rather:2 jaakkola:2 focus:1 noisiness:1 likelihood:1 seeger:7 learning4:1 baseline:1 benefitting:1 camp:2 posteriori:1 helpful:1 i0:1 typically:1 relation:3 transformed:1 quasi:1 germany:1 issue:2 classification:11 dual:2 html:1 plan:1 smoothing:2 special:2 constrained:2 field:3 equal:2 having:2 nicely:1 placing:2 look:1 report:2 spline:3 piecewise:1 divergence:2 ve:1 ourselves:1 tlx:2 interest:1 possibility:2 analyzed:1 chain:2 integral:2 closer:1 necessary:1 old:1 hyperprior:2 taylor:1 desired:1 column:1 logp:4 applicability:1 too:1 connect:1 ehi:1 density:1 winther:2 twonorm:2 siam:1 retain:1 randomized:1 probabilistic:1 informatics:1 together:1 quickly:1 squared:2 thesis:2 choose:2 dr:1 expert:1 tii:1 root:1 picked:1 lot:2 doing:1 start:1 di9:1 complicated:1 minimize:4 mlps:1 il:4 variance:3 who:1 efficiently:3 ensemble:3 correspond:1 maximized:1 bayesian:18 thumb:1 carlo:3 monitoring:1 submitted:1 mlearn:1 suffers:1 ed:2 leptograpsus:1 ty:4 energy:1 against:2 james:1 gain:1 dataset:1 knowledge:1 hilbert:2 agreed:1 sophisticated:2 cbms:1 attained:1 day:1 follow:1 methodology:1 done:1 evaluated:3 generality:1 stage:1 correlation:1 web:1 replacing:3 christopher:3 del:1 continuity:1 logistic:1 mode:5 karlsruhe:1 name:1 effect:2 normalized:3 true:1 y2:1 former:2 analytically:5 leibler:1 nonzero:1 neal:1 deal:1 unnormalized:2 mlrepository:1 criterion:3 hill:1 ay:1 tt:1 oid:1 performs:1 bring:1 fj:3 variational:14 svc:14 common:3 discriminants:2 functional:1 insensitive:2 jp:1 interpretation:1 kluwer:1 refer:1 cambridge:2 gibbs:1 llx:2 cv:1 smoothness:1 automatic:3 meila:1 consistency:1 grid:1 erich:1 had:1 funded:1 add:1 posterior:14 prime:1 discard:1 termed:1 certain:1 yi:9 accomplished:1 muller:1 seen:1 dai:2 somewhat:1 additional:1 greater:1 employed:1 maximize:2 ii:1 full:2 jrd:1 technical:2 adapt:3 plug:1 cross:4 determination:2 lin:1 equally:1 visit:1 marina:1 plugging:1 prediction:6 regression:4 basic:1 breast:1 essentially:1 expectation:4 kernel:21 normalization:1 represent:1 justified:1 semiparametric:1 rest:1 regional:1 archive:1 comment:1 vvv:2 jordan:1 call:1 easy:3 xj:1 architecture:1 restrict:1 wahba:2 reduce:1 idea:2 inner:1 handled:1 effort:1 render:1 peter:2 discr:1 ignored:2 gpc:3 detailed:1 amount:1 inspiring:1 svms:1 differentiability:1 http:3 problematic:1 nsf:1 track:1 broadly:1 nevertheless:1 clarity:1 changing:1 fraction:1 sum:1 tly:4 run:1 everywhere:2 parameterized:1 powerful:1 master:1 family:3 reader:1 almost:3 reasonable:1 forrest:1 separation:1 dy:5 scaling:2 comparable:1 bound:10 layer:1 fold:1 quadratic:3 nonnegative:1 annual:1 adapted:1 flat:1 integrand:1 department:1 combination:2 poor:1 conjugate:1 sollich:2 lld:1 wi:4 appealing:1 modification:1 making:1 restricted:1 rans:1 tractable:2 end:3 available:2 gaussians:1 apply:2 kwok:2 hyperpriors:1 generic:1 skilling:1 alternative:1 original:1 denotes:4 binomial:1 tony:1 running:1 unfortunate:1 graphical:1 newton:1 scholarship:1 sweep:1 costly:1 diagonal:2 grace:2 gradient:3 thank:1 outer:1 nondifferentiable:1 sensible:1 topic:1 chris:2 barber:2 discriminant:3 length:3 pointwise:1 relationship:1 minimizing:2 vladimir:1 lapl:1 ql:1 difficult:3 pima:3 relate:1 negative:1 design:2 implementation:1 proper:1 perform:1 allowing:1 upper:6 markov:2 datasets:2 finite:2 supporting:1 january:1 situation:2 hinton:2 reproducing:2 jebara:1 community:1 introduced:2 david:3 namely:2 nip:4 eigenfunction:1 able:3 bar:2 beyond:1 usually:2 challenge:1 memory:1 tau:1 ia:1 misclassification:2 natural:1 indicator:1 yid:9 improve:1 review:1 prior:6 wisconsin:2 loss:21 fully:1 interesting:4 var:2 geoffrey:1 validation:6 foundation:1 editor:1 lo:1 cancer:1 course:1 free:2 rasmussen:1 infeasible:1 keeping:1 bias:3 institute:1 wide:1 edinburgh:3 regard:2 van:2 dimension:4 xn:1 opper:2 made:1 adaptive:1 refinement:1 approximate:4 absorb:1 kullback:1 overfitting:1 conceptual:1 consuming:1 xi:8 discriminative:2 search:1 latent:2 continuous:1 table:2 expansion:4 excellent:1 investigated:1 diag:1 did:1 whole:2 noise:5 hyperparameters:4 n2:1 quadrature:2 fashion:1 wiley:1 wish:1 exponential:1 xl:1 lw:1 tin:1 loss2:1 bishop:1 utoronto:1 favourable:1 svm:5 evidence:7 sit:1 essential:1 ih:1 vapnik:1 phd:1 sparseness:1 margin:2 easier:1 wdbc:2 entropy:2 simply:2 failed:1 partially:1 applies:1 radford:1 nested:1 minimizer:3 change:1 determined:1 except:1 typical:1 averaging:1 miss:1 called:1 experimental:2 select:1 support:20 mark:1 latter:5 relevance:2 indian:1 evaluate:1 tested:3
790
1,723
Support Vector Method for Novelty Detection Bernhard Scholkopf*, Robert Williamson?, Alex Smola?, John Shawe-Taylor t , John Platt* ? * Microsoft Research Ltd., 1 Guildhall Street, Cambridge, UK Department of Engineering, Australian National University, Canberra 0200 t Royal Holloway, University of London, Egham, UK * Microsoft, 1 Microsoft Way, Redmond, WA, USA bsc/jplatt@microsoft.com, Bob.WilliamsoniAlex.Smola@anu.edu.au, john@dcs.rhbnc.ac.uk Abstract Suppose you are given some dataset drawn from an underlying probability distribution P and you want to estimate a "simple" subset S of input space such that the probability that a test point drawn from P lies outside of S equals some a priori specified l/ between 0 and 1. We propose a method to approach this problem by trying to estimate a function f which is positive on S and negative on the complement. The functional form of f is given by a kernel expansion in terms of a potentially small subset of the training data; it is regularized by controlling the length of the weight vector in an associated feature space. We provide a theoretical analysis of the statistical performance of our algorithm. The algorithm is a natural extension of the support vector algorithm to the case of unlabelled data. 1 INTRODUCTION During recent years, a new set of kernel techniques for supervised learning has been developed [8]. Specifically, support vector (SV) algorithms for pattern recognition, regression estimation and solution of inverse problems have received considerable attention. There have been a few attempts to transfer the idea of using kernels to compute inner products in feature spaces to the domain of unsupervised learning. The problems in that domain are, however, less precisely specified. Generally, they can be characterized as estimating junctions of the data which tell you something interesting about the underlying distributions. For instance, kernel PCA can be characterized as computing functions which on the training data produce unit variance outputs while having minimum norm in feature space [4] . Another kernel-based unsupervised learning technique, regularized principal manifolds [6], computes functions which give a mapping onto a lower-dimensional manifold minimizing a regularized quantization error. Clustering algorithms are further examples of unsupervised learning techniques which can be kernelized [4] . An extreme point of view is that unsupervised learning is about estimating densities. Clearly, knowledge of the density of P would then allow us to solve whatever problem can be solved on the basis of the data. The present work addresses an easier problem: it Support Vector Method for Novelty Detection 583 proposes an algorithm which computes a binary function which is supposed to capture regions in input space where the probability density lives (its support), i.e. a function such that most of the data will live in the region where the function is nonzero [5]. In doing so, it is in line with Vapnik's principle never to solve a problem which is more general than the one we actually need to solve. Moreover, it is applicable also in cases where the density of the data's distribution is not even well-defined, e.g. if there are singular components. Part of the motivation for the present work was the paper [1]. It turns out that there is a considerable amount of prior work in the statistical literature; for a discussion, cf. the full version of the present paper [3]. 2 ALGORITHMS We first introduce terminology and notation conventions. We consider training data Xl, ... , Xl E X, where fEN is the number of observations, and X is some set. For simplicity, we think of it as a compact subset of liN. Let ~ be a feature map X -t F, i.e. a map into a dot product space F such that the dot product in the image of ~ can be computed by evaluating some simple kernel [8] k(x, y) = (~(x) . ~(y)), (1) such as the Gaussian kernel (2) Indices i and j are understood to range over 1, ... ,f (in compact notation: 't, J E [fD. Bold face greek letters denote f-dimensional vectors whose components are labelled using normal face typeset. In the remainder of this section, we shall develop an algorithm which returns a function f that takes the value + 1 in a "small" region capturing most of the data points, and -1 elsewhere. Our strategy is to map the data into the feature space corresponding to the kernel, and to separate them from the origin with maximum margin. For a new point X, the value f(x) is determined by evaluating which side of the hyperplane it falls on, in feature space. Via the freedom to utilize different types of kernel functions, this simple geometric picture corresponds to a variety of nonlinear estimators in input space. To separate the data set from the origin, we solve the following quadratic program: ~llwl12 min wEF,eEiRt,PEiR subject to (w? ~(Xi)) + ;l L i ei 2:: P - ei, (3) P ei 2:: o. (4) Here, 1/ E (0, 1) is a parameter whose meaning will become clear later. Since nonzero slack variables ei are penalized in the objective function, we can expect that if wand p solve this problem, then the decision function f(x) = sgn((w . ~(x)) - p) will be positive for most examples Xi contained in the training set, while the SV type regularization term Ilwll will still be small. The actual trade-off between these two goals is controlled by 1/. Deriving the dual problem, and using (1), the solution can be shown to have an SV expansion f(x) = 'gn ( ~ a;k(x;, x) - p) (5) (patterns Xi with nonzero ll!i are called SVs), where the coefficients are found as the solution of the dual problem: main ~ L ll!ill!jk(Xi, Xj) ij subject to 0 ~ ll!i ~ :f' L ll!i = 1. (6) B. ScMlkop/, R. C. Williamson, A. J Smola, J Shawe-Taylor and J C. Platt 584 This problem can be solved with standard QP routines. It does, however, possess features that sets it apart from generic QPs, most notably the simplicity of the constraints. This can be exploited by applying a variant of SMO developed for this purpose [3]. The offset p can be recovered by exploiting that for any ll:i which is not at the upper or lower bound, the corresponding pattern Xi satisfies p = (w . <.P(Xi)) L:j ll:jk(Xj, Xi) . = Note that if v approaches 0, the upper boundaries on the Lagrange multipliers tend to infinity, i.e. the second inequality constraint in (6) becomes void. The problem then resembles the corresponding hard margin algorithm, since the penalization of errors becomes infinite, as can be seen from the primal objective function (3). It can be shown that if the data set is separable from the origin, then this algorithm will find the unique supporting hyperplane with the properties that it separates all data from the origin, and its distance to the origin is maximal among all such hyperplanes [3]. If, on the other hand, v approaches I, then the constraints alone only allow one solution, that where all ll:i are at the upper bound 1/ (v?). In this case, for kernels with integral I, such as normalized versions of (2), the decision function corresponds to a thresholded Parzen windows estimator. To conclude this section, we note that one can also use balls to describe the data in feature space, close in spirit to the algorithms of [2], with hard boundaries, and [7], with "soft margins." For certain classes of kernels, such as Gaussian RBF ones, the corresponding algorithm can be shown to be equivalent to the above one [3]. 3 THEORY In this section, we show that the parameter v characterizes the fractions of SVs and outliers (Proposition 1). Following that, we state a robustness result for the soft margin (Proposition 2) and error bounds (Theorem 5). Further results and proofs are reported in the full version of the present paper [3]. We will use italic letters to denote the feature space images of the corresponding patterns in input space, i.e. X i := <.P(Xi). Proposition 1 Assume the solution of (4) satisfies p =1= 0. The following statements hold: (i) v is an upper bound on the fraction of outliers. (ii) v is a lower bound on the fraction of SVs. (iii) Suppose the data were generated independently from a distribution P(x) which does not contain discrete components. Suppose, moreover, that the kernel is analytic and nonconstant. With probability 1, asymptotically, v equals both the fraction of SVs and the fraction of outliers. The proof is based on the constraints of the dual problem, using the fact that outliers must have Lagrange multipliers at the upper bound. Proposition 2 Local movements of outliers parallel to w do not change the hyperplane. We now move on to the subject of generalization. Our goal is to bound the probability that a novel point drawn from the same underlying distribution lies outside of the estimated region by a certain margin. We start by introducing a common tool for measuring the capacity of a class :r of functions that map X to lit Definition 3 Let (X, d) be a pseudo-metric space, I let A be a subset of X and f. > 0. A set B ~ X is an f.-cover for A if, for every a E A, there exists b E B such that d( a , b) ::::; f.. The f.-covering number of A, Nd(f., A), is the minimal cardinality of an f.-cover for A (if there is no such finite cover then it is defined to be 00). I i.e. with a distance function that differs from a metric in that it is only semidefinite 585 Support Vector Method for Novelty Detection The idea is that B should be finite but approximate all of A with respect to the pseudometric d. We will use the loo distance over a finite sample X = (Xl, .. ? , Xl) for the pseudometric in the space of functions, dx(f, g) = m~E[lllf(xd - g(xi)l. Let N(E,~, f) = SUPXEXI Ndx (E, ~). Below, logarithms are to base 2. Theorem 4 Consider any distribution P on X and any 0 E lR. Suppose Xl, ?.. , Xl are generated U.d. from P. Then with probability 1 - 6 over such an f-sample, if we find f E ~ such that f(Xi) ~ 0 for all i E [f), +, P{x : f(x) < 0 -,} :s; t(k + log 2(l), where k = rlog:Nb,~, 2f)1- We now consider the possibility that for a small number of points f(Xi) fails to exceed 0+,. This corresponds to having a non-zero slack variable ~i in the algorithm, where we take 0 + , = p / II w II and use the class of linear functions in feature space in the application of the theorem. There are well-known bounds for the log covering numbers of this class. Let f be a real valued function on a space X. Fix 0 E lR. For X E X, define d(x,J, ,) = max{O,O +, - f(x)}. Similarly for a training sequence X, we define 'D(X, f, ,) = L:xEX d(x, f, ,). Theorem 5 Fix 0 E lR. Consider a fixed but unknown probability distribution P on the input space X and a class of real valued functions ~ with range [a, b). Then with probability 1 - 6 over randomly drawn training sequences X of size f, for all, > 0 and any f E ~, P {x: f(x) < 0 - , and X ~ X} :s; t(k + log ~l ), where k = rlogN('V/2 ~ , U) I , + 64(b-a)'D(X,J,'y) -y2 log ( 8'D(X,J,-y) ell ) log (32l(b-a)2)1. -y2 The theorem bounds the probability of a new point falling in the region for which f(x) has value less than 0 - " this being the complement of the estimate for the support of the distribution. The choice of, gives a trade-off between the size of the region over which the bound holds (increasing, increases the size of the region) and the size of the probability with which it holds (increasing, decreases the size of the log covering numbers). The result shows that we can bound the probability of points falling outside the region of estimated support by a quantity involving the ratio of the log covering numbers (which can be bounded by the fat shattering dimension at scale proportional to ,) and the number of training examples, plus a factor involving the I-norm of the slack variables. It is stronger than related results given by [I], since their bound involves the square root of the ratio of the Pollard dimension (the fat shattering dimension when, tends to 0) and the number of training examples. The output of the algorithm described in Sec. 2 is a function f(x) = 2:::i aik(xi' x) which is greater than or equal to p - ~i on example Xi. Though non-linear in the input space, this function is in fact linear in the feature space defined by the kernel k . At the same time the 2-norm of the weight vector is given by B = J aT K a, and so we can apply the theorem with the function class ~ being those linear functions in the feature space with 2-norm bounded by B . If we assume that 0 is fixed, then, = p - 0, hence the support of the distribution is the set {x : f (x) ~ 0 - , = 20 - p}, and the bound gives the probability of a randomly generated point falling outside this set, in terms of the log covering numbers of the function class ~ and the sum of the slack variables ~i. Since the log covering numbers 586 B. SchOlkopj R. C. Williamson, A. 1. Smola, 1. Shawe-Taylor and 1. C. Platt 2 f) this gives a bound in terms at scale, /2 of the class ~ can be bounded by O( B:!F-Iog "Y of the 2-norm of the weight vector. Ideally, one would like to allow () to be chosen after the value of p has been determined, perhaps as a fixed fraction of that value. This could be obtained by another level of structural risk minimisation over the possible values of p or at least a mesh of some possible values. This result is beyond the scope of the current preliminary paper, but the form of the result would be similar to Theorem 5, with larger constants and log factors. Whilst it is premature to give specific theoretical recommendations for practical use yet, one thing is clear from the above bound. To generalize to novel data, the decision function to be used should employ a threshold TJ ? p, where TJ < 1 (this corresponds to a nonzero I)' 4 EXPERIMENTS We apply the method to artificial and real-world data. Figure 1 displays 2-D toy examples, and shows how the parameter settings influence the solution. Next, we describe an experiment on the USPS dataset of handwritten digits. The database contains 9298 digit images of size 16 x 16 = 256; the last 2007 constitute the test set. We trained the algorithm, using a Gaussian kernel (2) of width c = 0.5 . 256 (a common value for SVM classifiers on that data set, cf. [2]), on the test set and used it to identify outliers - it is folklore in the community that the USPS test set contains a number of patterns which are hard or impossible to classify, due to segmentation errors or mislabelling. In the experiment, we augmented the input patterns by ten extra dimensions corresponding to the class labels of the digits. The rationale for this is that if we disregarded the labels, there would be no hope to identify mislabelled patterns as outliers. Fig. 2 shows the 20 worst outliers for the USPS test set. Note that the algorithm indeed extracts patterns which are very hard to assign to their respective classes. In the experiment, which took 36 seconds on a Pentium II running at 450 MHz, we used a 11 value of 5%. Figure I: First two pictures: A single-class SVM applied to two toy problems; 11 = C = 0.5, domain: [-1, 1F. Note how in both cases, at least a fraction of 11 of all examples is in the estimated region (cf. table). The large value of 11 causes the additional data points in the upper left comer to have almost no influence on the decision function. For smaller values of 11, such as 0.1 (third picture), the points cannot be ignored anymore. Alternatively, one can force the algorithm to take these 'outliers' into account by changing the kernel width (2): in the fourth picture, using c = 0.1,11 = 0.5, the data is effectively analyzed on a different length scale which leads the algorithm to consider the outliers as meaningful points. Support Vector Method/or Novelty Detection 587 ~"1r~ftC ~"'J.. ~()nl~) 9 -507 1 -4580 -377 1 -282 7 -2162 -2003 -1869 -179 5 - -153 3 -143 6 -1286 - 3 0 -1177 -93 5 -78 0 -58 7 -52 6 -48 3 Figure 2: Outliers identified by the proposed algorithm, ranked by the negative output of the SVM (the argument of the sgn in the decision function). The outputs (for convenience in units of 10- 5 ) are written underneath each image in italics, the (alleged) class labels are given in bold face. Note that most of the examples are "difficult" in that they are either atypical or even mislabelled. 5 DISCUSSION One could view the present work as an attempt to provide an algorithm which is in line with Vapnik's principle never to solve a problem which is more general than the one that one is actually interested in. E.g., in situations where one is only interested in detecting novelty, it is not always necessary to estimate a full density model of the data. Indeed, density estimation is more difficult than what we are doing, in several respects. Mathematically speaking, a density will only exist if the underlying probability measure possesses an absolutely continuous distribution function. The general problem of estimating the measure for a large class of sets, say the sets measureable in Borel's sense, is not solvable (for a discussion, see e.g. [8]). Therefore we need to restrict ourselves to making a statement about the measure of some sets. Given a small class of sets, the simplest estimator accomplishing this task is the empirical measure, which simply looks at how many training points fall into the region of interest. Our algorithm does the opposite. It starts with the number of training points that are supposed to fall into the region, and then estimates a region with the desired property. Often, there will be many such regions - the solution becomes unique only by applying a regularizer, which in our case enforces that the region be small in a feature space associated to the kernel. This, of course, implies, that the measure of smallness in this sense depends on the kernel used, in a way that is no different to any other method that regularizes in a feature space. A similar problem, however, appears in density estimation already when done in input space. Let p denote a density on X. If we perform a (nonlinear) coordinate transformation in the input domain X, then the density values will change; loosely speaking, what remains constant is p(x) . dx, while dx is transformed, too. When directly estimating the probability measure of regions, we are not faced with this problem, as the regions automatically change accordingly. An attractive property of the measure of smallness that we chose to use is that it can also be placed in the context of regularization theory, leading to an interpretation of the solution as maximally smooth in a sense which depends on the specific kernel used [3]. The main inspiration for our approach stems from the earliest work of Vapnik and collaborators. They proposed an algorithm for characterizing a set of unlabelled data points by separating it from the origin using a hyperplane [9]. However, they quickly moved on to two-class classification problems, both in terms of algorithms and in the theoretical development of statistical learning theory which originated in those days. From an algorithmic point of view, we can identify two shortcomings of the original approach which may have caused research in this direction to stop for more than three decades. Firstly, the original 588 B. Scholkopf, R. C. Williamson, A. J Smola, J Shawe-Taylor and J C. Platt algorithm in was limited to linear decision rules in input space, secondly, there was no way of dealing with outliers. In conjunction, these restrictions are indeed severe - a generic dataset need not be separable from the origin by a hyperplane in input space. The two modifications that we have incorporated dispose of these shortcomings. Firstly, the kernel trick allows for a much larger class of functions by nonlinearly mapping into a high-dimensional feature space, and thereby increases the chances of separability from the origin. In particular, using a Gaussian kernel (2), such a separation exists for any data set Xl, ... , Xl: to see this, note that k(Xi, Xj) > 0 for all i, j, thus all dot products are positive, implying that all mapped patterns lie inside the same orthant. Moreover, since k(Xi, Xi) = 1 for all i, they have unit length. Hence they are separable from the origin. The second modification allows for the possibility of outliers. We have incorporated this 'softness' of the decision rule using the v-trick and thus obtained a direct handle on the fraction of outliers. We believe that our approach, proposing a concrete algorithm with well-behaved computational complexity (convex quadratic programming) for a problem that so far has mainly been studied from a theoretical point of view has abundant practical applications. To turn the algorithm into an easy-to-use black-box method for practicioners, questions like the selection of kernel parameters (such as the width of a Gaussian kernel) have to be tackled. It is our expectation that the theoretical results which we have briefly outlined in this paper will provide a foundation for this formidable task. Acknowledgement. Part of this work was supported by the ARC and the DFG (# Ja 37919-1), and done while BS was at the Australian National University and GMD FIRST. AS is supported by a grant of the Deutsche Forschungsgemeinschaft (Sm 62/1-1). Thanks to S. Ben-David, C. Bishop, C. Schnorr, and M. Tipping for helpful discussions. References [1] S. Ben-David and M. Lindenbaum. Learning distributions by their density levels: A paradigm for learning without a teacher. Journal of Computer and System Sciences, 55:171-182,1997. [2] B. SchOlkopf, C. Burges, and V. Vapnik. Extracting support data for a given task. In U. M. Fayyad and R. Uthurusamy, editors, Proceedings, First International Conference on Knowledge Discovery & Data Mining. AAAI Press, Menlo Park, CA, 1995. [3] B. SchOlkopf, J. Platt, J. Shawe-Taylor, AJ. Smola, and R.c. Williamson. Estimating the support of a high-dimensional distribution. TR MSR 99 - 87, Microsoft Research, Redmond, WA, 1999. [4] B. Scholkopf, A. Smola, and K.-R. Muller. Kernel principal component analysis. In B. SchOlkopf, C. Burges, and A. Smola, editors, Advances in Kernel Methods - Support Vector Learning. MIT Press, Cambridge, MA, 1999. 327 - 352. [5] B. SchOlkopf, R. Williamson, A. Smola, and J. Shawe-Taylor. Single-class support vector machines. In J. Buhmann, W. Maass, H. Ritter, and N. Tishby, editors, Unsupervised Learning, Dagstuhl-Seminar-Report 235, pages 19 - 20, 1999. [6] A. Smola, R. C. Williamson, S. Mika, and B. Scholkopf. Regularized principal manifolds. In Computational Learning Theory: 4th European Conference, volume 1572 of Lecture Notes in Artificial Intelligence, pages 214 - 229. Springer, 1999. [7] D.MJ. Tax and R.P.W. Duin. Data domain description by support vectors. In M. Verleysen, editor, Proceedings ESANN, pages 251 - 256, Brussels, 1999. D Facto. [8] V. Vapnik. Statistical Learning Theory. Wiley, New York, 1998. [9] V. Vapnik and A. Lerner. Pattern recognition using generalized portraits. Avtomatika i Telemekhanika, 24:774 -780, 1963.
1723 |@word msr:1 briefly:1 version:3 norm:5 stronger:1 nd:1 thereby:1 tr:1 contains:2 recovered:1 com:1 current:1 yet:1 dx:3 must:1 written:1 john:3 mesh:1 analytic:1 mislabelled:2 xex:1 alone:1 implying:1 intelligence:1 accordingly:1 lr:3 detecting:1 hyperplanes:1 firstly:2 direct:1 become:1 scholkopf:8 inside:1 introduce:1 notably:1 indeed:3 automatically:1 actual:1 window:1 cardinality:1 increasing:2 becomes:3 estimating:5 notation:2 underlying:4 moreover:3 bounded:3 formidable:1 deutsche:1 what:2 developed:2 proposing:1 whilst:1 transformation:1 pseudo:1 every:1 xd:1 softness:1 fat:2 classifier:1 platt:5 uk:3 unit:3 whatever:1 grant:1 facto:1 positive:3 engineering:1 understood:1 local:1 tends:1 black:1 plus:1 chose:1 au:1 resembles:1 studied:1 mika:1 limited:1 range:2 unique:2 practical:2 enforces:1 differs:1 digit:3 empirical:1 lindenbaum:1 onto:1 close:1 cannot:1 convenience:1 selection:1 nb:1 risk:1 live:1 applying:2 influence:2 impossible:1 context:1 equivalent:1 map:4 restriction:1 attention:1 independently:1 convex:1 simplicity:2 estimator:3 rule:2 deriving:1 handle:1 coordinate:1 controlling:1 suppose:4 aik:1 programming:1 origin:9 trick:2 recognition:2 jk:2 database:1 solved:2 capture:1 worst:1 region:16 trade:2 movement:1 decrease:1 dagstuhl:1 complexity:1 ideally:1 trained:1 basis:1 usps:3 comer:1 regularizer:1 describe:2 london:1 shortcoming:2 artificial:2 tell:1 outside:4 whose:2 larger:2 solve:6 valued:2 say:1 typeset:1 think:1 sequence:2 took:1 propose:1 product:4 maximal:1 remainder:1 tax:1 supposed:2 rlog:1 description:1 moved:1 exploiting:1 produce:1 qps:1 ben:2 develop:1 ac:1 ij:1 received:1 esann:1 involves:1 implies:1 australian:2 convention:1 direction:1 greek:1 sgn:2 ja:1 assign:1 fix:2 generalization:1 preliminary:1 collaborator:1 proposition:4 secondly:1 mathematically:1 extension:1 hold:3 normal:1 guildhall:1 mapping:2 scope:1 algorithmic:1 purpose:1 estimation:3 applicable:1 label:3 tool:1 hope:1 bsc:1 clearly:1 mit:1 gaussian:5 always:1 minimisation:1 earliest:1 conjunction:1 mainly:1 pentium:1 underneath:1 sense:3 helpful:1 ftc:1 kernelized:1 transformed:1 interested:2 dual:3 ill:1 among:1 classification:1 priori:1 verleysen:1 proposes:1 development:1 ell:1 equal:3 never:2 having:2 shattering:2 lit:1 look:1 unsupervised:5 park:1 report:1 few:1 employ:1 randomly:2 lerner:1 national:2 dfg:1 ourselves:1 microsoft:5 attempt:2 freedom:1 detection:4 practicioners:1 interest:1 fd:1 possibility:2 mining:1 severe:1 analyzed:1 extreme:1 nl:1 semidefinite:1 primal:1 tj:2 integral:1 necessary:1 respective:1 taylor:6 logarithm:1 loosely:1 desired:1 abundant:1 theoretical:5 minimal:1 instance:1 classify:1 soft:2 portrait:1 gn:1 cover:3 mhz:1 measuring:1 introducing:1 subset:4 jplatt:1 too:1 loo:1 tishby:1 reported:1 teacher:1 sv:3 thanks:1 density:11 international:1 ritter:1 off:2 wef:1 parzen:1 quickly:1 concrete:1 aaai:1 leading:1 return:1 toy:2 account:1 bold:2 sec:1 coefficient:1 caused:1 depends:2 later:1 view:4 root:1 doing:2 characterizes:1 start:2 parallel:1 alleged:1 square:1 accomplishing:1 variance:1 identify:3 generalize:1 handwritten:1 bob:1 definition:1 associated:2 proof:2 stop:1 dataset:3 knowledge:2 segmentation:1 routine:1 telemekhanika:1 actually:2 appears:1 tipping:1 supervised:1 day:1 maximally:1 done:2 though:1 box:1 smola:10 hand:1 ndx:1 ei:4 nonlinear:2 aj:1 perhaps:1 behaved:1 believe:1 usa:1 normalized:1 multiplier:2 contain:1 y2:2 regularization:2 hence:2 inspiration:1 nonzero:4 maass:1 attractive:1 ll:7 during:1 width:3 covering:6 generalized:1 trying:1 image:4 meaning:1 novel:2 common:2 functional:1 qp:1 volume:1 interpretation:1 cambridge:2 outlined:1 similarly:1 shawe:6 dot:3 base:1 something:1 recent:1 apart:1 certain:2 inequality:1 binary:1 life:1 exploited:1 fen:1 muller:1 seen:1 minimum:1 greater:1 additional:1 novelty:5 paradigm:1 ii:4 full:3 stem:1 smooth:1 unlabelled:2 characterized:2 lin:1 iog:1 controlled:1 variant:1 regression:1 involving:2 metric:2 expectation:1 kernel:24 want:1 void:1 singular:1 extra:1 posse:2 subject:3 tend:1 thing:1 spirit:1 extracting:1 structural:1 exceed:1 iii:1 easy:1 forschungsgemeinschaft:1 variety:1 xj:3 identified:1 restrict:1 opposite:1 inner:1 idea:2 rhbnc:1 pca:1 ltd:1 pollard:1 speaking:2 cause:1 svs:4 constitute:1 york:1 ignored:1 generally:1 clear:2 amount:1 ten:1 simplest:1 gmd:1 exist:1 dispose:1 estimated:3 discrete:1 shall:1 terminology:1 threshold:1 falling:3 drawn:4 changing:1 thresholded:1 utilize:1 asymptotically:1 fraction:8 year:1 sum:1 wand:1 inverse:1 letter:2 you:3 fourth:1 almost:1 separation:1 decision:7 capturing:1 bound:15 display:1 tackled:1 quadratic:2 duin:1 precisely:1 constraint:4 alex:1 infinity:1 argument:1 min:1 fayyad:1 pseudometric:2 separable:3 department:1 brussels:1 ball:1 smaller:1 separability:1 making:1 modification:2 b:1 outlier:14 remains:1 turn:2 slack:4 junction:1 apply:2 generic:2 egham:1 uthurusamy:1 anymore:1 robustness:1 original:2 clustering:1 cf:3 running:1 folklore:1 objective:2 move:1 already:1 quantity:1 question:1 avtomatika:1 strategy:1 italic:2 distance:3 separate:3 mapped:1 separating:1 capacity:1 street:1 manifold:3 length:3 index:1 ratio:2 minimizing:1 difficult:2 robert:1 potentially:1 statement:2 measureable:1 negative:2 unknown:1 perform:1 upper:6 observation:1 sm:1 arc:1 finite:3 orthant:1 supporting:1 situation:1 regularizes:1 incorporated:2 dc:1 community:1 david:2 complement:2 nonlinearly:1 specified:2 smo:1 address:1 beyond:1 redmond:2 below:1 pattern:10 program:1 royal:1 max:1 natural:1 force:1 regularized:4 ranked:1 solvable:1 buhmann:1 smallness:2 picture:4 extract:1 nonconstant:1 faced:1 prior:1 literature:1 geometric:1 acknowledgement:1 discovery:1 expect:1 lecture:1 rationale:1 interesting:1 proportional:1 penalization:1 foundation:1 principle:2 editor:4 elsewhere:1 penalized:1 course:1 placed:1 last:1 supported:2 side:1 allow:3 burges:2 fall:3 face:3 characterizing:1 boundary:2 dimension:4 evaluating:2 world:1 computes:2 premature:1 far:1 approximate:1 compact:2 bernhard:1 dealing:1 conclude:1 xi:16 alternatively:1 continuous:1 decade:1 table:1 schnorr:1 transfer:1 mj:1 ca:1 menlo:1 expansion:2 williamson:7 european:1 domain:5 main:2 motivation:1 augmented:1 fig:1 canberra:1 borel:1 wiley:1 fails:1 seminar:1 originated:1 xl:8 lie:3 atypical:1 third:1 theorem:7 specific:2 bishop:1 offset:1 svm:3 exists:2 ilwll:1 quantization:1 vapnik:6 effectively:1 anu:1 margin:5 disregarded:1 easier:1 simply:1 lagrange:2 llwl12:1 contained:1 recommendation:1 springer:1 corresponds:4 satisfies:2 chance:1 ma:1 goal:2 rbf:1 labelled:1 considerable:2 hard:4 change:3 specifically:1 determined:2 infinite:1 hyperplane:5 principal:3 called:1 meaningful:1 holloway:1 support:15 absolutely:1
791
1,724
Spiking Boltzmann Machines Geoffrey E. Hinton Gatsby Computational Neuroscience Unit University College London London WCIN 3AR, UK hinton@gatsby. ucl. ac. uk Andrew D. Brown Department of Computer Science University of Toronto Toronto, Canada andy@cs.utoronto.ca Abstract We first show how to represent sharp posterior probability distributions using real valued coefficients on broadly-tuned basis functions. Then we show how the precise times of spikes can be used to convey the real-valued coefficients on the basis functions quickly and accurately. Finally we describe a simple simulation in which spiking neurons learn to model an image sequence by fitting a dynamic generative model. 1 Population codes and energy landscapes A perceived object is represented in the brain by the activities of many neurons, but there is no general consensus on how the activities of individual neurons combine to represent the multiple properties of an object. We start by focussing on the case of a single object that has multiple instantiation parameters such as position, velocity, size and orientation. We assume that each neuron has an ideal stimulus in the space of instantiation parameters and that its activation rate or probability of activation falls off monotonically in all directions as the actual stimulus departs from this ideal. The semantic problem is to define exactly what instantiation parameters are being represented when the activities of many such neurons are specified. Hinton, Rumelhart and McClelland (1986) consider binary neurons with receptive fields that are convex in instantiation space. They assume that when an object is present it activates all of the neurons in whose receptive fields its instantiation parameters lie. Consequently, if it is known that only one object is present, the parameter values of the object must lie within the feasible region formed by the intersection of the receptive fields of the active neurons. This will be called a conjunctive distributed representation. Assuming that each receptive field occupies only a small fraction of the whole space, an interesting property of this type of "coarse coding" is that the bigger the receptive fields, the more accurate the representation. However, large receptive fields lead to a loss of resolution when several objects are present simultaneously. When the sensory input is noisy, it is impossible to infer the exact parameters of objects so it makes sense for a perceptual system to represent the probability distribution across parameters rather than just a single best estimate or a feasible region. The full probability distribution is essential for correctly combining infor- Spiking Boltzmann Machines E(x) P(X) 123 Figure 1: a) Energy landscape over a onedimensional space. Each neuron adds a dimple (dotted line) to the energy landscape (solid line). b) The corresponding probability density. Where dimples overlap the corresponding probability density becomes sharper. Since the dimples decay to zero, the location of a sharp probability peak is not affected by distant dimples and multimodal distributions can be represented. mation from different times or different Sources. One obvious way to represent this distribution (Anderson and van Essen, 1994) is to allow each neuron to represent a fairly compact probability distribution over the space of instantiation parameters and to treat the activity levels of neurons as (unnormalized) mixing proportions. The semantics of this disjunctive distributed representation is precise, but the percepts it allows are not because it is impossible to represent distributions that are sharper than the individual receptive fields and, in high-dimensional spaces, the individual fields must be broad in order to cover the space. Disjunctive representations are used in Kohonen's self-organizing map which is why it is restricted to very low dimensional latent spaces. The disjunctive model can be viewed as an attempt to approximate arbitrary smooth probability distributions by adding together probability distributions contributed by each active neuron. Coarse coding suggests a multiplicative approach in which the addition is done in the domain of energies (negative log probabilities). Each active neuron contributes an energy landscape over the whole space of instantiation parameters. The activity level of the neuron multiplies its energy landscape and the landscapes for all neurons in the population are added (Figure 1). If, for example, each neuron has a full covariance Gaussian tuning function, its energy landscape is a parabolic bowl whose curvature matrix is the inverse of the covariance matrix. The activity level of the neuron scales the inverse covariance matrix. If there are k instantiation parameters then only k + k(k + 1)/2 real numbers are required to span the space of means and inverse covariance matrices. So the real-valued activities of O(k2) neurons are sufficient to represent arbitrary full covariance Gaussian distributions over the space of instantiation parameters. Treating neural activities as multiplicative coefficients on additive contributions to energy landscapes has a number of advantages. Unlike disjunctive codes, vague distributions are represented by low activities so significant biochemical energy is only required when distributions are quite sharp. A central operation in Bayesian inference is to combine a prior term with a likelihood term or to combine two conditionally independent likelihood terms. This is trivially achieved by adding two energy landscapes l . lWe thank Zoubin Ghahramani for pointing out that another important operation, convolving a probability distribution with Gaussian noise, is a difficult non-linear operation on the energy landscape. 124 2 G. E. Hinton and A. D. Brown Representing the coefficients on the basis functions To perform perception at video rates, the probability distributions over instantiation parameters need to be represented at about 30 frames per second. This seems difficult using relatively slow spiking neurons because it requires the real-valued multiplicative coefficients on the basis functions to be communicated accurately and quickly using all-or-none spikes. The trick is to realise that when a spike arrives at another neuron it produces a postsynaptic potential that is a smooth function of time. So from the perspective of the postsynaptic neuron, the spike has been convolved with a smooth temporal function. By adding a number of these smooth functions together, with appropriate temporal offsets, it is possible to represent any smoothly varying sequence of coefficient values on a basis function, and this makes it possible to represent the temporal evolution of probability distributions as shown in Figure 2. The ability to vary the location of a spike in the single dimension of time thus allows real-valued control of the representation of probability distributions over multiple spatial dimensions. b) a) neuron 2 time .~ >" I ~ 'iii OOiS Q) II: , , Encoded Value Time Figure 2: a)Two spiking neurons centered at 0 and 1 can represent the time-varying mean and standard deviation on a single spatial dimension. The spikes are first convolved with a temporal kernel and the resulting activity values are treated as exponents on Gaussian distributions centered at 0 and 1. The ratio of the activity values determines the mean and the sum of the activity values determines the inverse variance. b) The same method can be used for two (or more) spatial dimensions. Time flows from top to bottom. Each spike makes a contribution to the energy landscape that resembles an hourglass (thin lines). The waist of the hourglass corresponds to the time at which the spike has its strongest effect on some post-synaptic population. By moving the hourglasses in time, it is possible to get whatever temporal cross-sections are desired (thick lines) provided the temporal sampling rate is comparable to the time course of the effect of a spike. Our proposed use of spike timing to convey real values quickly and accurately does not require precise coincidence detection, sub-threshold oscillations, modifiable time delays, or any of the other paraphernalia that has been invoked to explain how the brain could make effective use of the single, real-valued degree of freedom in the timing of a spike (Hopfield, 1995). The coding scheme we have proposed would be far more convincing if we could show how it was learned and could demonstrate that it was effective in a simulation. There are two ways to design a learning algorithm for such spiking neurons. We could work in the relatively low-dimensional space of the instantiation parameters and design the learning to produce the right representations and interactions between representations in this space. Or we could treat this space as an implicit emergent property of the network and design the learning algorithm to optimize Spiking Boltzmann Machines 125 some objective function in the much higher-dimensional space of neural activities in the hope that this will create representations that can be understood using the implicit space of instantiation parameters. We chose the latter approach. 3 A learning algorithm for restricted Boltzmann machines Hinton (1999) describes a learning algorithm for probabilistic generative models that are composed of a number of experts. Each expert specifies a probability distribution over the visible variables and the experts are combined by multiplying these distributions together and renormalizing. (1) where d is a data vector in a discrete space, Om is all the parameters of individual model m, Pm(d\Om) is the probability of d under model m, and i is an index over all possible vectors in the data space. The coding scheme we have described is just a product of experts in which each spike is an expert. We first summarize the Product of Experts learning rule for a restricted Boltzmann machine (RBM) which consists of a layer of stochastic binary visible units connected to a layer of stochastic binary hidden units with no intralayer connections. We then extend RBM's to deal with temporal data. In an RBM, each hidden unit is an expert. When it is off it specifies a uniform distribution over the states of the visible units . When it is on, its weight to each visible unit specifies the log odds that the visible unit is on. Multiplying together the distributions specified by different hidden units is achieved by adding the log odds. Inference in an RBM is much easier than in a causal belief net because there is no explaining away. The hidden states, S j, are conditionally independent given the visible states, Si, and the distribution of Sj is given by the standard logistic function a: p(Sj = 1) = a(L:i WijSi). Conversely, the hidden states of an RBM are marginally dependent so it is easy for an RBM to learn population codes in which units may be highly correlated. It is hard to do this in causal belief nets with one hidden layer because the generative model of a causal belief net assumes marginal independence. An RBM can be trained by following the gradient of the log likelihood of the data: (2) where < SiSj >0 is the expected value of SiSj when data is clamped on the visible units and the hidden states are sampled from their conditional distribution given the data, and < SiSj >00 is the expected value of SiSj after prolonged Gibbs sampling that alternates between sampling from the conditional distribution of the hidden states given the visible states and vice versa. This learning rule not work well because the sampling noise in the estimate of < SiSj >00 swamps the gradient. It is far more effective to maximize the difference between the log likelihood of the data and the log likelihood of the one-step reconstructions of the data that are produced by first picking binary hidden states from their conditional distribution given the data and then picking binary visible states from their conditional distribution given the hidden states. The gradient of the log G. E. Hinton and A. D. Brown 126 likelihood of the one-step reconstructions is complicated because changing a weight changes the probability distribution of the reconstructions: + (3) where Ql is the distribution of the one-step reconstructions of the training data and Qoo is the equilibrium distribution (i.e. the stationary distribution of prolonged Gibbs sampling). Fortunately, the cumbersome third term is sufficiently small that ignoring it does not prevent the vector of weight changes from having a positive cosine with the true gradient of the difference of the log likelhoods so the following very simple learning rule works much better than Eq. 2. (4) 4 Restricted Boltzmann machines through time Using a restricted Boltzmann machine we can represent time by spatializing it, i.e. taking each visible unit, i, and hidden unit, j, and replicating them through time with the constraint that the weight WijT between replica t of i and replica t + T of j does not depend on t. To implement the desired temporal smoothing, we also force the weights to be a smooth function of T that has the shape of the temporal kernel, shown in Figure 3. The only remaining degree of freedom in the weights between replicas of i and replicas of j is the scale of the temporal kernel and it is this scale that is learned. The replicas of the visible and hidden units still form a bipartite graph and the probability distribution over the hidden replicas can be inferred exactly without considering data that lies further into the future than the width of the temporal kernel. One problem with the restricted Boltzmann machine when we spatialize time is that hidden units at one time step have no memory of their states at previous time steps; they only see the data. If we were to add undirected connections between hidden units at different time steps, then the architecture would return to a fully connected Boltzmann machine in which the hidden units are no longer conditionally independent given the data. A useful trick borrowed from Elman nets is to allow the hidden units to see their previous states, but to treat these observations like data that cannot be modified by future hidden states. Thus, the hidden states may still be inferred independently without resorting to Gibbs sampling. The connections between hidden layer weights also follow the time course of the temporal kernel. These connections act as a predictive prior over the hidden units. It is important to note that these forward connections are not required for the network to model a sequence, but only for the purposes of extrapolating into the future. Figure 3: The form of the temporal kernel. 127 Spiking Boltzmann Machines Now the probability that Sj(t) = 1 given the states of the visible units is, P(Sj(t) = 1) = u (~W,jh,(t) + ~ W,;h,(t)) . where hi(t) is the convolution of the history of visible unit i with the temporal kernel, 00 T=O and hk(t), the convolution of the hidden unit history, is computed similarly. 2 Learning the weights follows immediately from this formula for doing inference. In the positive phase the visible units are clamped at each time step and the posterior of the hidden units conditioned on the data is computed (we assume zero boundary conditions for time before t = 0). Then in the negative phase we sample from the posterior of the hidden units, and compute the distribution over the visible units at each time step given these hidden unit states. In each phase the correlations between the hidden and visible units are computed and the learning rule is, 00 AWij = 00 L L r(7) ((Sj(t)Si(t - 7))0 - (Sj(t)Si(t - 7))1) . t=O T=O 5 Results We trained this network on a sequence of 8x8 synthetic images of a Gaussian blob moving in a circular path. In the following diagrams we display the time sequence of images as a matrix. Each row of the matrix represents a single image with its pixels stretched out into a vector in scanline order, and each column is the time course of a single pixel. The intensity f the pixel is represented by the area of the white patch. We used 20 hidden units. Figure 5a shows a segment (200 time steps) of the time series which was used in training. In this sequence the period of the blob is 80 time steps. Figure 5b shows how the trained model reconstructs the data after we sample from the hidden layer units. Once we have trained the model it is possible to do forecasting by clamping visible layer units for a segment of a sequence and then doing iterative Gibbs sampling to generate future points in the sequence. Figure 5c shows that given 50 time steps from the series, the model can predict reasonably far into the future, before the pattern dies out. One problem with these simulations is that we are treating the real valued intensities in the images as probabilities. While this works for the blob images, where the values can be viewed as the probabilities of pixels in a binary image being on, this is not true for more natural images. 6 Discussion In our initial simulations we used a causal sigmoid belief network (SBN) rather than a restricted Boltzmann machine. Inference in an SBN is much more difficult than in an RBM. It requires Gibbs sampling or severe approximations, and even if a temporal kernel is used to ensure that a replica of a hidden unit at one time 2Computing the conditional probability distribution over the visible units given the hidden states is done in a similar fashion, with the caveat that the weights in each direction must be symmetric. Thus, the convolution is done using the reverse kernel. G. E. Hinton and A. D. Brown 128 a) c) b) Figure 4: a) The original data, b) reconstruction of the data, and c) prediction of the data given 50 time steps of the sequence. The black line indicates where the prediction begins. has no connections to replicas of visible units at very different times, the posterior distribution of the hidden units still depends on data far in the future. The Gibbs sampling made our SBN simulations very slow and the sampling noise made the learning far less effective than in the RBM. Although the RBM simulations seem closer to biological plausibility, they too suffer from a major problem. To apply the learning procedure it is necessary to reconstruct the data from the hidden states and we do not know how to do this without interfering with the incoming datastream. In our simulations we simply ignored this problem by allowing a visible unit to have both an observed value and a reconstructed value at the same time. Acknowledgements We thank Zoubin Ghahramani, Peter Dayan, Rich Zemel, Terry Sejnowski and Radford Neal for helpful discussions. This research was funded by grants from the Gatsby Foundation and NSERC. References Anderson, C.H. & van Essen, D.C (1994). Neurobiological computational systems. In J.M Zureda, R.J. Marks, & C.J. Robinson (Eds.), Computational Intelligence Imitating Life 213-222. New York: IEEE Press. Hinton, G. E. (1999) Products of Experts. ICANN 99: Ninth international conference on Artificial Neural Networks, Edinburgh, 1-6. Hinton, G. E., McClelland, J. L., & Rumelhart, D. E. (1986) Distributed representations. In Rumelhart, D. E. and McClelland, J. L., editors, Parallel Distributed Processing: Explorations in the Microstructure of Cognition. Volume 1: Foundations, MIT Press, Cambridge, MA. Hopfield, J. (1995). Pattern recognition computation using action potential timing for stimulus representation. Nature, 376, 33-36.
1724 |@word proportion:1 seems:1 simulation:7 covariance:5 solid:1 awij:1 initial:1 series:2 tuned:1 activation:2 si:3 conjunctive:1 must:3 distant:1 additive:1 visible:20 shape:1 treating:2 hourglass:3 extrapolating:1 stationary:1 generative:3 intelligence:1 caveat:1 coarse:2 toronto:2 location:2 consists:1 fitting:1 combine:3 expected:2 elman:1 brain:2 prolonged:2 actual:1 considering:1 becomes:1 provided:1 begin:1 what:1 temporal:15 act:1 exactly:2 k2:1 uk:2 control:1 unit:35 whatever:1 grant:1 positive:2 before:2 understood:1 timing:3 treat:3 path:1 black:1 chose:1 resembles:1 suggests:1 conversely:1 implement:1 communicated:1 procedure:1 area:1 zoubin:2 get:1 cannot:1 impossible:2 optimize:1 map:1 independently:1 convex:1 resolution:1 immediately:1 rule:4 population:4 swamp:1 exact:1 trick:2 velocity:1 rumelhart:3 recognition:1 bottom:1 disjunctive:4 observed:1 coincidence:1 sbn:3 region:2 connected:2 dynamic:1 trained:4 depend:1 segment:2 predictive:1 bipartite:1 basis:5 vague:1 multimodal:1 bowl:1 hopfield:2 emergent:1 represented:6 describe:1 london:2 effective:4 sejnowski:1 artificial:1 zemel:1 whose:2 quite:1 encoded:1 valued:7 reconstruct:1 ability:1 noisy:1 sequence:9 advantage:1 blob:3 net:4 ucl:1 reconstruction:5 interaction:1 product:3 kohonen:1 combining:1 organizing:1 mixing:1 wcin:1 produce:2 renormalizing:1 object:8 andrew:1 ac:1 borrowed:1 eq:1 c:1 direction:2 thick:1 stochastic:2 centered:2 occupies:1 exploration:1 require:1 microstructure:1 biological:1 sufficiently:1 equilibrium:1 cognition:1 predict:1 pointing:1 major:1 vary:1 purpose:1 perceived:1 vice:1 create:1 hope:1 mit:1 activates:1 gaussian:5 mation:1 modified:1 rather:2 dimple:4 qoo:1 varying:2 sisj:5 likelihood:6 indicates:1 hk:1 sense:1 helpful:1 inference:4 dependent:1 dayan:1 biochemical:1 hidden:32 semantics:1 infor:1 pixel:4 orientation:1 exponent:1 multiplies:1 spatial:3 smoothing:1 fairly:1 marginal:1 field:8 once:1 having:1 sampling:10 represents:1 broad:1 thin:1 future:6 stimulus:3 composed:1 simultaneously:1 individual:4 phase:3 attempt:1 freedom:2 detection:1 essen:2 highly:1 circular:1 severe:1 arrives:1 accurate:1 andy:1 closer:1 necessary:1 desired:2 causal:4 lwe:1 column:1 ar:1 cover:1 deviation:1 uniform:1 delay:1 too:1 synthetic:1 combined:1 density:2 peak:1 international:1 probabilistic:1 off:2 picking:2 together:4 quickly:3 central:1 reconstructs:1 convolving:1 expert:8 return:1 potential:2 coding:4 coefficient:6 depends:1 multiplicative:3 doing:2 scanline:1 start:1 complicated:1 parallel:1 contribution:2 om:2 formed:1 variance:1 percept:1 landscape:11 waist:1 bayesian:1 accurately:3 produced:1 none:1 marginally:1 multiplying:2 history:2 explain:1 strongest:1 cumbersome:1 synaptic:1 ed:1 energy:12 realise:1 obvious:1 rbm:10 sampled:1 higher:1 follow:1 done:3 anderson:2 just:2 implicit:2 correlation:1 logistic:1 wijsi:1 effect:2 brown:4 true:2 evolution:1 symmetric:1 semantic:1 neal:1 deal:1 conditionally:3 white:1 self:1 width:1 unnormalized:1 cosine:1 demonstrate:1 image:8 invoked:1 sigmoid:1 spiking:8 volume:1 extend:1 onedimensional:1 significant:1 versa:1 gibbs:6 cambridge:1 stretched:1 tuning:1 trivially:1 pm:1 resorting:1 similarly:1 replicating:1 funded:1 moving:2 longer:1 add:2 curvature:1 posterior:4 perspective:1 reverse:1 binary:6 life:1 fortunately:1 maximize:1 focussing:1 monotonically:1 period:1 ii:1 multiple:3 full:3 infer:1 smooth:5 plausibility:1 cross:1 post:1 bigger:1 prediction:2 represent:11 kernel:9 achieved:2 addition:1 diagram:1 source:1 unlike:1 undirected:1 flow:1 seem:1 odds:2 ideal:2 iii:1 easy:1 independence:1 architecture:1 forecasting:1 suffer:1 peter:1 york:1 action:1 ignored:1 useful:1 mcclelland:3 generate:1 specifies:3 dotted:1 neuroscience:1 correctly:1 per:1 modifiable:1 broadly:1 discrete:1 affected:1 threshold:1 changing:1 prevent:1 replica:8 graph:1 fraction:1 sum:1 inverse:4 parabolic:1 patch:1 oscillation:1 dy:1 comparable:1 layer:6 hi:1 display:1 activity:13 constraint:1 span:1 relatively:2 department:1 alternate:1 across:1 describes:1 postsynaptic:2 restricted:7 imitating:1 know:1 operation:3 apply:1 away:1 appropriate:1 convolved:2 original:1 top:1 assumes:1 remaining:1 ensure:1 ghahramani:2 objective:1 added:1 spike:12 receptive:7 gradient:4 thank:2 consensus:1 assuming:1 code:3 index:1 ratio:1 convincing:1 difficult:3 ql:1 sharper:2 negative:2 design:3 boltzmann:11 contributed:1 perform:1 allowing:1 neuron:24 observation:1 convolution:3 hinton:9 precise:3 frame:1 ninth:1 sharp:3 arbitrary:2 canada:1 intensity:2 inferred:2 required:3 specified:2 connection:6 learned:2 robinson:1 perception:1 pattern:2 summarize:1 memory:1 video:1 belief:4 terry:1 overlap:1 treated:1 force:1 natural:1 representing:1 scheme:2 x8:1 prior:2 acknowledgement:1 loss:1 fully:1 interesting:1 geoffrey:1 foundation:2 degree:2 sufficient:1 editor:1 interfering:1 row:1 course:3 allow:2 jh:1 fall:1 explaining:1 taking:1 distributed:4 van:2 boundary:1 dimension:4 edinburgh:1 rich:1 sensory:1 forward:1 made:2 far:5 sj:6 approximate:1 compact:1 reconstructed:1 neurobiological:1 active:3 instantiation:12 incoming:1 latent:1 iterative:1 why:1 nature:1 learn:2 reasonably:1 ca:1 ignoring:1 contributes:1 intralayer:1 domain:1 icann:1 whole:2 noise:3 convey:2 fashion:1 gatsby:3 slow:2 sub:1 position:1 lie:3 clamped:2 perceptual:1 third:1 formula:1 departs:1 utoronto:1 offset:1 decay:1 essential:1 adding:4 conditioned:1 clamping:1 easier:1 smoothly:1 intersection:1 simply:1 nserc:1 radford:1 corresponds:1 determines:2 ma:1 conditional:5 viewed:2 consequently:1 feasible:2 hard:1 change:2 called:1 datastream:1 college:1 mark:1 latter:1 correlated:1
792
1,725
A Winner-Take-All Circuit with Controllable Soft Max Property Shih-Chii Lin Institute for Neuroinformatics, ETHjUNIZ Winterthurstrasse 190, CH-8057 Zurich Switzerland shih@ini.phys.ethz.ch Abstract I describe a silicon network consisting of a group of excitatory neurons and a global inhibitory neuron. The output of the inhibitory neuron is normalized with respect to the input strengths. This output models the normalization property of the wide-field directionselective cells in the fly visual system. This normalizing property is also useful in any system where we wish the output signal to code only the strength of the inputs, and not be dependent on the number of inputs. The circuitry in each neuron is equivalent to that in Lazzaro's winner-take-all (WTA) circuit with one additional transistor and a voltage reference. Just as in Lazzaro's circuit, the outputs of the excitatory neurons code the neuron with the largest input. The difference here is that multiple winners can be chosen. By varying the voltage reference of the neuron, the network can transition between a soft-max behavior and a hard WTA behavior. I show results from a fabricated chip of 20 neurons in a 1.2J.Lm CMOS technology. 1 Introduction Lazzaro and colleagues (Lazzaro, 1988) were the first to implement a hardware model of a winner-take-all (WTA) network. This network consists of N excitatory cells that are inhibited by a global signal. Improvements of this network with addition of positive feedback and lateral connections have been described (Morris, 1998; Indiveri, 1998). The dynamics and stability properties of networks of coupled excitatory and inhibitory neurons have been analyzed by many (Amari, 1982; Grossberg, 1988). Grossberg described conditions under which these networks will exhibit WTA behavior. Lazzaro's network computes a single winner as reflected by the outputs of the excitatory cells. Several winners can be chosen by using more localized inhibition. In this work, I describe two variants of a similar architecture where the outputs of the excitatory neurons code the relative input strengths as in a soft-max computation. The relative values of the outputs depend on the number of inputs, their relative strengths and two parameter settings in the network. The global inhibitory 8.-c. Liu 718 Figure 1: Network model of recurrent inhibitory network. signal can also be used as an output. This output saturates with increasing number of active inputs, and the saturation level depends on the input strengths and parameter settings. This normalization property is similar to the normalization behavior of the wide-field direction-selective cells in the fly visual system. These cells code the temporal frequency of the visual inputs and are largely independent of the stimulation size. The circuitry in each neuron in the silicon network is equivalent to that in Lazzaro et. al.'s hard WTA network with an additional transistor and a voltage reference. By varying the voltage reference, the network can transition between a soft-max computation and a hard WTA computation. In the two variants, the outputs of the excitatory neurons either code the strength of the inputs or are normalized with respect to a constant bias current. Results from a fabricated network of 20 neurons in a 1.2J.Lm AMI CMOS show the different regimes of operation. 2 Network with Global Inhibition The generic architecture of a recurrent network with excitatory neurons and a single inhibitory neuron is shown in Figure 1. The excitatory neurons receive an external input, and they synapse onto a global inhibitory neuron. The inhibitory neuron, in turn, inhibits the excitatory neurons. The dynamics of the network is described as follows: dYi dt N = -Yi + ei - ~ g(~ WjYj) (1) j=l where Wj is the weight of the synapse between the jth excitatory neuron and the inhibitory neuron, and Yj is the state of the jth neuron. Under steady-state conditions, Yi = ei - YT, where YT = g(L:~l WjYj)? Assume a linear relationship between YT and Yj, and letting Wj = W, _ N ~ _ YT - W ~ Yj - "N W L..Jj=l ej 1 + wN j=l As N increases, YT = L:;l ej ? If all inputs have the same level, e, then YT = e. 719 A Winner-Take-All Circuit with Controllable Soft Max Property Figure 2: First variant of the architecture. Here we show the circuit for two excitatory neurons and the global inhibition neuron, M 4 ? The circuit in each excitatory neuron consists of an input current source, h, and transistors, M1 to M 3 . The inhibitory transistor is a fixed current source, lb . The inputs to the inhibitory transistor, 101 and I~2 are normalized with respect to lb. 3 First Variant of Network with Fixed Current Source In Sections 3 and 4, I describe two variants of the architecture shown in Figure 1. The two variants differ in the way that the inhibition signal is generated. The first network in Figure 2 shows the circuitry for two excitatory neurons and the inhibition neuron. Each excitatory neuron is a linear threshold unit and consists of an input current, h, and transistors, Ml, M 2 , and M 3 . The state of the neuron is represented by the current, I r1 . The diode-connected transistor, M 2 , introduces a rectifying nonlinearity into the system since Ir1 cannot be negative. The inhibition current, Ir, is sunk by M 1 , and is determined by the gate voltage, VT. The inhibition neuron consists of a current source, Ib, and VT is determined by the corresponding current, Ir1 and the corresponding transistor, M3 in each neuron. Notice that IT cannot be greater than the largest input to the network and the inputs to this network can only be excitatory. The input currents into the transistor, M 4 , are defined as 101 and 102 and are normalized with respect to the current source, h. In the hard WTA condition, the output current of the winning neuron is equal to the bias current, h. This network exhibits either a soft-maximum behavior or a hard WTA behavior depending on the value of an external bias, Va. The inhibition current, IT, is derived as: (2) where N is the number of "active" excitatory neurons (that is, neurons whose Ii > IT), Ii is the same input current to each neuron, and Ia = Ioe",vQ/uT. In deriving the above equation, we assumed that K, = 1. The inhibition current, IT, is Iri x Ial h? a linear combination of the states of the neurons because Ir = 2:f Figure 3(a) shows the response of the common-node voltage, VT, as a function of the number of inputs for different input values measured from a fabricated silicon network of 20 neurons. The input current to each neuron is provided by a pFET transistor that is driven by the gate voltage, Yin. All input currents are equal in this figure. The saturation behavior of the network as a function of the number s.-c. 720 Liu 0.8,---~-~-~-~-~----, Vin=3.9V -' . 0.7 0.6 Vin=4.3V .II'~''''--.'' 5 ......... ?? ' .. - - - ... ... ..... .. 10 15 20 Number of inputs (a) 25 30 5 10 15 20 25 30 Number of inputs (b) Figure 3: (a) Common-node voltage, VT, as a function of the number of input stimuli. Va = O.8V. (b) Common-node voltage, VT, as a function of the number of inputs with an input voltage of 4.3V and Vb = O.7V. The curves correspond to different values of Va . of inputs can be seen in the different traces and the saturation level increases as Vin decreases. As seen in Equation 2, the point at which the response saturates is dependent on the ratio, h / I a. In Figure 3(b), I show how the curve saturates at different points for different values of Va and a fixed hand Vin. In Figure 4, I set all inputs to zero except for two inputs, Vin1 and Vin2 that are set to the same value. I measured 101 and 101 as a function of Va as shown in Figure 4(a). The four curves correspond to four values of Vin. Initially both currents 101 and 102 are equal as is expected in the soft-max condition. As Va increases, the network starts exhibiting a WTA behavior. One of the output currents finally goes to zero above a critical value of Va. This critical value increases for higher input currents because of transistor backgate effects. In Figure 4(b), I show how the output currents respond as a function of the differential voltage between the two inputs as shown in Figure 4. Here, I fixed one input at 4.3V and swept the second input differentially around it. The different curves correspond to different values of Va. For a low value of Va, the linear differential input range is about lOOmV. This linear range decreases as Va is increased (corresponding to the WTA condition). 4 Second Variant with Diode-Connected Inhibition Transistor In the second variant shown in Figure 5, the current source, M4 is replaced by a diode-connected transistor and the output currents, 10i' follow the magnitude of the input currents. The inhibition current, Ir, can be expressed as follows: (3) where la is defined in Section 3. We sum Equation 3 over all neurons and assuming equal inputs, we get Ir = J'LJri x la. This equation shows that the feedback signal has a square root dependence on the neuron states. As we will see, this causes the feedback signal to saturate quickly with the number of inputs. 721 A Winner-Take-All Circuit with Controllable Soft Max Property 6 r-... 5 2.5 Va=O ~ ~4 ~ , 2 -: 1.5 ".\'0" 5~'~\ /~. " I Va=O 6V Va=O.7V Al\',/' ..s 2 < -0.2 l~ ' ~ ,. <\ / ) .. \ \> .,.~J \ \ '" , 0.5 Va=O.4V ,i -0.1 ,,~ 0 0.1 Vio2-Viol (V) 0.2 0.3 (b) (a) Figure 4: (a) Output currents, 101 and 1 02 , as a function of Va: for a subthreshold bias current and Yin = 4.0V to 4.3V. (b) Outputs, 101 and 1 02 , as a function of the differential input voltage, ~ Vin, with Yinl = 4.3V. Figure 5: Second variant of network. The schematic shows two excitatory neurons with diode-connected inhibition transistor. Substituting lri = Ii - IT in Equation 3, we solve for Jr, N IT = -1a: N + (Ia: N )2 + 41a: L Ii (4) From measurements from a fabricated circuit with 20 neurons, I show the dependence of VT (the natural logarithm of Jr) on the number of inputs in Figure 6(a). The output saturates quickly with the number of inputs and the level of saturation increases with increased input strengths. All the inputs have the same value. The network can also act as a WTA by changing Va:. Again, all inputs are set to zero except for two inputs whose gate voltages are both set at 4.2V. As shown in Figure 6(b), the output currents, 101 and 1 02 , are initially equal, and as Va: increases above 0.6V, the output currents split apart and eventually, 102 = OA. The final value of 101 depends on the maximum input current. This data shows that the network acts as a WTA circuit when Va: > 0.73V. If I set Vin2 = 4.25V instead, the output currents split at a lower value of Va:. s.-c. 722 Liu 0.45.--~-~-~-~-~------, 5 Vinl=4.2V. Vin2=4.25V 4 $ ~3 2 0.20'---:-2-~4-~6:--8=---1-:':0:----:'12 Number of inputs (a) 0.6 0.7 0.8 0.9 Va (V) (b) Figure 6: (a) Common-node voltage, VT, as a function of the number of inputs for input voltages, 3.9V, 4.06V, and 4.3V for Va = O.4V. (b) Outputs, 101 and 1 02 , as a function of Va for Vinl = 4.2V, Vin2 = 4.25V for the 2 curves with asterisks and for Vinl = Vin2 = 4.2V for the 2 curves with circles. 5 Inhibition The WTA property arises in both variants of this network if the gain parameter, Va, is increased so that the diode-connected transistor, M 2 , can be ignored. Both variants then reduce to Lazzaro's network. In the first variant, the feedback current (Ir) is a linear combination ofthe neuron states. However, when the gain parameter is increased so that M2 can be ignored, the feedback current is now a nonlinear combination of the input states so the WTA behavior is exhibited by these reduced networks. Under hard WTA conditions, if Ir is initially smaller than all the input currents, the capacitances C at the nodes Vr1 and Vr2 are charged up by the difference between the individual input current and IT, i.e., d~t = liCIT. Since the inhibition current is a linear combination of Iri and Iri is exponential in Vri , we can see that IT is a sum of the exponentials of the input currents, h Hence the feedback current is nonlinear in the input currents. Another way of viewing this condition in electronic terms is that in the soft WTA condition, the output node of each neuron is a softimpedance node, or a low-gain node. In the hard WTA case, the output node is now a high-impedance node or a high-gain node. Any input differences are immediately amplified in the circuit. 6 Discussion Hahnloser (Hahnloser, 1998) recently implemented a silicon network of linear threshold excitatory neurons that are coupled to a global inhibitory neuron. The inhibitory signal is a linear combination of the output states of the excitatory neurons. This network does not exhibit WTA behavior unless the excitatory neurons include a self-excitatory term. The inhibition current in his network is also generated via a diode-connected transistor. The circuitry in two variants described here is more compact than the circuitry in his network. Recurrent networks with the architecture described in this paper have been proposed by Reichardt and colleagues (Reichardt, 1983) in modelling the aggregation property A Winner-Take-All Circuit with Controllable Soft Max Property 723 of the wide-field direction-selective cells in flies. The synaptic inputs are inhibited by a wide-field cell that pools all the synaptic inputs. Similar networks have also been used to model cortical processing, for example, orientation selectivity (Douglas, 1995). The network implemented here can model the aggregation property of the directionselective cells in the fly. By varying a voltage reference, the network implements either a soft-max computation or a hard WTA computation. This circuitry will be useful in hardware models of cortical processing or motion processing in invertebrates. Acknowledgments I thank Rodney Douglas for supporting this work, and the MOSIS foundation for fabricating this circuit. I also thank Tobias Delbriick for proofreading this document. This work was supported in part by the Swiss National Foundation Research SPP grant and the U.S. Office of Naval Research. References Amari, S., and Arbib, M. A., "Competition and cooperation in neural networks," New York, Springer-Verlag, 1982. Grossberg, W., "Nonlinear neural networks: Principles, mechanisms, and architectures," Neural Networks, 1, 17-61, 1988. Hanhloser, R., "About the piecewise analysis of networks of linear threshold neurons," Neural Networks, 11,691- 697, 1988. Hahnloser, R., "Computation in recurrent networks of linear threshold neurons: Theory, simulation and hardware implementation," Ph.D. Thesis, Swiss Federal Institute of Technology, 1998. Lazzaro, J., Ryckebusch, S. Mahowald, M.A., and Mead. C., "Winner-take-all networks of O(n) complexity," In Tourestzky, D. (ed), Advances in Neural Information Processing Systems 1, San Mateo, CA: Morgan Kaufman Publishers, pp. 703-711, 1988. Morris, T .G., Horiuchi, T. and Deweerth, S.P., "Object-based selection within an analog VLSI visual attention system," IEEE Trans. on Circuits and Systems II, 45:12, 1564-1572, 1998. Indiveri, G., "Winner-take-all networks with lateral excitation," Neuromorphic Systems Engineering, Editor, Lande, TS., 367-380, Kluwer Academic, Norwell, MA, 1998. Reichardt, W., Poggio, T., and Hausen, K., "Figure-ground discrimination by relative movement in the visual system of the fly," BioI. Cybern., 46, 1-30, 1983. Douglas, RJ., Koch, C., Mahowald, M., Martin, KAC., and Suarez, HH., "Recurrent excitation in neocortical circuits," Science, 269:5226,981-985, 1995.
1725 |@word simulation:1 liu:3 document:1 current:41 discrimination:1 ial:1 fabricating:1 node:11 differential:3 consists:4 expected:1 behavior:10 increasing:1 provided:1 circuit:14 directionselective:2 kaufman:1 fabricated:4 temporal:1 act:2 unit:1 grant:1 positive:1 engineering:1 mead:1 mateo:1 range:2 grossberg:3 acknowledgment:1 yj:3 implement:2 swiss:2 get:1 onto:1 cannot:2 selection:1 vr2:1 cybern:1 equivalent:2 charged:1 yt:6 go:1 iri:3 attention:1 immediately:1 m2:1 ioe:1 deriving:1 his:2 stability:1 suarez:1 fly:5 wj:2 connected:6 decrease:2 movement:1 complexity:1 tobias:1 dynamic:2 depend:1 vr1:1 chip:1 represented:1 horiuchi:1 describe:3 neuroinformatics:1 whose:2 solve:1 amari:2 final:1 ir1:2 transistor:16 amplified:1 competition:1 differentially:1 r1:1 cmos:2 object:1 depending:1 recurrent:5 measured:2 vin2:5 implemented:2 diode:6 differ:1 switzerland:1 direction:2 exhibiting:1 viewing:1 around:1 koch:1 ground:1 lm:2 circuitry:6 substituting:1 largest:2 federal:1 ej:2 varying:3 voltage:16 sunk:1 office:1 derived:1 indiveri:2 naval:1 improvement:1 modelling:1 lri:1 dependent:2 initially:3 vlsi:1 selective:2 orientation:1 field:4 equal:5 stimulus:1 piecewise:1 inhibited:2 national:1 individual:1 m4:1 replaced:1 consisting:1 tourestzky:1 introduces:1 analyzed:1 dyi:1 norwell:1 poggio:1 unless:1 logarithm:1 circle:1 increased:4 soft:11 neuromorphic:1 mahowald:2 pool:1 quickly:2 again:1 thesis:1 external:2 depends:2 root:1 start:1 aggregation:2 vin:6 rodney:1 rectifying:1 square:1 ir:6 largely:1 correspond:3 subthreshold:1 ofthe:1 chii:1 phys:1 synaptic:2 ed:1 colleague:2 frequency:1 pp:1 gain:4 ut:1 higher:1 dt:1 follow:1 reflected:1 response:2 synapse:2 just:1 deweerth:1 hand:1 ei:2 nonlinear:3 effect:1 normalized:4 hence:1 self:1 steady:1 excitation:2 ini:1 neocortical:1 motion:1 recently:1 common:4 stimulation:1 winner:11 vin1:1 analog:1 m1:1 kluwer:1 silicon:4 measurement:1 nonlinearity:1 inhibition:15 driven:1 apart:1 selectivity:1 verlag:1 vt:7 yi:2 swept:1 seen:2 morgan:1 additional:2 greater:1 signal:7 ii:6 multiple:1 rj:1 academic:1 lin:1 va:23 schematic:1 lande:1 variant:13 normalization:3 cell:8 receive:1 addition:1 source:6 publisher:1 exhibited:1 split:2 wn:1 architecture:6 arbib:1 reduce:1 york:1 cause:1 lazzaro:8 jj:1 ignored:2 useful:2 morris:2 hardware:3 ph:1 reduced:1 kac:1 inhibitory:13 notice:1 group:1 shih:2 four:2 threshold:4 changing:1 douglas:3 mosis:1 sum:2 respond:1 electronic:1 vb:1 strength:7 invertebrate:1 proofreading:1 inhibits:1 martin:1 combination:5 pfet:1 jr:2 smaller:1 wta:19 equation:5 zurich:1 vq:1 turn:1 eventually:1 mechanism:1 hh:1 letting:1 operation:1 generic:1 gate:3 include:1 capacitance:1 ryckebusch:1 dependence:2 exhibit:3 thank:2 lateral:2 oa:1 assuming:1 code:5 relationship:1 ratio:1 trace:1 negative:1 implementation:1 neuron:50 t:1 supporting:1 saturates:4 lb:2 delbriick:1 vri:1 connection:1 trans:1 spp:1 regime:1 saturation:4 max:9 ia:2 critical:2 natural:1 technology:2 coupled:2 reichardt:3 relative:4 localized:1 asterisk:1 foundation:2 principle:1 editor:1 excitatory:22 cooperation:1 supported:1 jth:2 bias:4 institute:2 wide:4 feedback:6 curve:6 cortical:2 transition:2 computes:1 hausen:1 san:1 compact:1 ml:1 global:7 active:2 assumed:1 impedance:1 ca:1 controllable:4 wish:1 winning:1 exponential:2 ib:1 saturate:1 normalizing:1 magnitude:1 yin:2 visual:5 expressed:1 springer:1 ch:2 ma:1 hahnloser:3 bioi:1 hard:8 determined:2 except:2 ami:1 la:2 m3:1 arises:1 ethz:1
793
1,726
A Variational Bayesian Framework for Graphical Models Hagai Attias hagai@gatsby.ucl.ac.uk Gatsby Unit, University College London 17 Queen Square London WC1N 3AR, U.K. Abstract This paper presents a novel practical framework for Bayesian model averaging and model selection in probabilistic graphical models. Our approach approximates full posterior distributions over model parameters and structures, as well as latent variables, in an analytical manner. These posteriors fall out of a free-form optimization procedure, which naturally incorporates conjugate priors. Unlike in large sample approximations, the posteriors are generally nonGaussian and no Hessian needs to be computed. Predictive quantities are obtained analytically. The resulting algorithm generalizes the standard Expectation Maximization algorithm, and its convergence is guaranteed. We demonstrate that this approach can be applied to a large class of models in several domains, including mixture models and source separation. 1 Introduction A standard method to learn a graphical model 1 from data is maximum likelihood (ML). Given a training dataset, ML estimates a single optimal value for the model parameters within a fixed graph structure. However, ML is well known for its tendency to overfit the data. Overfitting becomes more severe for complex models involving high-dimensional real-world data such as images, speech, and text. Another problem is that ML prefers complex models, since they have more parameters and fit the data better. Hence, ML cannot optimize model structure. The Bayesian framework provides, in principle, a solution to these problems. Rather than focusing on a single model, a Bayesian considers a whole (finite or infinite) class of models. For each model, its posterior probability given the dataset is computed. Predictions for test data are made by averaging the predictions of all the individual models, weighted by their posteriors. Thus, the Bayesian framework avoids overfitting by integrating out the parameters. In addition, complex models are automatically penalized by being assigned a lower posterior probability, therefore optimal structures can be identified. Unfortunately, computations in the Bayesian framework are intractable even for lWe use the term 'model' to refer collectively to parameters and structure. H. Attias 210 very simple cases (e.g. factor analysis; see [2]). Most existing approximation methods fall into two classes [3]: Markov chain Monte Carlo methods and large sample methods (e.g., Laplace approximation). MCMC methods attempt to achieve exact results but typically require vast computational resources, and become impractical for complex models in high data dimensions. Large sample methods are tractable, but typically make a drastic approximation by modeling the 'posteriors over all parameters as Normal, even for parameters that are not positive definite (e.g., covariance matrices). In addition, they require the computation ofthe Hessian, which may become quite intensive. In this paper I present Variational Bayes (VB), a practical framework for Bayesian computations in graphical models. VB draws together variational ideas from intractable latent variables models [8] and from Bayesian inference [4,5,9], which, in turn, draw on the work of [6]. This framework facilitates analytical calculations of posterior distributions over the hidden variables, parameters and structures. The posteriors fall out of a free-form optimization procedure which naturally incorporates conjugate priors, and emerge in standard forms, only one of which is Normal. They are computed via an iterative algorithm that is closely related to Expectation Maximization (EM) and whose convergence is guaranteed. No Hessian needs to be computed. In addition, averaging over models to compute predictive quantities can be performed analytically. Model selection is done using the posterior over structure; in particular, the BIC/MDL criteria emerge as a limiting case. 2 General Framework We restrict our attention in this paper to directed acyclic graphs (DAGs, a.k.a. Bayesian networks). Let Y = {y., ... ,YN} denote the visible (data) nodes, where n = 1, ... , N runs over the data instances, and let X = {Xl, ... , XN} denote the hidden nodes. Let e denote the parameters, which are simply additional hidden nodes with their own distributions. A model with a fixed structure m is fully defined by the joint distribution p(Y, X, elm). In a DAG, this joint factorizes over the nodes, i.e. p(Y,X I e,m) = TIiP(Ui I pai,Oi,m), where Ui E YUX, pai is the set of parents of Ui, and Oi E e parametrize the edges directed toward Ui. In addition, we usually assume independent instances, p(Y, X Ie, m) = TIn p(y n, Xn Ie, m). We shall also consider a set of structures m E M, where m controls the number of hidden nodes and the functional forms of the dependencies p( Ui I pai , 0i, m), including the range of values assumed by each node (e.g., the number of components in a mixture model). Associated with the set of structures is a structure prior p(m). Marginal likelihood and posterior over parameters. For a fixed structure m, we are interested in two quantities. The first is the parameter posterior distribution p(e I Y,m). The second is the marginal likelihood p(Y I m), also known as the evidence assigned to structure m by the data. In the following, the reference to m is usually omitted but is always implied. Both quantities are obtained from the joint p(Y, X, elm). For models with no hidden nodes the required computations can often be performed analytically. However, in the presence of hidden nodes, these quantities become computationally intractable. We shall approximate them using a variational approach as follows. Consider the joint posterior p(X, elY) over hidden nodes and parameters. Since it is intractable, consider a variational posterior q(X, elY), which is restricted to the factorized form q(X, elY) = q(X I Y)q(e I Y) , (1) wher"e given the data, the parameters and hidden nodes are independent. This A Variational Baysian Frameworkfor Graphical Models 211 restriction is the key: It makes q approximate but tractable. Notice that we do not require complete factorization, as the parameters and hidden nodes may still be correlated amongst themselves. We compute q by optimizing a cost function Fm[q] defined by Fm[q] = ! dE> q(X)q(E? log ~~i~(:j ~ logp(Y I m) , (2) where the inequality holds for an arbitrary q and follows from Jensen's inequality (see [6]); it becomes an equality when q is the true posterior. Note that q is always understood to include conditioning on Y as in (1). Since Fm is bounded from above by the marginal likelihood, we can obtain the optimal posteriors by maximizing it w.r.t. q. This can be shown to be equivalent to minimizing the KL distance between q and the true posterior. Thus, optimizing Fm produces the best approximation to the true posterior within the space of distributions satisfying (1), as well as the tightest lower bound on the true marginal likelihood. Penalizing complex models. To see that the VB objective function Fm penalizes complexity, it is useful to rewrite it as Fm = (log p(Y, X I E? q(X) )X,9 - KL[q(E? II p(E?] , (3) where the average in the first term on the r.h.s. is taken w.r.t. q(X, E?. The first term corresponds to the (averaged) likelihood. The second term is the KL distance between the prior and posterior over the parameters. As the number of parameters increases, the KL distance follows and consequently reduces Fm. This penalized likelihood interpretation becomes transparent in the large sample limit N -7 00, where the parameter posterior is sharply peaked about the most probable value E> = E>o. It can then be shown that the KL penalty reduces to (I E>o 1/2) log N, which is linear in the number of parameters I E>o I of structure m. Fm then corresponds precisely the Bayesian information criterion (BIC) and the minimum description length criterion (MDL) (see [3]). Thus, these popular model selection criteria follow as a limiting case of the VB framework. Free-form optimization and an EM-like algorithm. Rather than assuming a specific parametric form for the posteriors, we let them fall out of free-form optimization of the VB objective function. This results in an iterative algorithm directly analogous to ordinary EM. In the E-step, we compute the posterior over the hidden nodes by solving 8Fm/8q(X) = 0 to get q(X) ex e(log p(Y,XI9?e , (4) where the average is taken w.r.t. q(E?. In the M-step, rather than the 'optimal' parameters, we compute the posterior distribution over the parameters by solving 8Fm/8q(E? = 0 to get q(E? ex e(IOgp(y,X I9?x p (E? , (5) where the average is taken w.r.t. q(X). This is where the concept of conjugate priors becomes useful. Denoting the exponential term on the r.h.s. of (5) by f(E?, we choose the prior p(E? from a family of distributions such that q(E? ex f(E?p(E? belongs to that same family. p(E? is then said to be conjugate to f(E?. This procedure allows us to select a prior from a fairly large family of distributions (which includes non-informative ones as limiting cases) H. Attias 212 and thus not compromise generality, while facilitating mathematical simplicity and elegance. In particular, learning in the VB framework simply amounts to updating the hyperparameters, i.e., transforming the prior parameters to the posterior parameters. We point out that, while the use of conjugate priors is widespread in statistics, so far they could only be applied to models where all nodes were visible. Structure posterior. To compute q(m) we exploit Jensen's inequality once again to define a more general objective function, .1'[q] l:mEM q(m) [.1'm + logp(m)jq(m)] ~ 10gp(Y), where now q = q(X I m, Y)q(8 I m, Y)q( m I Y) . After computing .1'm for each m EM, the structure posterior is obtained by free-form optimization of .1': q(m) ex e:Frnp(m) . (6) Hence, prior assumptions about the likelihood of different structures, encoded by the prior p(m), affect the selection of optimal model structures performed according to q( m), as they should. Predictive quantities. The ultimate goal of Bayesian inference is to estimate predictive quantities, such as a density or regression function. Generally, these quantities are computed by averaging over all models, weighting each model by its posterior. In the VB framework, exact model averaging is approximated by replacing the true posterior p(8 I Y) by the variational q(8 I Y). In density estimation, for example, the density assigned to a new data point Y is given by p(y I Y) = J d8 p(y I 8) q(8 I Y) . In some situations (e.g. source separation), an estimate of hidden node values x from new data y may be required. The relevant quantity here is the conditional p(x I y, Y), from which the most likely value of hidden nodes is extracted. VB approximates it by p(x I y, Y) ex J d8 p(y, x I 8) q(8 I Y). 3 Variational Bayes Mixture Models Mixture models have been investigated and analyzed extensively over many years. However, the well known problems of regularizing against likelihood divergences and of determining the required number of mixture components are still open. Whereas in theory the Bayesian approach provides a solution, no satisfactory practical algorithm has emerged from the application of involved sampling techniques (e.g., [7]) and approximation methods [3] to this problem. We now present the solution provided by VB. We consider models of the form m P(Yn I 8,m) = LP(Yn I Sn = s,8) p(sn = s I 8), (7) s=1 where Yn denotes the nth observed data vector, and Sn denotes the hidden component that generated it. The components are labeled by s = 1, ... , m, with the structure parameter m denoting the number of components. Whereas our approach can be applied to arbitrary models, for simplicity we consider here Normal component distributions, P(Yn I Sn = s, 8) = N(JJ. s' r 8), where I-Ls is the mean and r 8 the precision (inverse covariance) matrix. The mixing proportions are P(Sn = S I 8) = 'Tr s' In hindsight, we use conjugate priors on the parameters 8 = {'Trs, I-Ls' rs}. The mixing proportions are jointly Dirichlet, p( {'Trs}) = V(..~O), the means (conditioned on the preCisions) are Normal, p(l-Ls Irs) = N(pO, f3 0 r s), and the precisions are Wishart, p(r s) = W(v O, ~O). We find that the parameter posterior for a fixed m A Variational Baysian Framework/or Graphical Models 213 factorizes into q(8) = q({1I"s})flsq(J.?s,rs). The posteriors are obtained by the following iterative algorithm, termed VB-MOG. E-step. Compute the responsibilities for instance n using (4): = S I Yn) ex ffs r!/2 e-(Yn-P,)Tr,(Yn-P,)/2 e- d / 2f3? , (8) here X = S and q(S) = TIn q(Sn). This expression resembles the re- ,: == q(sn noting that sponsibilities in ordinary MLj the differences stem from integrating out the parameters. The special quantities in (8) are logffs == (log1l"s) = 1/1()..s) -1/1CLJs' )..s,), d logrs == (log I rs I) = Li=l1/1(lIs + 1 - i)/2) - log 1 ~s 1 +dlog2, and i\ == (r s) = IIs~;l, where 1/1(x) = dlog r(x)/dx is the digamma function, and the averages (-} are taken w.r.t. q(8). The other parameters are described below. M-step. Compute the parameter posterior in two stages. First, compute the quantities 1I"s 1 N " n , = N 'L..J's J.?s = 1 N " n Yn , N 'L..J's ~ Lis N = " n enS ' N1 'L..J's (9) n=l where C~ = (Yn - f.ts)(Yn - f.ts)T and fls = N7r s. This stage is identical to the M-step in ordinary EM where it produces the new parameters. In VB, however, the quantities in (9) only help characterize the new parameter posteriors. These posteriors are functionally identical to the priors but have different parameter values. The mixing proportions are jointly Dirichlet, q( {11" s}) = D( {)..s}), the means are Normal, q(J..t s Irs) = N(ps' /3srs), and the precisions are Wishart, p(rs) = W(lI s , ~s). The posterior parameters are updated in the second stage, using the simple rules )..s lis fls +)..0, = - Ns + 0 II, s n=1 /3s = fls + /30 , (10) aT a ~s = NsEs + N s/3 (J.?s - P )(I's - P ) /(Ns + f3 ) + ~ . Ps = (flsf.ts + /3opO)/(Ns +~) , s n=l - - - 0 - 0- ? The final values of the posterior parameters form the output of the VB-MOG. We remark that (a) Whereas no specific assumptions have been made about them, the parameter posteriors emerge in suitable, non-trivial (and generally non-Normal) functional forms. (b) The computational overhead of the VB-MOG compared to EM is minimal. (c) The covariance of the parameter posterior is O(l/N), and VBMOG reduces to EM (regularized by the priors) as N ~ 00. (d) VB-MOG has no divergence problems. (e) Stability is guaranteed by the existence of an objective function. (f) Finally, the approximate marginal likelihood F m , required to optimize the number of components via (6), can also be obtained in closed form (omitted). Predictive Density. Using our posteriors, we can integrate out the parameters and show that the density assigned by the model to a new data vector Y is a mixture of Student-t distributions, m (11) s=1 where component S has Ws = lis + 1 - d d.o.f., mean Ps' covariance As = ?/3s + 1)//3sws)~s, and proportion 7rs = )..s/ Ls' )..s" (11) reduces to a MOG as N ~ 00. Nonlinear Regression. We may divide each data vector into input and output parts, Y = (yi,y o ), and use the model to estimate the regression function yO = f(yi) and error spheres. These may be extracted from the conditional p(yO I yt, Y) = L:n=l Ws tw~ (yO I p~, A~), which also turns out to be a mixture of Student-t distributions, with means p~ being linear, and covariances A~ and mixing proportions Ws nonlinear, in yi, and given in terms of the posterior parameters. H Attias 214 Buffalo post offIce digits Misclasslflcation rate histogram 1 , - - - - -- -- - - - - -- , 0.8 0.6 0 .4 0 .2 0 0 ( ,, ,, -, - _1 - 0 .05 ~ , ,, ,, 0 .1 Figure 1: VB-MOG applied to handwritten digit recognition. VB-MOG was applied to the Boston housing dataset (UCI machine learning repository), where 13 inputs are used to predict the single output, a house's price. 100 random divisions of the N = 506 dataset into 481 training and 25 test points were used, resulting in an average MSE of 11.9. Whereas ours is not a discriminative method, it was nevertheless competitive with Breiman's (1994) bagging technique using regression trees (MSE=11.7). For comparison, EM achieved MSE=14.6. Classification. Here, a separate parameter posterior is computed for each class c from a training dataset yc. Test data vector y is then classified according to the conditional p(c I y, {yC}), which has a form identical to (11) (with c-dependent parameters) multiplied by the relative size of yc. VB-MOG was applied to the Buffalo post office dataset, which contains 1100 examples for each digit 0 - 9. Each digit is a gray-level 8 x 8 pixel array (see examples in Fig. 1 (left)). We used 10 random 500-digit batches for training, and a separate batch of 200 for testing. An average misclassification rate of .018 was obtained using m = 30 components; EM achieved .025. The misclassification histograms (VB=solid, EM=dashed) are shown in Fig. 1 (right). 4 VB and Intractable Models: a Blind Separation Example The discussion so far assumed that a free-form optimization of the VB objective function is feasible. Unfortunately, for many interesting models, in particular models where ordinary ML is intractable, this is not the case. For such models, we modify the VB procedure as follows: (a) Specify a parametric functional form for the posterior over the hidden nodes q(X) , and optimize w.r.t. its parameters, in the spirit of [8J. (b) Let the parameter posterior q(8) fall out of free-form optimization, as before. We illustrate this approach in the context of the blind source separation (BSS) problem (see, e.g., [1]). This problem is described by Yn = HX n + Un , where Xn is an unobserved m-dim source vector at instance n, H is an unknown mixing matrix, and the noise Un is Normally distributed with an unknown precision >'1. The task is to construct a source estimate xn from the observed d-dim data y. The sources are independent and non-Normally distributed. Here we assume the high-kurtosis distribution p(xi) ex: cosh-\xf /2) , which is appropriate for modeling speech sources. One important but heretofore unresolved problem in BSS is determining the number m of sources from data. Another is to avoid overfitting the mixing matrix. Both problems, typical to ML algorithms, can be remedied using VB. It is the non-Normal nature ofthe sources that renders the source posterior p(X I Y) intractable even before a Bayesian treatment. We use a Normal variational posterior q(X) = TIn N(x n lPn' r n) with instance-dependent mean and precision. The mixing matrix posterior q(H) then emerges as Normal. For simplicity, >. is optimized rather than integrated out. The reSUlting VB-BSS algorithm runs as follows: A Variational Baysian Framework for Graphical Models log PrIm) o source reconstruction errOr o ' -1000 -5 -2000 -10 -3000 -4000 215 -15 2 4 6 8 m 10 12 -roO~--~5~--~1-0--~ 15 SNR(dB) Figure 2: Application of VB to blind source separation algorithm (see text). E-step. Optimize the variational mean Pn by iterating to convergence, for each n, the fixed-point equation XfI:T(Yn - HPn) - tanhpn /2 = C- I Pn , where C is the source covariance conditioned on the data. The variational precision matrix turns out to be n-independent: r n = A. T AA. + 1/2 + C- I . M-step. Update the mean and precision of the posterior q(H) (rules omitted). This algorithm was applied to ll-dim data generated by linearly mixing 5 lOOmseclong speech and music signals obtained from commercial CDs. Gaussian noise were added at different SNR levels. A uniform structure prior p( m) = 1/ K for m ~ K was used. The resulting posterior over the number of sources (Fig. 2 (left)) is peaked at the correct value m = 5. The sources were then reconstructed from test data via p(x I y, Y). The log reconstruction error is plotted vs. SNR in Fig. 2 (right, solid). The ML error (which includes no model averaging) is also shown (dashed) and is larger, reflecting overfitting. 5 Conclusion The VB framework is applicable to a large class of graphical models. In fact, it may be integrated with the junction tree algorithm to produce general inference engines with minimal overhead compared to ML ones. Dirichlet, Normal and Wishart posteriors are not special to models treated here but emerge as a general feature. Current research efforts include applications to multinomial models and to learning the structure of complex dynamic probabilistic networks. Acknowledgements I thank Matt Beal, Peter Dayan, David Mackay, Carl Rasmussen, and especially Zoubin Ghahramani, for important discussions. References [1) Attias, H. (1999). Independent Factor Analysis. Neural Computation 11, 803-85l. [2) Bishop, C.M. (1999). Variational Principal Component Analysis. Proc. 9th ICANN. [3) Chickering, D.M. & Heckerman, D. (1997) . Efficient approximations for the marginal likelihood of Bayesian networks with hidden variables. Machine Learning 29, 181-212. [4) Hinton, G.E. & Van Camp, D. (1993). Keeping neural networks simple by minimizing the description length of the weights. Proc. 6th COLT, 5-13. [5) Jaakkola, T. & Jordan, M.L (1997). Bayesian logistic regression: A variational approach. Statistics and Artificial Intelligence 6 (Smyth, P. & Madigan, D., Eds). [6) Neal, R.M. & Hinton, G.E. (1998). A view of the EM algorithm that justifies incremental, sparse, and other variants. Learning in Graphical Models, 355-368 (Jordan, M.L, Ed). Kluwer Academic Press, Norwell, MA. [7) Richardson, S. & Green, P.J. (1997). On Bayesian analysis of mixtures with an unknown number of components. Journal of the Royal Statistical Society B, 59, 731-792. [8) Saul, L.K., Jaakkola, T., & Jordan, M.I. (1996). Mean field theory of sigmoid belief networks. Journal of Artificial Intelligence Research 4, 61-76. [9) Waterhouse, S., Mackay, D., & Robinson, T. (1996). Bayesian methods for mixture of experts. NIPS-8 (Touretzky, D.S. et aI, Eds). MIT Press.
1726 |@word repository:1 proportion:5 open:1 heretofore:1 r:5 covariance:6 tr:2 solid:2 contains:1 denoting:2 ours:1 existing:1 current:1 dx:1 visible:2 informative:1 update:1 v:1 intelligence:2 provides:2 node:16 mathematical:1 become:3 overhead:2 manner:1 themselves:1 automatically:1 becomes:4 provided:1 bounded:1 factorized:1 hindsight:1 unobserved:1 elm:2 impractical:1 uk:1 control:1 unit:1 normally:2 yn:13 positive:1 before:2 understood:1 modify:1 limit:1 resembles:1 factorization:1 range:1 averaged:1 directed:2 practical:3 testing:1 definite:1 digit:5 procedure:4 integrating:2 madigan:1 zoubin:1 get:2 cannot:1 selection:4 context:1 optimize:4 restriction:1 equivalent:1 yt:1 maximizing:1 attention:1 l:4 simplicity:3 rule:2 array:1 stability:1 laplace:1 limiting:3 updated:1 analogous:1 commercial:1 exact:2 smyth:1 carl:1 satisfying:1 approximated:1 updating:1 recognition:1 xi9:1 labeled:1 observed:2 wher:1 transforming:1 ui:5 complexity:1 dynamic:1 rewrite:1 solving:2 compromise:1 predictive:5 division:1 po:1 joint:4 london:2 monte:1 artificial:2 quite:1 whose:1 encoded:1 emerged:1 larger:1 roo:1 statistic:2 richardson:1 gp:1 jointly:2 final:1 beal:1 housing:1 analytical:2 kurtosis:1 ucl:1 reconstruction:2 unresolved:1 relevant:1 uci:1 mixing:8 achieve:1 description:2 convergence:3 parent:1 p:3 produce:3 incremental:1 help:1 illustrate:1 ac:1 closely:1 correct:1 require:3 hx:1 transparent:1 probable:1 hagai:2 hold:1 normal:10 predict:1 omitted:3 estimation:1 proc:2 applicable:1 weighted:1 mit:1 always:2 gaussian:1 rather:4 avoid:1 pn:2 breiman:1 factorizes:2 jaakkola:2 office:2 yo:3 likelihood:11 digamma:1 camp:1 dim:3 inference:3 dependent:2 dayan:1 typically:2 integrated:2 hidden:15 w:3 jq:1 interested:1 pixel:1 classification:1 colt:1 special:2 fairly:1 mackay:2 marginal:6 field:1 once:1 f3:3 construct:1 sampling:1 identical:3 pai:3 peaked:2 divergence:2 individual:1 n1:1 attempt:1 irs:2 severe:1 mdl:2 mixture:9 analyzed:1 chain:1 wc1n:1 norwell:1 edge:1 tree:2 divide:1 penalizes:1 re:1 plotted:1 minimal:2 lwe:1 instance:5 modeling:2 ar:1 logp:2 queen:1 maximization:2 ordinary:4 cost:1 snr:3 uniform:1 characterize:1 dependency:1 density:5 ie:2 probabilistic:2 together:1 nongaussian:1 again:1 choose:1 wishart:3 d8:2 expert:1 li:6 de:1 student:2 includes:2 ely:3 blind:3 performed:3 view:1 responsibility:1 closed:1 competitive:1 bayes:2 hpn:1 square:1 oi:2 ofthe:2 bayesian:17 handwritten:1 carlo:1 classified:1 touretzky:1 ed:3 against:1 involved:1 naturally:2 associated:1 elegance:1 dataset:6 treatment:1 popular:1 emerges:1 mlj:1 reflecting:1 focusing:1 follow:1 specify:1 done:1 generality:1 stage:3 overfit:1 replacing:1 nonlinear:2 widespread:1 logistic:1 gray:1 matt:1 concept:1 true:5 analytically:3 hence:2 assigned:4 equality:1 satisfactory:1 neal:1 ll:1 criterion:4 complete:1 demonstrate:1 l1:1 image:1 variational:15 novel:1 sigmoid:1 functional:3 multinomial:1 conditioning:1 interpretation:1 approximates:2 kluwer:1 functionally:1 refer:1 dag:2 ai:1 posterior:48 own:1 optimizing:2 belongs:1 termed:1 inequality:3 yi:3 minimum:1 additional:1 dashed:2 ii:3 signal:1 full:1 reduces:4 stem:1 frameworkfor:1 xf:1 academic:1 calculation:1 sphere:1 post:2 prediction:2 involving:1 regression:5 variant:1 expectation:2 mog:8 histogram:2 achieved:2 addition:4 whereas:4 source:15 unlike:1 sr:1 facilitates:1 db:1 incorporates:2 spirit:1 jordan:3 presence:1 noting:1 affect:1 fit:1 bic:2 identified:1 restrict:1 fm:10 idea:1 iogp:1 attias:5 intensive:1 xfi:1 expression:1 ultimate:1 effort:1 penalty:1 render:1 peter:1 speech:3 hessian:3 jj:1 prefers:1 remark:1 generally:3 useful:2 iterating:1 amount:1 extensively:1 cosh:1 notice:1 shall:2 key:1 nevertheless:1 penalizing:1 vast:1 graph:2 year:1 run:2 inverse:1 family:3 separation:5 i9:1 draw:2 vb:25 bound:1 fl:3 guaranteed:3 trs:2 precisely:1 sharply:1 according:2 conjugate:6 heckerman:1 em:11 lp:1 tw:1 restricted:1 dlog:1 taken:4 computationally:1 resource:1 equation:1 turn:3 tractable:2 drastic:1 generalizes:1 parametrize:1 tightest:1 junction:1 multiplied:1 appropriate:1 batch:2 existence:1 bagging:1 denotes:2 dirichlet:3 include:2 graphical:9 sw:1 music:1 exploit:1 ghahramani:1 especially:1 society:1 implied:1 objective:5 added:1 quantity:12 parametric:2 said:1 amongst:1 distance:3 separate:2 remedied:1 thank:1 considers:1 trivial:1 toward:1 assuming:1 length:2 minimizing:2 unfortunately:2 unknown:3 markov:1 finite:1 t:3 buffalo:2 situation:1 hinton:2 arbitrary:2 david:1 required:4 kl:5 optimized:1 baysian:3 engine:1 nip:1 robinson:1 usually:2 below:1 yc:3 including:2 green:1 royal:1 belief:1 suitable:1 misclassification:2 treated:1 regularized:1 nth:1 sn:7 text:2 prior:15 acknowledgement:1 determining:2 relative:1 waterhouse:1 fully:1 interesting:1 acyclic:1 integrate:1 principle:1 lpn:1 cd:1 penalized:2 free:7 rasmussen:1 keeping:1 fall:5 saul:1 emerge:4 sparse:1 distributed:2 van:1 dimension:1 xn:4 world:1 avoids:1 bs:3 made:2 far:2 reconstructed:1 approximate:3 dlog2:1 ml:9 overfitting:4 mem:1 assumed:2 discriminative:1 xi:1 un:2 latent:2 iterative:3 learn:1 nature:1 correlated:1 mse:3 investigated:1 complex:6 domain:1 icann:1 linearly:1 whole:1 noise:2 hyperparameters:1 facilitating:1 fig:4 en:1 gatsby:2 n:4 precision:8 exponential:1 xl:1 house:1 chickering:1 weighting:1 tin:3 tiip:1 specific:2 bishop:1 prim:1 jensen:2 evidence:1 intractable:7 conditioned:2 justifies:1 boston:1 simply:2 likely:1 collectively:1 aa:1 corresponds:2 extracted:2 ma:1 conditional:3 goal:1 consequently:1 price:1 feasible:1 infinite:1 typical:1 averaging:6 principal:1 tendency:1 select:1 college:1 mcmc:1 regularizing:1 ex:7
794
1,727
The Relaxed Online Maximum Margin Algorithm Yi Li and Philip M. Long Department of Computer Science National University of Singapore Singapore 119260, Republic of Singapore {liyi,p/ong}@comp.nus.edu.sg Abstract We describe a new incremental algorithm for training linear threshold functions: the Relaxed Online Maximum Margin Algorithm, or ROMMA. ROMMA can be viewed as an approximation to the algorithm that repeatedly chooses the hyperplane that classifies previously seen examples correctly with the maximum margin. It is known that such a maximum-margin hypothesis can be computed by minimizing the length of the weight vector subject to a number of linear constraints. ROMMA works by maintaining a relatively simple relaxation of these constraints that can be efficiently updated. We prove a mistake bound for ROMMA that is the same as that proved for the perceptron algorithm. Our analysis implies that the more computationally intensive maximum-margin algorithm also satisfies this mistake bound; this is the first worst-case performance guarantee for this algorithm. We describe some experiments using ROMMA and a variant that updates its hypothesis more aggressively as batch algorithms to recognize handwritten digits. The computational complexity and simplicity of these algorithms is similar to that of perceptron algorithm , but their generalization is much better. We describe a sense in which the performance of ROMMA converges to that of SVM in the limit if bias isn't considered. 1 Introduction The perceptron algorithm [10, 11] is well-known for its simplicity and effectiveness in the case of linearly separable data. Vapnik's support vector machines (SVM) [13] use quadratic programming to find the weight vector that classifies all the training data correctly and maximizes the margin, i.e. the minimal distance between the separating hyperplane and the instances. This algorithm is slower than the perceptron algorithm, but generalizes better. On the other hand, as an incremental algorithm, the perceptron algorithm is better suited for online learning, where the algorithm repeatedly must classify patterns one at a time, then finds out the correct classification, and then updates its hypothesis before making the next prediction. In this paper, we design and analyze a new simple online algorithm called ROMMA (the Relaxed Online Maximum Margin Algorithm) for classification using a linear threshold 499 The Relaxed Online Maximum Margin Algorithm function. ROMMA has similar time complexity to the perceptron algorithm, but its generalization performance in our experiments is much better on average. Moreover, ROMMA can be applied with kernel functions. We conducted experiments similar to those performed by Cortes and Vapnik [2] and Freund and Schapire [3] on the problem of handwritten digit recognition. We tested the standard perceptron algorithm, the voted perceptron algorithm (for details, see [3]) and our new algorithm, using the polynomial kernel function with d = 4 (the choice that was best in [3]). We found that our new algorithm performed better than the standard perceptron algorithm, had slightly better performance than the voted perceptron. For some other research with aims similar to ours, we refer the reader to [9,4,5,6]. The paper is organized as foIlows. In Section 2, we describe ROMMA in enough detail to determine its predictions, and prove a mistake bound for it. In Section 3, we describe ROMMA in more detail. In Section 4, we compare the experimental results of ROMMA and an aggressive variant of ROMMA with the perceptron and the voted perceptron algorithms. 2 2.1 A mistake-bound analysis The online algorithms For concreteness, our analysis will concern the case in which instances (also called patterns) and weight vectors are in R n . Fix n EN. In the standard online learning model [7], learning proceeds in trials. In the tth trial, the algorithm is first presented with an instance it ERn . Next, the algorithm outputs a prediction Yt of the classification of it. Finally, the algorithm finds out the correct classification Yt E {-1 , 1}. If Yt =I=- Yt, then we say that the algorithm makes a mistake. It is worth emphasizing that in this model, when making its prediction for the tth trial, the algorithm only has access to instance-classification pairs for previous trials. All of the online algorithms that we will consider work by maintaining a weight vector WI which is updated between trials, and predicting Yt sign( Wt . it), where sign( z) is 1 if z is positive, -1 if z is negative, and 0 otherwise.! = The perceptron algorithm. The perceptron algorithm, due to Rosenblatt [10, 11], starts off with Wi = O. When its prediction differs from the label Yt, it updates its weight vector by Wt+i = Wt + Ytit. If the prediction is correct then the weight vector is not changed. The next three algorithms that we will consider assume that all of the data seen by the online algorithm is collectively linearly separable, i.e. that there is a weight vector u such that for all each trial t, Yt = sign( u . xd. When kernel functions are used, this is often the case in practice. The ideal online maximum margin algorithm. On each trial t, this algorithm chooses a weight vector Wt for which for all previous trials s ::; t, sign( Wt . is) = Ys, and which maximizes the minimum distance of any is to the separating hyperplane. It is known [1, 14] that this can be implemented by choosing Wt to minimize Ilwdl subject to the constraints that Ys (Wt . xs ) ;::: 1 for all s ::; t. These constraints define a convex polyhedron in weight space which we will refer to as Pt. The relaxed online maximum margin algorithm. This is our new algorithm. The first difference is that trials in which mistakes are not made are ignored. The second difference 'The prediction of 0, which ensures a mistake, is to make the proofs simpler. The usual mistake bound proof for the perceptron algorithm goes through with this change. Y. Li and P. M Long 500 is in how the algorithm responds to mistakes. The relaxed algorithm starts off like the ideal algorithm. Before the second trial, it sets W2 to be the shortest weight vector such that Yl (W2 . i l ) 2:: 1. If there is a mistake on the second trial, it chooses W3 as would the ideal algorithm, to be the smallest element of (1) However, if the third trial is a mistake, then it behaves differently. Instead of choosing W4 to be the smallest element of {w: yI(w? i l ) 2:: I} n {w: Y2(W. i 2 ) 2:: I} n {w: Y3(W? i3) 2:: I} , it lets W4 be the smallest element of {w: W3 . W 2:: JJw3112} n {w: Y3(W. i3) 2:: I}. This can be thought of as, before the third trial, replacing the polyhedron defined by (1) with the halfspace {w : W3 ? W 2:: JJW3JJ2} (see Figure 1). Note that this halfspace contains the polyhedron of (1); in fact, it contains any convex set whose smallest element is W3. Thus, it can be thought of as the least restrictive convex constraint for which the smallest satisfying weight vector is W3. Let us call this halfspace H 3 . The algorithm continues in this manner. If the tth trial is a mistake, then Wt+l is chosen to be the smallest element of H t n {w : Yt(w? it) 2:: I}, and Ht+l is set to be {w : Wt+l . W 2:: IIwt+lJJ2}. If the tth trial is not a Wt and Ht+l H t . We will mistake, then Wt+l call H t the old constraint, and {w : Yt (w . it) 2: I} the new constraint. Figure 1: In ROMMA, a convex polyhedron in weight space is replaced with the halfspace with the same smallest element. Note that after each mistake, this algorithm needs only to solve a quadratic programming problem with two linear constraints. In fact, there is a simple closed-form expression for Wt+l as a function of Wt, it and Yt that enables it to be computed incrementally using time similar to that of the perceptron algorithm. This is described in Section 3. = = The relaxed online maximum margin algorithm with aggressive updating. The algorithm is the same as the previous algorithm, except that an update is made after any trial in which yt{Wt . it} < 1, not just after mistakes. 2.2 Upper bound on the number of mistakes made Now we prove a bound on the number of mistakes made by ROMMA. As in previous mistake bound proofs (e.g. [8]), we will show that mistakes result in an increase in a "measure of progress", and then appeal to a bound on the total possible progress. Our proof will use the squared length of Wt as its measure of progress. First we will need the following lemmas. Lemma 1 On any run of ROMMA on linearly separable data, if trial t was a mistake, then the new constraint is binding at the new weight vector; i.e. Yt (Wt+l . it) = 1. Proof: For the purpose of contradiction suppose the new constraint is not binding at the new weight vector Wt+l. Since Wt fails to satisfy this constraint, the line connecting Wt+l and Wt intersects with the border hyperplane of the new constraint, and we denote the aWt + (l-a)Wt+l, 0 < a < 1. intersecting point q. Then Wq can be represented as Wq w = 501 The Relaxed Online Maximum Margin Algorithm Since the square of Euclidean length II . 1\2 is a convex function, the following holds: IIwql\2 ~ allwtll 2 + (1 - a) IIwt+d2 Since Wt is the unique smallest member of H t and Wt+1 which implies i= Wt, we have IIwtl12 < IIwt+11l2, (2) Since Wt and Wt+1 are both in H t , Wq is too, and hence (2) contradicts the definition of 0 Wt+1? Lemma 2 On any run of ROMMA on linearly separable data, if trial t was a mistake, and not the first one, then the old constraint is binding at the new weight vector, i.e. Wt + 1 . Wt = IIwtV Proof: Let At be the plane of weight vectors that make the new constraint tight, i.e. At = {tV : Yt(w? xd = I}. By Lemma 1, Wt+1 E At . Let at = Ytxtlllxtll2 be the element of At that is perpendicular to it. Then each wE At satisfies IIwII 2 lIatll2 + IIw - at 1 2, and therefore the length of a vector W in At is minimized when W = at and is monotone in the distance from W to at. Thus, if the old constraint is not binding, then Wt+1 = at. since otherwise the solution could be improved by moving Wt+1 a little bit toward at. But the old constraint requires that (Wt . Wt+d 2: IIwtll2, and if Wt+1 = at Ytxtlllxtll2, this means that Wt . (YtxtlllxtIl2) 2: Ilwtll2. Rearranging, we get Yt(Wt . xd 2: IIxtll211wtlI2 > 0 (IIXtll > 0 follows from the fact that the data is linearly separable, and IIwt!\ > 0 follows from the fact that there was at least one previous mistake). But since trial t was a mistake, Yt (Wt . Xt) ~ 0, a contradiction. 0 = = Now we're ready to prove the mistake bound. Theorem 3 Choose mEN, and a sequence (Xl, Yd,???, (xm , Ym) of patternclassijicationpairsinRn x {-1,+1}. LetR = maxtl\xtli. Ifthereisaweightvector ii such that Yt (ii . Xt) 2: 1 for all 1 ~ t ~ m, then the number of mistakes made by ROMMA on (Xl, yd, .. . , (xm, Ym) is at most R211ii1l 2. Proof: First, we claim that for all t, ii E H t . This is easily seen since ii satisfies all the constraints that are ever imposed on a weight vector, and therefore all relaxations of such constraints. Since Wt is the smallest element of Ht. we have IIwtll ~ lliill. = = We have W2 Ylxdllid 2, and therefore IIw211 1/lIx1\\ 2:: 1/ R which implies IIw2112 2: 1/ R2. We claim that if any trial t > 1 is a mistake, then IIWt+1112 2: IIwtlI 2 + 1/ R2. This will imply by induction that after M mistakes, the squared length of the algorithm's weight vector is at least M / R2, which, since all of the algorithm's weight vectors are no longer than ii, will complete the proof. Choose an index t > 1 of a trial in which a mistake is made. Let At = {tV : Yt (w . it) = I} and B t = {w : (w . Wt) = IIwtIl2}. By Lemmas 1 and 2, Wt+1 EAt n B t . B The distance from Wt to At (call?it pe) satisfies IYt(xt . we) -11 Pt = I xtll 2: Figure 2: At, B t , and Pt 1 lIitll 2: 1 R' (3) since the fact that there was a mistake in trial t implies Yt(Xt . Wt) ~ O. Also, since Wt+1 E At. (4) 502 Y. Li and P. M. Long Because Wt is the normal vector of B t and Wt+1 E B t , we have + I Wt+1 - Wt1l 2. IIWt+d2 - IIWt 112 = IIWt+! IIWt+1112= IIWtll2 Thus, applying (3) and (4), we have which, as discussed above, completes the proof. - welI2 2: p; 2: 1/ R2, 0 Using the fact, easily proved using induction, that for all t, Pt ~ H t , we can easily prove the following, which complements analyses of the maximum margin algorithm using independence assumptions [1, 14, 12]. Details are omitted due to space constraints. Theorem 4 Choose mEN, and a sequence (x\, yd,"', (im , Ym) of patternclassification pairs in R n x {-I , +1}. Let R = maXt lIitli. If there is a weight vector ii such that Yt (ii . it) 2: 1 for all 1 ::; t ::; m, then the number of mistakes made by the ideal online maximum margin algorithm on (Xl, yd, .. " (xm, Ym) is at most R211ii1l 2. In the proof of Theorem 3, if an update is made and Yt (Wt . id < 1 - 0 instead of Yt (Wt . it) ::; 0, then the progress made can be seen to be at least 02/ R2. This can be applied to prove the following. Theorem 5 Choose 0 > 0, mEN, and a sequence (Xl, Y1) , ... , (X m, Ym) of patternclassification pairs in R n x {-I, + I}. Let R = maXt lIiell. If there is a weight vector ii such that Yt (ii . Xt) 2: 1 for aliI::; t ::; m, then if (i1' yI), ... , (im, Ym) are presented on line the number of trials in which aggressive ROMMA has Yt (Wt . it) < 1 - 0 is at most R2I1iiIl 2/0 2. Theorem 5 implies that, in a sense, repeatedly cycling through a dataset using aggressive ROMMA will eventually converge to SVM; note however that bias is not considered. 3 An efficient implementation When the prediction of ROMMA differs from the expected label, the algorithm chooses Wt+! to minimize IIWt+!1I subject to AWt+! = b, where A = (~f) and b = ( 11~:"2 ) . Simple calculation shows that Wt+! AT (AAT)-lb ("wtIl2(Yt - (Wt ? it)) ) ~ ( IIxtII211Wtll2 - Yt(Wt . it)) ~ lIitll211Wtll2 - (Wt . ie)2 Wt + IIxtll211Wtll2 _ (Wt . XtP Xt? If on trials t in which a mistake is made, Ct (5) = Since the computations required by ROMMA involve inner products together with a few operations on scalars, we can apply the kernel method to our algorithm, efficiently solving the original problem in a very high dimensional space. Computationally, we only need to modify the algorithm by replacing each inner product computation (ii . Xj) with a kernel function computation IC (ii, Xj). To make a prediction for the tth trial, the algorithm must compute the inner product between Xt and prediction vector Wt. In order to apply the kernel function, as in [1, 3], we store each prediction vector Wt in an implicit manner, as the weighted sum of examples on which 503 The Relaxed Online Maximum Margin Algorithm mistakes occur during the training. In particular. each Wt is represented as Wt = n t-l (t_ l ) (t-l) n IT WI + L djxj Cj J=1 J=1 C n=J+l (Wt+l The above formula may seem daunting; however, making use of the recurrence ?x) = Ct (Wt . x) + dt (Xt . x). it is obvious that the complexity of our new algorithm is similar to that of perceptron algorithm. This was born out by our experiments. The implementation for aggressive ROMMA is similar to the above. 4 Experiments We did some experiments using the ROMMA and aggressive ROMMA as batch algorithms on the MNIST OCR database. 2 We obtained a batch algorithm from our online algorithm in. the usual way, making a number of passes over the dataset and using the final weight vector to classify the test data. Every example in this database has two parts, the first is a 28 x 28 matrix which represents the image of the corresponding digit. Each entry in the matrix takes value from {O, . . . , 255}. The second part is a label taking a value from {O,? .. , g} . The dataset consists of 60, 000 training examples and 10,000 test examples. We adopt the following polynomial kernel: K(Xi, Xj) = (1 + (Xi? Xj))d. This corresponds to using an expanded collection of features including all products of at most d components of the original feature vector (see [14]). Let us refer to the mapping from the original feature vector to the expanded feature vector as <1>. Note that one component of <I> (x) is always 1, and therefore the component of the weight vector corresponding to that component can be viewed as a bias. In our experiments, we set WI = <1>(6') rather than (5 to speed up the learning of the coefficient corresponding to the bias. We chose d = 4 since in experiments on the same problem conducted in [3 , 2], the best results occur with this value. To cope with multiclass data, we trained ROMMA and aggressive ROMMA once for each of the 10 labels. Classification of an unknown pattern is done according to the maximum output of these ten classifiers. As every entry in the image matrix takes value from {O , ? .. , 255}, the order of magnitude of K(x , x) is at least 10 26 , which might cause round-off error in the computation of Ci and di . We scale the data by dividing each entry with 1100 when training with ROMMA. Table 1: Experimental results on MNIST data T=l Err MisNo percep voted-percep ROMMA agg-ROMMA agg-ROMMACNC) 2.84 2.26 2.48 2.14 2.05 7970 7970 7963 6077 5909 T=2 Err MisNo 2.27 1.88 1.96 1.82 1.76 10539 10539 9995 7391 6979 T=3 Err MisNo 1.99 1.76 1.79 1.71 1.67 11945 11945 10971 7901 7339 T =4 Err MisNo 1.85 1.69 1.77 1.67 1.63 12800 12800 11547 8139 7484 Since the performance of online learning is affected by the order of sample sequence, all the results shown in Table 1 average over 10 random permutations. The columns marked 2National Institute for Standards and Technology, special database http://www.research.att.com/... yanniocr for information on obtaining this dataset. 3. See 504 Y Li and P M. Long "MisNo" in Table 1 show the total number of mistakes made during the training for the 10 labels. Although online learning would involve only one epoch, we present results for a batch setting until four epochs (T in Table 1 represents the number of epochs). To deal with data which are linearly inseparable in the feature space, and also to improve generalization, Friess et al [4] suggested the use of quadratic penalty in the cost function, which can be implemented using a slightly different kernel function [4, 5]: iC(Xk ' Xj) = K(Xk, Xj) + c5kj ).., where c5kj is the Kronecker delta function. The last row in Table 1 is the result of aggressive ROMMA using this method to control noise ().. = 30 for 10 classifiers). We conducted three groups of experiments, one for the perceptron algorithm (denoted "percep"), the second for the voted perceptron (denoted "voted-percep") whose description is in [3], the third for ROMMA, aggressive ROMMA (denoted "agg-ROMMA"), and aggressive ROMMA with noise control (denoted "agg-ROMMA(NC)"). Data in the third group are scaled. All three groups set 'lih = <1>(0). The results in Table 1 demonstrate that ROMMA has better performance than the standard perceptron, aggressive ROMMA has slightly better performance than the voted perceptron. Aggressive ROMMA with noise control should not be compared with perceptrons without noise control. Its presentation is used to show what performance our new online algorithm could achieve (of course it's not the best, since all 10 classifiers use the same )"). A remarkable phenomenon is that our new algorithm behaves very well at the first two epochs. References [1] B. E. Boser, I. M. Guyon, and V. N. Vapnik. A training algorithm for optimal margin classifiers. Proceedings of the 1992 Workshop on Computational Learning Theory, pages 144-152, 1992. [2] C. Cortes and V. Vapnik. Support-vector networks. Machine Learning, 20(3):273-297,1995. [3] y. Freund and R. E. Schapire. Large margin classification using the perceptron algorithm. Proceedings of the 1998 Conference on Computational Learning Theory, 1998. [4] T. T. Friess, N. Cristianini, and C. Campbell. The kernel adatron algorithm: a fast and simple learning procedure for support vector machines. In Proc. 15th Int. Con! on Machine Learning. Morgan Kaufman Publishers, 1998. [5] S. S. Keerthi, S. K. Shevade, C. Bhattacharyya, and K. R. K. Murthy. A fast iterative nearest point algorithm for support vector machine c1assiifer design. Technical report, Indian Institute of Science, 99. TR-ISL-99-03 . [6] Adam Kowalczyk. Maximal margin perceptron. In Smola, Bartlett, Scholkopf, and Schuurmans, editors, Advances in Large Margin Classifiers, 1999. MIT-Press. [7] N. Littlestone. Learning quickly when irrelevant attributes abound: a new linear-threshold algorithm. Machine Learning, 2:285-318, 1988. [8] N. Littlestone. Mistake Bounds and Logarithmic Linear-threshold Learning Algorithms. PhD thesis, UC Santa Cruz, 1989. [9] John C. Platt. Fast training of support vector machines using sequential minimal optimization. In B. Scholkopf, C. Burges, A. Smola, editors, Advances in Kernel Methods: Support Vector Machines , 1998. MIT Press. [10] F. Rosenblatt. The perceptron: A probabilistic model for information storage and organization in the brain. Psychological Review, 65:386-407, 1958. [11] F. Rosenblatt. Principles of Neurodynamics: Perceptrons and the Theory of Brain Mechanisms . Spartan Books, Washington, D. C., 1962. [12] J. Shawe-Taylor, P. Bartlett, R. Williamson, and M. Anthony. A framework for structural risk minimization. In Proc. of the 1996 Conference on Computational Learning Theory, 1996. [13] V. N. Vapnik. Estimation of Dependencies based on Empirical Data. Springer Verlag, 1982. [14] V. N. Vapnik. The Nature of Statistical Learning Theory . Springer, 1995.
1727 |@word trial:25 polynomial:2 d2:2 t_:1 tr:1 born:1 contains:2 att:1 ours:1 bhattacharyya:1 err:4 percep:4 com:1 must:2 john:1 cruz:1 enables:1 update:5 plane:1 xk:2 simpler:1 scholkopf:2 prove:6 consists:1 manner:2 expected:1 brain:2 little:1 abound:1 classifies:2 moreover:1 maximizes:2 what:1 kaufman:1 guarantee:1 adatron:1 y3:2 every:2 xd:3 classifier:5 scaled:1 platt:1 control:4 before:3 positive:1 aat:1 modify:1 mistake:34 limit:1 id:1 yd:4 might:1 chose:1 perpendicular:1 unique:1 practice:1 differs:2 digit:3 procedure:1 empirical:1 w4:2 thought:2 get:1 storage:1 risk:1 applying:1 www:1 imposed:1 yt:25 go:1 convex:5 simplicity:2 contradiction:2 updated:2 pt:4 suppose:1 programming:2 hypothesis:3 element:8 recognition:1 satisfying:1 updating:1 continues:1 fries:2 database:3 worst:1 ensures:1 complexity:3 cristianini:1 ong:1 trained:1 tight:1 solving:1 easily:3 differently:1 represented:2 intersects:1 fast:3 describe:5 spartan:1 choosing:2 whose:2 solve:1 say:1 otherwise:2 final:1 online:20 sequence:4 product:4 maximal:1 achieve:1 description:1 incremental:2 converges:1 adam:1 nearest:1 progress:4 dividing:1 implemented:2 implies:5 correct:3 attribute:1 fix:1 generalization:3 im:2 hold:1 considered:2 ic:2 normal:1 mapping:1 claim:2 inseparable:1 adopt:1 smallest:9 omitted:1 purpose:1 estimation:1 proc:2 label:5 weighted:1 minimization:1 mit:2 always:1 aim:1 i3:2 rather:1 polyhedron:4 sense:2 letr:1 i1:1 classification:7 denoted:4 special:1 uc:1 once:1 washington:1 represents:2 minimized:1 report:1 few:1 national:2 recognize:1 replaced:1 keerthi:1 organization:1 lih:1 old:4 euclidean:1 littlestone:2 re:1 taylor:1 minimal:2 psychological:1 instance:4 classify:2 column:1 cost:1 republic:1 entry:3 conducted:3 too:1 dependency:1 chooses:4 ie:1 probabilistic:1 off:3 yl:1 connecting:1 ym:6 together:1 quickly:1 intersecting:1 squared:2 thesis:1 choose:4 book:1 li:4 aggressive:12 coefficient:1 int:1 satisfy:1 performed:2 closed:1 analyze:1 start:2 halfspace:4 minimize:2 square:1 voted:7 efficiently:2 handwritten:2 comp:1 worth:1 murthy:1 definition:1 obvious:1 proof:10 di:1 con:1 proved:2 dataset:4 organized:1 cj:1 campbell:1 dt:1 improved:1 daunting:1 done:1 just:1 implicit:1 smola:2 until:1 shevade:1 hand:1 replacing:2 incrementally:1 y2:1 hence:1 aggressively:1 deal:1 round:1 during:2 recurrence:1 complete:1 demonstrate:1 image:2 iiw:1 behaves:2 discussed:1 refer:3 shawe:1 had:1 moving:1 access:1 longer:1 irrelevant:1 store:1 verlag:1 yi:3 seen:4 minimum:1 morgan:1 relaxed:9 determine:1 shortest:1 converge:1 ii:12 technical:1 calculation:1 long:4 y:2 prediction:11 variant:2 kernel:10 completes:1 publisher:1 w2:3 romma:39 pass:1 subject:3 member:1 effectiveness:1 seem:1 call:3 structural:1 ideal:4 enough:1 independence:1 xj:6 w3:5 inner:3 multiclass:1 intensive:1 expression:1 bartlett:2 penalty:1 cause:1 repeatedly:3 ignored:1 santa:1 involve:2 ten:1 tth:5 schapire:2 http:1 singapore:3 sign:4 delta:1 correctly:2 rosenblatt:3 affected:1 group:3 four:1 threshold:4 ht:3 relaxation:2 concreteness:1 monotone:1 sum:1 run:2 reader:1 guyon:1 bit:1 bound:11 ct:2 quadratic:3 occur:2 constraint:19 kronecker:1 speed:1 separable:5 eat:1 expanded:2 relatively:1 ern:1 department:1 tv:2 according:1 slightly:3 contradicts:1 iiwii:1 wi:4 making:4 computationally:2 previously:1 awt:2 eventually:1 mechanism:1 generalizes:1 operation:1 apply:2 ocr:1 kowalczyk:1 batch:4 slower:1 original:3 maintaining:2 restrictive:1 usual:2 responds:1 cycling:1 distance:4 separating:2 philip:1 toward:1 induction:2 length:5 index:1 iyt:1 minimizing:1 nc:1 negative:1 design:2 implementation:2 unknown:1 upper:1 ever:1 y1:1 lb:1 isl:1 complement:1 pair:3 required:1 boser:1 nu:1 suggested:1 proceeds:1 pattern:3 xm:3 including:1 predicting:1 improve:1 technology:1 imply:1 ready:1 isn:1 alii:1 epoch:4 sg:1 l2:1 review:1 freund:2 permutation:1 men:3 remarkable:1 principle:1 editor:2 maxt:2 row:1 course:1 changed:1 last:1 bias:4 burges:1 perceptron:24 institute:2 taking:1 made:11 collection:1 iiwt:10 cope:1 xi:2 iterative:1 table:6 neurodynamics:1 nature:1 rearranging:1 obtaining:1 schuurmans:1 williamson:1 anthony:1 did:1 linearly:6 border:1 noise:4 en:1 fails:1 xtp:1 xl:4 pe:1 third:4 theorem:5 emphasizing:1 formula:1 xt:8 appeal:1 x:1 svm:3 cortes:2 r2:5 concern:1 workshop:1 mnist:2 vapnik:6 sequential:1 ci:1 phd:1 magnitude:1 margin:19 suited:1 logarithmic:1 agg:4 scalar:1 collectively:1 binding:4 springer:2 corresponds:1 satisfies:4 viewed:2 marked:1 presentation:1 change:1 except:1 hyperplane:4 wt:63 lemma:5 called:2 total:2 experimental:2 perceptrons:2 wq:3 support:6 indian:1 tested:1 phenomenon:1
795
1,728
Wiring optimization in the brain Dmitri B. Chklovskii Sloan Center for Theoretical Neurobiology The Salk Institute La Jolla, CA 92037 mitya@salk.edu Charles F. Stevens Howard Hughes Medical Institute and Molecular Neurobiology Lab The Salk Institute La Jolla, CA 92037 stevens@salk.edu Abstract The complexity of cortical circuits may be characterized by the number of synapses per neuron. We study the dependence of complexity on the fraction of the cortical volume that is made up of "wire" (that is, ofaxons and dendrites), and find that complexity is maximized when wire takes up about 60% of the cortical volume. This prediction is in good agreement with experimental observations. A consequence of our arguments is that any rearrangement of neurons that takes more wire would sacrifice computational power. Wiring a brain presents formidable problems because of the extremely large number of connections: a microliter of cortex contains approximately 105 neurons, 109 synapses, and 4 km ofaxons, with 60% of the cortical volume being taken up with "wire", half of this by axons and the other half by dendrites. [ 1] Each cortical neighborhood must have exactly the right balance of components; if too many cell bodies were present in a particular mm cube, for example, insufficient space would remain for the axons, dendrites and synapses. Here we ask "What fraction of the cortical volume should be wires (axons + dendrites)?" We argue that physiological properties ofaxons and dendrites dictate an optimal wire fraction of 0.6, just what is actually observed. To calculate the optimal wire fraction, we start with a real cortical region containing a fixed number of neurons, a mm cube, for example, and imagine perturbing it by adding or subtracting synapses and the axons and dendrites needed to support them. The rules for perturbing the cortical cube require that the existing circuit connections and function remain intact (except for what may have been removed in the perturbation), that no holes are created, and that all added (or subtracted) synapses are typical of those present; as wire volume is added, the volume of the cube of course increases. The ratio of the number of synapses per neuron in the perturbed cortex to that in the real cortex is denoted by 8, a parameter we call the relative complexity. We require that the volume of non-wire components (cell bodies, blood vessels, glia, etc) is unchanged by our perturbation and use 4> to denote the volume fraction of the perturbed cortical region that is made up of wires (axons + dendrites; 4> can vary between zero and one), with the fraction for the real brain being 4>0. The relation between relative complexity 8 and wire volume fraction 4> is given by the equation (derived in Methods) 1 (1-4?2/34> -,\5 1 - 4>0 4>0? 8-- (I) D. B. Chklovskii and C. F Stevens 104 2 -- :\ = .9 ( I) ?1 )( ..!! Q.. :\=1 1 E o u O~----r---~-----.----~--~ 0.0 0.2 0.4 0.6 0.8 1.0 Wire fraction (+) Figure I: Relative complexity (0) as a function of volume wire fraction (e/?. The graphs are calculated from equation (1) for three values of the parameter A as indicated; this parameter determines the average length of wire associated with a synapse (relative to this length for the real cortex, for which (A = 1). Note that as the average length of wire per synapse increases, the maximum possible complexity decreases. For the following discussion assume that A = 1; we return to the meaning of this parameter later. To derive this equation two assumptions are made. First, we suppose that each added synapse requires extra wire equal to the average wire length and volume per synapse in the unperturbed cortex. Second, because adding wire for new synapses increases the brain volume and therefore increases the distance axons and dendrites must travel to maintain the connections they make in the real cortex, all of the dendrite and unmyelinated axon diameters are increased in proportion to the square of their length changes in order to maintain the intersynaptic conduction times[2] and dendrite cable lengths[3] as they are in the actual cortex. If the unmyelinated axon diameters were not increased as the axons become longer, for example, the time for a nerve impulse to propagate from one synapse to the next would be increased and we would violate our rule that the existing circuit and its function be unchanged. We note that the vast majority of cortical axons are unmyelinated.[l] The plot of as a function of e/> is parabolic-like (see Figure 1) with a maximum value at e/> = 0.6, a point at which dO/de/> = O. This same maximum value is found for any possible value of e/>o, the real cortical wire fraction. o Why does complexity reach a maximum value at a particular wire fraction? When wire and synapses are added, a series of consequences can lead to a runaway situation we call the wiring catastrophe. If we start with a wire fraction less than 0.6, adding wire increases the cortical volume, increased volume makes longer paths for axons to reach their targets which requires larger diameter wires (to keep conduction delays or cable attenuation constant from one point to another), the larger wire diameters increase cortex volume which means wires must be longer, etc. While the wire fraction e/> is less than 0.6, increasing complexity is accompanied by finite increases in e/>. At e/> = 0.6 the rate at which wire fraction increases with complexity becomes infinite de/>/dO --t 00); we have reached the wiring catastrophe. At this point, adding wire becomes impossible without decreasing complexity or making other changes - like decreasing axon diameters - that alter cortical function. The physical cause of the catastrophe is a slow growth of conduction velocity and dendritic cable length with diameter combined with the requirement that the conduction times between synapses (and dendrite cable lengths) be unchanged in the perturbed cortex. We assumed above that each synapse requires a certain amount of wire, but what if we could 105 Wiring Optimization in the Brain add new synapses using the wire already present? We do not know what factors determine the wire volume needed to support a synapse, but if the average amount of wire per synapse could be less (or more) than that in the actual cortex, the maximum wire fraction would still be 0.6. Each curve in Figure 1 corresponds to a different assumed average wire length required for a synapse (determined by A), and the maximum always occurs at 0.6 independent of A. In the following we consider only situations in which A is fixed. For a given A, what complexity should we expect for the actual cortex? Three arguments favor the maximum possible complexity. The greatest complexity gives the largest number of synapses per neuron and this permits more bits of information to be represented per neuron. Also, more synapses per neuron decreases the relative effect caused by the loss or malfunction of a single synapse. Finally, errors in the local wire fraction would minimally affect the local complexity because d(} / dqJ = 0 at if> = 0.6. Thus one can understand why the actual cortex has the wire fraction we identify as optimal. [ 1] This conclusion that the wire fraction is a maximum in the real cortex has an interesting consequence: components of an actual cortical circuit cannot be rearranged in a way that needs more wire without eliminating synapses or reducing wire diameters. For example, if intermixing the cell bodies of left and right eye cells in primate primary visual cortex (rather than separating them in ocular dominance columns) increased the average length of the wire[4] the existing circuit could not be maintained just by a finite increase in volume. This happens because a greater wire length demanded by the rearrangement of the same circuit would require longer wire per synapse, that is, an increased A. As can be seen from Figure I, brains with A > 1 can never achieve the complexity reached at the maximum of the A = 1 curve that corresponds to the actual cortex. Our observations support the notion that brains are arranged to minimize wire length. This idea, dating back to Cajal[5], has recently been used to explain why retinotopic maps exist[6],[7], why cortical regions are separated, why ocular dominance columns are present in primary visual cortex[4],[8],[9] and why the cortical areas and flat worm ganglia are placed as they are. [ 10-13] We anticipate that maximal complexity/minimal wire length arguments will find further application in relating functional and anatomical properties of brain. Methods The volume of the cube of cortex we perturb is V, the volume of the non-wire portion is W (assumed to be constant), the fraction of V consisting of wires is if>, the total number of synapses is N, the average length of axonal wire associated with each synapse is s, and the average axonal wire volume per unit length is h; the corresponding values for dendrites are indicated by primes (s' and h'). The unperturbed value for each variable has a 0 subscript; thus the volume of the cortical cube before it is perturbed is Vo = Wo + No(soho + s~h~). (2) We now define a "virtual" perturbation that we use to explore the extent to which the actual cortical region contains an optimal fraction of wire. If we increase the number of synapses by a factor () and the length of wire associated with each synapse by a factor A, then the perturbed cortical cube's volume becomes Vo = Wo + A(} (Nosoho ~ + Nos~h~ ~~) (V/VO)1/3 . (3) This equation allows for the possibility that the average wire diameter has been perturbed and increases the length of all wire segments by the "mean field" quantity (V/VO)1/3 to take account of the expansion of the cube by the added wire; we require our perturbation disperses the added wire as uniformly as possible throughout the cortical cube. D. B. Chklovskii and C. F. Stevens 106 To simplify this relation we must eliminate h/ ho and h' / h&; we consider these terms in tum. When we perturb the brain we require that the average conduction time (s/u, where u is the conduction velocity) from one synapse to the next be unchanged so that s/u so/uo, or = ~ uo = !.... = '\so(V/VO)1/3 = '\(V/VO)1/3. So (4) So Because axon diameter is proportional to the square of conduction velocity u and the axon volume per unit length h is proportional to diameter squared, h is proportional to u 4 and the ratio h/ ho can be written as (5) For dendrites, we require that their length from one synapse to the next in units of the cable length constant be unchanged by the perturbation. The dendritic length constant is proportional to the square root of the dendritic diameter d, so 8/.fd = 80/ ViIO or ~ = (;~)2 = (,\(V/VO)1/3f =,\2(V/Vo)2/3. (6) Because dendritic volumes per unit length (h and h') vary as the square of the diameters, we have that ~~ = (~) 2 =,\4 (V/VO)4/3. (7) The equation (2) can thus be rewritten as V = Wo + No(soh o + s~h~)B,\5 (V/Vo)5/3 . (8) Divide this equation by Vo, define v = VIVo, and recognize that Wo/Vo that 4>0 = No(soho + s&h&)/Vo ; the result is = (1 - 4>0) and (9) Because the non-wire volume is required not to change with the perturbation, we know that 4?; substitute this in equation (9) and rearrange to give Wo = (1 - 4>o)Vo = (1 - 4?V which means that v = (1 - 4>0)/(1 - B = ~ ( 1,\5 4> ) 2/3 .!t 1 - 4>0 4>0 . (1) the equation used in the main text. We have assumed that conduction velocity and the dendritic cable length constant vary exactly with the square root of diameter[2],[ 14] but if the actual power were to deviate slightly from 112 the wire fraction that gives the maximum complexity would also differ slightly from 0.6. Acknowledgments This work was supported by the Howard Hughes Medical Institute and a grant from NIH to C.F.S. D.C. was supported by a Sloan Fellowship in Theoretical Neurobiology. Wiring Optimization in the Brain 107 References [1] Braitenberg. V. & Schuz. A. Cortex: Statistics and Geometry of Neuronal Connectivity (Springer. 1998). [2] Rushton. W.A.H. A Theory of the Effects of Fibre Size in Medullated Nerve. J. Physiol. 115. 101-122 (1951). [3] Bekkers. J.M. & Stevens. C.F. Two different ways evolution makes neurons larger. Prog Brain Res 83. 37-45 (1990). [4] Mitchison. G. Neuronal branching patterns and the economy of cortical wiring. Proc R Soc Lond B Bioi Sci 245.151-158 (1991). [5] Cajal. S.R.Y. Histology of the Nervous System 1-805 (Oxford University Press, 1995). [6] Cowey. A. Cortical maps and visual perception: the Grindley Memorial Lecture. Q J Exp Phychol 31. 1-17 (1979). [7] Allman J.M. & Kaas J.H. The organization of the second visual area (V II) in the owl monkey: a second order transformation of the visual hemifield. Brain Res 76: 247-65 (1974). [8] Durbin. R. & Mitchison. G. A dimension reduction framework for understanding cortical maps. Nature 343. 644-647 (1990). [9] Mitchison. G. Axonal trees and cortical architecture. Trends Neurosci 15. 122-126 (1992). [10] Young. M.P. Objective analysis of the topological organization of the primate cortical visual system. Nature 358. 152-154 (1992). [11] Cherniak. C. Loca1 optimization of neuron arbors. Bioi Cybern 66. 503-510 (1992). [12] Cherniak. C. Component placement optimization in the brain. J Neurosci 14.2418-2427 (1994). [13J Cherniak. C. Neural component placement. Trends Neurosci 18.522-527 (1995). [14J Rall. W. in Handbook of Physiology, The Nervous Systems, Cellular Biology of Neurons (ed. Brookhart. J.M.M .? V.B.) 39-97 (Am. Physiol. Soc .? Bethesda. MD. 1977).
1728 |@word eliminating:1 proportion:1 km:1 propagate:1 reduction:1 contains:2 series:1 existing:3 must:4 written:1 physiol:2 plot:1 half:2 nervous:2 become:1 sacrifice:1 brain:13 decreasing:2 rall:1 actual:8 increasing:1 becomes:3 retinotopic:1 cherniak:3 circuit:6 formidable:1 what:6 monkey:1 transformation:1 attenuation:1 growth:1 exactly:2 unit:4 medical:2 uo:2 grant:1 before:1 rushton:1 local:2 consequence:3 oxford:1 subscript:1 path:1 approximately:1 minimally:1 acknowledgment:1 hughes:2 area:2 physiology:1 dictate:1 cannot:1 impossible:1 cybern:1 map:3 center:1 rule:2 notion:1 imagine:1 suppose:1 target:1 agreement:1 velocity:4 trend:2 observed:1 calculate:1 region:4 decrease:2 removed:1 complexity:18 segment:1 represented:1 separated:1 neighborhood:1 larger:3 favor:1 statistic:1 subtracting:1 maximal:1 achieve:1 requirement:1 derive:1 soc:2 differ:1 stevens:5 runaway:1 virtual:1 owl:1 require:6 dendritic:5 anticipate:1 mm:2 exp:1 viio:1 vary:3 proc:1 travel:1 largest:1 soh:1 always:1 rather:1 derived:1 am:1 economy:1 eliminate:1 relation:2 denoted:1 cube:9 equal:1 field:1 never:1 biology:1 alter:1 braitenberg:1 simplify:1 cajal:2 recognize:1 geometry:1 consisting:1 maintain:2 rearrangement:2 organization:2 fd:1 possibility:1 rearrange:1 tree:1 divide:1 re:2 theoretical:2 minimal:1 increased:6 column:2 delay:1 too:1 conduction:8 perturbed:6 combined:1 hemifield:1 connectivity:1 squared:1 containing:1 return:1 account:1 de:2 accompanied:1 caused:1 sloan:2 later:1 root:2 lab:1 kaas:1 reached:2 start:2 portion:1 vivo:1 intermixing:1 square:5 minimize:1 maximized:1 identify:1 explain:1 synapsis:15 reach:2 ed:1 ocular:2 associated:3 ask:1 malfunction:1 actually:1 back:1 nerve:2 tum:1 synapse:15 arranged:1 just:2 indicated:2 impulse:1 effect:2 evolution:1 wiring:7 branching:1 maintained:1 ofaxons:3 vo:14 meaning:1 recently:1 charles:1 nih:1 functional:1 physical:1 perturbing:2 soho:2 volume:25 relating:1 bekkers:1 cortex:18 longer:4 etc:2 add:1 histology:1 jolla:2 prime:1 certain:1 cowey:1 seen:1 greater:1 determine:1 ii:1 violate:1 memorial:1 characterized:1 molecular:1 prediction:1 cell:4 fellowship:1 chklovskii:3 extra:1 call:2 axonal:3 allman:1 affect:1 architecture:1 idea:1 wo:5 cause:1 amount:2 diameter:13 rearranged:1 exist:1 per:12 anatomical:1 dominance:2 blood:1 vast:1 graph:1 dmitri:1 fraction:21 fibre:1 prog:1 parabolic:1 throughout:1 bit:1 topological:1 durbin:1 placement:2 flat:1 argument:3 extremely:1 lond:1 glia:1 remain:2 slightly:2 bethesda:1 cable:6 making:1 primate:2 mitya:1 happens:1 taken:1 equation:8 needed:2 know:2 rewritten:1 permit:1 subtracted:1 ho:2 substitute:1 schuz:1 perturb:2 unchanged:5 objective:1 added:6 already:1 occurs:1 quantity:1 primary:2 dependence:1 md:1 distance:1 separating:1 sci:1 majority:1 argue:1 extent:1 cellular:1 length:23 insufficient:1 ratio:2 balance:1 neuron:11 wire:57 observation:2 howard:2 finite:2 situation:2 neurobiology:3 perturbation:6 required:2 connection:3 pattern:1 perception:1 power:2 greatest:1 eye:1 created:1 dating:1 text:1 deviate:1 understanding:1 relative:5 loss:1 expect:1 lecture:1 interesting:1 proportional:4 course:1 placed:1 supported:2 understand:1 institute:4 curve:2 calculated:1 cortical:25 dimension:1 made:3 keep:1 handbook:1 assumed:4 mitchison:3 demanded:1 why:6 nature:2 ca:2 dendrite:13 expansion:1 vessel:1 main:1 neurosci:3 body:3 neuronal:2 salk:4 axon:14 slow:1 young:1 unperturbed:2 physiological:1 adding:4 hole:1 explore:1 ganglion:1 visual:6 catastrophe:3 springer:1 corresponds:2 determines:1 bioi:2 change:3 typical:1 except:1 infinite:1 determined:1 reducing:1 uniformly:1 worm:1 total:1 experimental:1 la:2 arbor:1 intact:1 support:3
796
1,729
Topographic Transformation as a Discrete Latent Variable Nebojsa Jojic Beckman Institute University of Illinois at Urbana www.ifp.uiuc.edu/",jojic Brendan J. Frey Computer Science University of Waterloo www.cs.uwaterloo.ca/ ... frey Abstract Invariance to topographic transformations such as translation and shearing in an image has been successfully incorporated into feedforward mechanisms, e.g., "convolutional neural networks", "tangent propagation". We describe a way to add transformation invariance to a generative density model by approximating the nonlinear transformation manifold by a discrete set of transformations. An EM algorithm for the original model can be extended to the new model by computing expectations over the set of transformations. We show how to add a discrete transformation variable to Gaussian mixture modeling, factor analysis and mixtures of factor analysis. We give results on filtering microscopy images, face and facial pose clustering, and handwritten digit modeling and recognition. 1 Introduction Imagine what happens to the point in the N-dimensional space corresponding to an N-pixel image of an object, while the object is deformed by shearing. A very small amount of shearing will move the point only slightly, so deforming the object by shearing will trace a continuous curve in the space of pixel intensities. As illustrated in Fig. la, extensive levels of shearing will produce a highly nonlinear curve (consider shearing a thin vertical line ), although the curve can be approximated by a straight line locally. Linear approximations of the transformation manifold have been used to significantly improve the performance of feedforward discriminative classifiers such as nearest neighbors (Simard et al., 1993) and multilayer perceptrons (Simard et al., 1992). Linear generative models (factor analysis, mixtures of factor analysis) have also been modified using linear approximations of the transformation manifold to build in some degree of transformation invariance (Hinton et al., 1997). In general, the linear approximation is accurate for transformations that couple neighboring pixels, but is inaccurate for transformations that couple nonneighboring pixels. In some applications (e.g., handwritten digit recognition), the input can be blurred so that the linear approximation becomes more robust. For significant levels of transformation, the nonlinear manifold can be better modeled using a discrete approximation. For example, the curve in Fig. 1a can be N. Jojic and B. J. Frey 478 (b) (c) (e) (d) p(z) ~ Figure 1: (a) An N-pixel greyscale image is represented by a point (unfilled disc) in an Ndimensional space. When the object being imaged is deformed by shearing. the point moves along a continuous curve. Locally. the curve is linear. but high levels of shearing produce a highly nonlinear curve. which we approximate by discrete points (filled discs) indexed bye. (b) A graphical model showing how a discrete transformation variable can be added to a density model p(z) for a latent image z to model the observed image x . The Gaussian pdf p(xle, z) captures the eth transformation plus a small amount of pixel noise. (We use a box to represent variables that have Gaussian conditional pdfs.) We have explored (c) transformed mixtures of Gaussians. where c is a discrete cluster index; (d) transformed component analysis (TeA). where y is a vector of Gaussian factors. some of which may model locally linear transformation perturbations; and (e) mixtures of transformed component analyzers. or transformed mixtures of factor analyzers. e represented by a set of points (filled discs). In this approach, a discrete set of possible transformations is specified beforehand and parameters are learned so that the model is invariant to the set of transformations. This approach has been used to design "convolutional neural networks" that are invariant to translation (Le Cun et al., 1998) and to develop a general purpose learning algorithm for generative topographic maps (Bishop et al., 1998) . We describe how invariance to a discrete set of known transformations (like translation) can be built into a generative density model and we show how an EM algorithm for the original density model can be extended to the new model by computing expectations over the set of transformations. We give results for 5 different types of experiment involving translation and shearing. 2 Transformation as a Discrete Latent Variable We represent transformation f by a sparse transformation generating matrix G e that operates on a vector of pixel intensities. For example, integer-pixel translations of an image can be represented by permutation matrices. Although other types of transformation matrix may not be accurately represented by permutation matrices, many useful types of transformation can be represented by sparse transformation matrices. For example, rotation and blurring can be represented by matrices that have a small number of nonzero elements per row (e.g., at most 6 for rotations). The observed image x is linked to the nontransformed latent image z and the transformation index f E {I, ... , L} as follows: p(xlf, z) = N(x; Gez, w), (1) where W is a diagonal matrix of pixel noise variances. Since the probability of a transformation may depend on the latent image, the joint distribution over the latent image z, the transformation index f and the observed image x is p(x, f, z) = N(x; Gez, w)P(flz)p(z). (2) The corresponding graphical model is shown in Fig. lb. For example, to model noisy transformed images of just one shape, we choose p(z) to be a Gaussian distribution. Topographic Transformation as a Discrete Latent Variable 479 2.1 Transformed mixtures of Gaussians (TMG). Fig. lc shows the graphical model for a TMG, where different clusters may havp. different transformation probabilities. Cluster c has mixing proportion 7rc , mean /-t c and diagonal covariance matrix ~ c. The joint distribution is (3) where the probability of transformation f for cluster c is Plc. Marginalizing over the latent image gives the cluster/transformation conditional likelihood, (4) which can be used to compute p(x) and the cluster/transformation responsibility P(f, clx). This likelihood looks like the likelihood for a mixture of factor analyzers (Ghahramani and Hinton, 1997). However, whereas the likelihood computation for N latent pixels takes order N 3 time in a mixture of factor analyzers, it takes linear time, order N, in a TMG, because Gl~cG'I + W is sparse. 2.2 Transformed component analysis (TCA). Fig. Id shows the graphical model for TCA (or "transformed factor analysis"). The latent image is modeled using linearly combined Gaussian factors, y. The joint distribution is p(x, f, z, y) = N(x; Glz, w)N(z; /-t + Ay, ~ )N(y; 0, I)Pl, (5) where /-t is the mean of the latent image, A is a matrix of latent image components (the factor loading matrix) and ~ is a diagonal noise covariance matrix for the latent image. Marginalizing over the factors and the latent image gives the transformation conditional likelihood, p(xlf) = N(x; Gl/-t, Gl(AA T + ~)G'I + w), (6) which can be used to compute p(x) and the transformation responsibility p(flx). + ~)G'I is not sparse, so computing this likelihood exactly takes N 3 time. However, the likelihood can be computed in linear time if we assume IGl(AA T + f))G'I + wi ~ IGl(AAT + ~)G'II, which corresponds to assuming that the observed noise is smaller than the variation due to the latent image, or that the observed noise is accounted for by the latent noise model, ~. In our experiments, this approximation did not lead to degenerate behavior and produced useful models. Gl(AA T By setting columns of A equal to the derivatives of /-t with respect to continuous transformation parameters, a TCA can accommodate both a local linear approximation and a discrete approximation to the transformation manifold. 2.3 Mixtures of transformed component analyzers (MTCA). A combination of a TMG and a TCA can be used to jointly model clusters, linear components and transformations. Alternatively, a mixture of Gaussians that is invariant to a discrete set of transformations and locally linear transformations can be obtained by combining a TMG with a TCA whose components are all set equal to transformation derivatives. The joint distribution for the combined model in Fig. Ie is p(x, f, z, c, y) = N(x; GlZ, w)N(z; /-t c + AcY, ~c)N(y; 0, I)Plc7rc. The cluster/transformation likelihood is p(xlf,c) = N(X;Gl/-tc,Gl(AcA; ~c)G'I + w), which can be approximated in linear time as for TCA. (7) + 480 3 N. Jojic and B. J. Frey Mixed Transformed Component Analysis (MTCA) We present an EM algorithm for MTCA; EM algorithms for TMG or TCA emerge by setting the number of factors to 0 or setting the number of clusters to 1. Let 0 represent a parameter in the generative model. For Li.d. data, the derivative of the log-likelihood of a training set Xl, ... ,XT with respect to 0 can be written 810gp(XI80, ... ,XT) IJ)I ] = ~ E [8 80 logp(xt, c, (., Z, Y Xt , T '"' (8) t=l where the expectation is taken over p(c, f, z, ylxt). The EM algorithm iteratively solves for a new set of parameters using the old parameters to compute the expectations. This procedure consistently increases the likelihood of the training data. By setting (8) to 0 and solving for the new parameter values, we obtain update equations based on the expectations given in the Appendix. Notation: (-) = .!. Ei=l (.) is a sufficient statistic computed by averaging over the training set; diag(A) gives a vector containing the diagonal elements of matrix A; diag(a) gives a diagonal matrix whose diagonal contains the elements of vector a; and a 0 h gives the element-wise product of vectors a and h. Denoting the updated parameters by "-", we have ire = (P(cJXt)), he = (P(flxt, c)), (9) _ (P(clxt)E[z - AeyIXt,c]) J.?e = (P(cIXt)) , (10) diag( (P( cIXt)E[(z - J.?e - AeY) 0 (z - J.?e - Aey)IXt, cD) ~e = (P(cIXt)) , (11) ~ = diag( (E[(Xt -GiZ)O(Xt - Giz)IXtD), (12) - Ae = (P(cJxdE[(z - J.?e)yTlxtl)(P(cIXt)E[yyTlxtD-I. (13) To reduce the number of parameters, we will sometimes assume Pic does not depend on c or even that Pic is held constant at a uniform distribution. 4 Experiments 4.1 Filtering Images from a Scanning Electron Microscope (SEM). SEM images (e.g., Fig. 2a) can have a very low signal to noise ratio due to a high variance in electron emission rate and modulation of this variance by the imaged material (Golem and Cohen, 1998). To reduce noise, multiple images are usually averaged and the pixel variances can be used to estimate certainty in rendered structures. Fig. 2b shows the estimated means and variances of the pixels from 230 140 x 56 SEM images like the ones in Fig. 2a. In fact, averaging images does not take into account spatial uncertainties and filtering in the imaging process introduced by the electron detectors and the high-speed electrical circuits. We trained a single-cluster TMG with 5 horizontal shifts and 5 vertical shifts on the 230 SEM images using 30 iterations of EM. To keep the number of parameters almost equal to the number of parameters estimated using simple averaging, the transformation probabilities were not learned and the pixel variances in the observed image were set equal after each M step. So, TMG had 1 more parameter. Fig. 2c shows the mean and variance learned by the TMG. Compared to simple averaging, the TMG finds sharper, more detailed structure. The variances are significantly lower, indicating that the TMG produces a more confident estimate of the image. 481 Topographic Transformation as a Discrete Latent Variable (a) (b) (e) Figure 2: (a) 140 x 56 pixel SEM images. (b) The mean and variance of the image pixels. (c) The mean and variance found by a TMG reveal more structure and less uncertainty. (a) -.- f I .~ ,. !. I\:~ .. --- f= , ... VI -t., -, --i..--.-~.. -.. .... , ~ -~_1~_. "'t~ - . ~ r f :s.- .,f~ ,.. -' tr I. .. ~ ~ <,.\ Sf' ~ ~'-,fll" .."iIIio . . !~V '. ~ .. ~;'''; ..,,,,-L ~- r . .. . -.~ ~ ~ . ,.?:" ~I ,.". , .:p. ~ f" ...-" . . . ,-,;I iI' '.j (d) t l~ ijIt . , . .l ...... . :or -. ,. ,,~: y: .", . .'-?, FO ,.,io ~ ~ ! .. .~ .<t --.,j .jIo ;.'~. ....# ... r; ~. ..,. .,. ..... ~ .~. ...... O ' J .. ...,.. ... ........ ~ . .', , ,:.~ ? (e) (b) (f) (e) (g) Figure 3: (a) Frontal face images of two people. (b) Cluster means learned by a TMG and (c) a mixture of Gaussians. (d) Images of one person with different poses. (e) Cluster means learned by a TMG. (f) Less detailed cluster means learned by a mixture of Gaussians. (g) Mean and first 4 principal components of the data. which mostly model lighting and translation. 4.2 Clustering Faces and Poses. Fig. 3a shows examples from a training set of 400 jerky images of two people walking across a cluttered background. We trained a TMG with 4 clusters, 11 horizontal shifts and 11 vertical shifts using 15 iterations of EM after initializing the weights to small, random values. The loop-rich MATLAB script executed in 40 minutes on a 500MHz Pentium processor. Fig. 3b shows the cluster means, which include two sharp representations of each person's face, with the background clutter suppressed. Fig. 3c shows the much blurrier means for a mixture of Gaussians trained using 15 iterations of EM. Fig. 3d shows examples from a training set of 400 jerky images of one person with different poses. We trained a TMG with 5 clusters, 11 horizontal shifts and 11 vertical shifts using 40 iterations of EM. Fig. 3e shows the cluster means, which capture 4 poses and mostly suppress the background clutter. The mean for cluster 4 includes part of the background, but this cluster also has a low mixing proportion of 0.1. A traditional mixture of Gaussians trained using 40 iterations of EM finds blurrier means, as shown in Fig. 3f. The first 4 principal components mostly try to account for lighting and translation, as shown in Fig. 3g. 482 N Jojic and B. J. Frey (d) (e) (f) Figure 4: Modeling handwritten digits. (a) Means and components and (b) the sheared + translated means (dimmed transformations have low probability) for each of 10 TCA models trained on 200 examples of each digit. (c) Means and components of 10 FA models trained on the same data. (d) Digits generated from the 10 TCA models and (e) the 10 FA models. (f) The means for a mixture of 10 Gaussians, a mixture of 10 factor analyzers and a 10-ciuster TMG trained on all 2000 digits. In each case, the best of 10 experiments was selected. 4.3 Modeling Handwritten Digits. We performed both supervised and unsupervised learning experiments on 8 x 8 greyscale versions of 2000 digits from the CEDAR CDROM (Hull, 1994). Although the preprocessed images fit snugly in the 8 x 8 window, there is wide variation in "writing angle" (e.g., the vertical stroke of the 7 is at different angles). So, we produced a set of 29 shearing+translation transformations (see the top row of Fig. 4b) to use in transformed density models. In our supervised learning experiments, we trained one 10-component TCA on each class of digit using 30 iterations of EM. Fig. 4a shows the mean and 10 components for each of the 10 models. The lower 10 rows of images in Fig. 4b show the sheared and translated means. In cases where the transformation probability is below 1%, the image is dimmed. We also trained one lO-component factor analyzer on each class of digit using 30 iterations of EM. The means and components are shown in Fig. 4c. The means found by TCA are sharper and whereas the components found by factor analysis often account for writing angle (e.g., see the components for 7) the components found by TCA tend to account for line thickness and arc size. Fig. 4d and e show digits that were randomly generated from the TCAs and the factor analyzers. Since different components in the factor analyzers account for different stroke angles, the simulated digits often have an extra stroke, whereas digits simulated from the TCAs contain fewer spurious strokes. To test recognition performance, we trained 10-component factor analyzers and TCAs on 200 examples of each digit using 50 iterations of EM. Each set of models used Bayes rule to classify 1000 test patterns and while factor analysis gave an error rate of 3.2%, TCA gave an error rate of only 2.7%. In our unsupervised learning experiments, we fit 10-cluster mixture models to the entire set of 2000 digits to see which models could identify all 10 digits. We tried a mixture of 10 Gaussians, a mixture of 10 factor analyzers and a lO-cluster TMG. In each case, 10 models were trained using 100 iterations of EM and the model with 483 Topographic Transformation as a Discrete Latent Variable the highest likelihood was selected and is shown in Fig. 4f. Compared to the TMG, the first two methods found blurred and repeated classes. After identifying each cluster with its most prevalent class of digit, we found that the first two methods had error rates of 53% and 49%, but the TMG had a much lower error rate of 26%. 5 Summary In many learning applications, we know beforehand that the data includes transformations of an easily specified nature (e.g., shearing of digit images). If a generative density model is learned from the data, the model must extract a model of both the transformations and the more interesting and potentially useful structure. We described a way to add transformation invariance to a generative density model by approximating the transformation manifold with a discrete set of points. This releases the generative model from needing to model the transformations. 5 different types of experiment show that the method is effective and quite efficient. Although the time needed by this method scales exponentially with the dimensionality of the transformation manifold, we believe that it will be useful in many practical applications and that it illustrates what is possible with a generative model that incorporates a latent transformation variable. We are exploring the performance of a faster variational learning method and extending the model to time series. Acknowledgements. We used CITO, NSERC, NSF and Beckman Foundation grants. References C. M. Bishop, M. Svensen and C. K. I. Williams 1998. GTM: The generative topographic mapping. Neural Computation 10:1, 215- 235 . G. E. Hinton, P. Dayan and M. Revow 1997. Modeling the manifolds of images of handwritten digits. IEEE 1rans. on Neural Networks 8, 65- 74. Z. Ghahramani and G. E. Hinton 1997. The EM algorithm for mixtures of factor analyzers. University of Toronto Technical Report CRG-TR-96-1. Available at www.gatsby.ucl.ac.uk/ ... zoubin. R. Golem and I. Cohen 1998. Scanning electron microscope image enhancement. School of Computer and Electrical Engineering project report, Ben-Gurion University. J. J . Hull 1994. A database for handwritten text recognition research. IEEE 1rans. on Pattern Analysis and Machine Intelligence 16:5, 550-554. Y. Le Cun, L. Bottou, Y. Bengio and P. Haffner 1998. Gradient-based learning applied to document recognition. Proceedings of the IEEE 86:11, November, 2278-2324. P. Y. Simard, B. Victorri, Y. Le Cun and J. Denker 1992. Tangent Prop - A formalism for specifying selected invariances in an adaptive network. In Advances in Neural Information Processing Systems 4, Morgan Kaufmann, San Mateo, CA. P. Y. Simard, Y. Le Cun and J. Denker 1993. Efficient pattern recognition using a new transformation distance. In S. J. Hanson, J . D. Cowan and C. L. Giles, Advances in Neural Information Processing Systems 5, Morgan Kaufmann, San Mateo, CA. Appendix: The Sufficient Statistics Found in the E-Step The sufficient statistics for the M-Step are computed in the E-Step using sparse linear algebra during a single pass through the training set. Before making this pass, the following matrices are computed: Ot,c = COV(zlx,y,l,c) = (~;l +G~-?-lGt)-\ (3t,c = COV(ylx,l,c) = (I+A~~;JAc? A~~;lOt.c~;lAc)-l. For each case in the training set, P(c,llxt) is first computed for each combination of c, l, before computing E[ylxt,l, c] = {3l,cA~ ~;;-l [Ot ,cGi-?-lXt - (I-Ot,c~;;-l )I-'c]' E[zlxt, l, c] = I-'c +Ot,cG~-?-l (Xt - Gtl-'c) +Ot~-l Ac{3t,cA~ ~;;-lOt , cG~-?-J (Xt - Gtl-'c), E[(2J-I.'c}:(2J-I.'c) IXt ,l, c] = (E[zlxt ,l,cH.&c}c(E[zlxt ,l,cH.&c)+diag(Ot,c)+diag(Ol , c~;l Ac{3t,cA~ ~;lOt,c), E[(z-l-'c)y'lxt , l, c] = (E[zlxt,l, c]-l-'c)E[Ylxt,l, c]' + Ot,c~;l A c{3t ,c' The expectations needed in (10)-(13) are then computed from P(clxt}E[z - AcylXt, c] = Et P(c, lIXt)(E[zIXt, l, c] - AcE[Ylxt,l, cD, P(clxt}E[(z-l-'cAcy)o(z-l-'c-AcY) IXt, c] = Et P(c, llxt} {E[(z-l-'c)o(z-I-'C>lxt,l, c] +diag(Acl3t,cA~) - 2diag(AcE[(zl-'C>y'lxt, l, +(AcE[Ylxt, e, c])o(AcE[ylxt, e, cD}, E[(Xt-Gtz)o(Xt-Gtz)lxtl Ec,t P(c,lIXt) {(XtGtE[zlxt, l, c]) 0 (Xt - GtE[zIXt, l, c]) + diag(GtOt,cGi) + diag(GtOt,c~;l Ac{3t,cA~~;lOt,cG~)}, P(clxt}E[(z-l-')y'lxt, c] = El P(c,llxt}E[(z-l-')y'lxt, l, c], P(clxt)E[yy'lxt, c] = Et P(c,llxt}{3t,c + Et P(c, llxt)E[ylxt.e, c]E[ylxt.l, c]'. en =
1729 |@word deformed:2 version:1 proportion:2 loading:1 tried:1 covariance:2 tr:2 accommodate:1 tmg:20 golem:2 contains:1 series:1 denoting:1 document:1 written:1 must:1 gurion:1 shape:1 update:1 nebojsa:1 generative:10 selected:3 fewer:1 intelligence:1 ire:1 toronto:1 rc:1 along:1 jac:1 shearing:11 behavior:1 uiuc:1 ol:1 window:1 becomes:1 project:1 notation:1 circuit:1 what:2 transformation:57 certainty:1 exactly:1 classifier:1 uk:1 zl:1 grant:1 before:2 aat:1 frey:5 local:1 engineering:1 io:1 id:1 modulation:1 plus:1 mateo:2 specifying:1 cacy:1 averaged:1 practical:1 digit:19 procedure:1 eth:1 significantly:2 zoubin:1 writing:2 www:3 map:1 williams:1 cluttered:1 identifying:1 rule:1 variation:2 updated:1 imagine:1 element:4 recognition:6 approximated:2 walking:1 dimmed:2 database:1 observed:6 electrical:2 capture:2 initializing:1 highest:1 plc:1 flz:1 trained:12 depend:2 solving:1 algebra:1 blurring:1 tca:13 translated:2 easily:1 joint:4 represented:6 gtm:1 describe:2 effective:1 whose:2 quite:1 ace:4 statistic:3 cov:2 topographic:7 gp:1 jointly:1 noisy:1 ucl:1 product:1 neighboring:1 combining:1 loop:1 mixing:2 degenerate:1 lxt:7 cluster:22 enhancement:1 extending:1 produce:3 generating:1 ben:1 object:4 develop:1 ac:4 svensen:1 pose:5 ij:1 nearest:1 school:1 solves:1 c:1 hull:2 material:1 crg:1 exploring:1 pl:1 mapping:1 electron:4 purpose:1 beckman:2 waterloo:1 successfully:1 gaussian:6 modified:1 release:1 emission:1 pdfs:1 consistently:1 prevalent:1 likelihood:11 pentium:1 brendan:1 cg:4 dayan:1 el:1 inaccurate:1 entire:1 spurious:1 transformed:11 pixel:15 spatial:1 equal:4 look:1 blurrier:2 unsupervised:2 thin:1 acy:2 report:2 randomly:1 lac:1 highly:2 mixture:21 held:1 accurate:1 beforehand:2 facial:1 filled:2 indexed:1 old:1 column:1 modeling:5 classify:1 formalism:1 giles:1 mhz:1 logp:1 cedar:1 uniform:1 mtca:3 ixt:3 thickness:1 scanning:2 combined:2 confident:1 person:3 density:7 ie:1 containing:1 choose:1 simard:4 derivative:3 li:1 account:5 includes:2 blurred:2 vi:1 script:1 try:1 performed:1 responsibility:2 lot:4 linked:1 aca:1 bayes:1 jerky:2 convolutional:2 variance:10 kaufmann:2 identify:1 handwritten:6 accurately:1 disc:3 produced:2 lighting:2 straight:1 processor:1 stroke:4 detector:1 tcas:3 fo:1 couple:2 dimensionality:1 supervised:2 box:1 just:1 horizontal:3 ei:1 nonlinear:4 propagation:1 gtot:2 llxt:5 reveal:1 believe:1 contain:1 jojic:5 imaged:2 nonzero:1 iteratively:1 unfilled:1 illustrated:1 during:1 pdf:1 ay:1 image:40 wise:1 variational:1 ifp:1 rotation:2 cohen:2 exponentially:1 he:1 significant:1 xlf:3 illinois:1 analyzer:12 had:3 add:3 lgt:1 morgan:2 signal:1 ii:2 multiple:1 needing:1 technical:1 faster:1 clx:1 involving:1 multilayer:1 ae:1 expectation:6 iteration:9 represent:3 sometimes:1 microscopy:1 microscope:2 background:4 whereas:3 victorri:1 extra:1 ot:7 tend:1 cowan:1 incorporates:1 integer:1 feedforward:2 bengio:1 fit:2 gave:2 reduce:2 haffner:1 shift:6 matlab:1 useful:4 detailed:2 ylx:1 amount:2 clutter:2 locally:4 lixt:2 nsf:1 estimated:2 per:1 yy:1 discrete:16 tea:1 gtz:2 preprocessed:1 imaging:1 angle:4 uncertainty:2 flx:1 almost:1 appendix:2 fll:1 zlxt:5 speed:1 rendered:1 jio:1 combination:2 smaller:1 slightly:1 em:15 across:1 suppressed:1 wi:1 cun:4 making:1 happens:1 invariant:3 taken:1 equation:1 rans:2 mechanism:1 needed:2 know:1 available:1 gaussians:9 denker:2 original:2 top:1 clustering:2 include:1 graphical:4 ghahramani:2 build:1 approximating:2 move:2 added:1 fa:2 diagonal:6 traditional:1 gradient:1 distance:1 simulated:2 cgi:2 igl:2 manifold:8 assuming:1 modeled:2 index:3 ratio:1 mostly:3 executed:1 sharper:2 greyscale:2 potentially:1 trace:1 suppress:1 design:1 vertical:5 urbana:1 arc:1 november:1 extended:2 incorporated:1 hinton:4 perturbation:1 lb:1 sharp:1 intensity:2 pic:2 introduced:1 specified:2 extensive:1 xle:1 hanson:1 learned:7 zlx:1 usually:1 below:1 pattern:3 cdrom:1 built:1 ndimensional:1 improve:1 extract:1 text:1 acknowledgement:1 tangent:2 marginalizing:2 permutation:2 mixed:1 interesting:1 filtering:3 foundation:1 degree:1 sufficient:3 cd:3 translation:8 row:3 lo:2 summary:1 accounted:1 gl:6 bye:1 institute:1 neighbor:1 wide:1 face:4 emerge:1 cito:1 sparse:5 curve:7 rich:1 adaptive:1 san:2 ec:1 approximate:1 keep:1 discriminative:1 alternatively:1 continuous:3 latent:19 nature:1 robust:1 ca:8 sem:5 bottou:1 diag:10 did:1 ylxt:8 linearly:1 uwaterloo:1 noise:8 repeated:1 fig:23 en:1 gatsby:1 lc:1 sf:1 xl:1 minute:1 bishop:2 xt:11 showing:1 explored:1 zixt:2 illustrates:1 sheared:2 tc:1 gtl:2 nserc:1 aa:3 corresponds:1 ch:2 prop:1 conditional:3 revow:1 operates:1 averaging:4 gte:1 principal:2 pas:2 invariance:6 la:1 deforming:1 perceptrons:1 indicating:1 people:2 frontal:1
797
173
356 USING BACKPROPAGATION WITH TEMPORAL WINDOWS TO LEARN THE DYNAMICS OF THE CMU DIRECT-DRIVE ARM II K. Y. Goldberg and B. A. Pearlmutter School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 ABSTRACT Computing the inverse dynamics of a robot ann is an active area of research in the control literature. We hope to learn the inverse dynamics by training a neural network on the measured response of a physical ann. The input to the network is a temporal window of measured positions; output is a vector of torques. We train the network on data measured from the first two joints of the CMU Direct-Drive Arm II as it moves through a randomly-generated sample of "pick-and-place" trajectories. We then test generalization with a new trajectory and compare its output with the torque measured at the physical arm. The network is shown to generalize with a root mean square error/standard deviation (RMSS) of 0.10. We interpreted the weights of the network in tenns of the velocity and acceleration filters used in conventional control theory. INTRODUCTION Dynamics is the study of forces. The dynamic response of a robot arm relates joint torques to the position, velocity, and acceleration of its links. In order to control an ann at high spee<L it is important to model this interaction. In practice however, the dynamic response is extremely difficult to predict. A dynamic controller for a robot ann is shown in Figure 1. Feedforward torques for a desired trajectory are computed off-line using a model of arm dynamics and applied to the joints at every cycle in an effort to linearize the resulting system. An independent feedback loop at each joint is used to correct remaining errors and compensate for external disturbances. See ([3]) for an introduction to dynamiC control of robot arms. Conventional control theory has difficulty addressing physical effects such as friction [I], backlash [2], torque non-linearity (especially dead zone and saturation) [2], highfrequency dynamics [2], sampling effects [7], and sensor noise [7]. We propose to use a three-layer backpropagation network [4] with sigmoid thresholds to fill the box marked "inverse arm" in Figure 1. We will treat the arm as an unknown non-linear transfer function to be represented by weights of the network.
173 |@word effect:2 especially:1 move:1 correct:1 filter:1 highfrequency:1 pick:1 link:1 generalization:1 pearlmutter:1 predict:1 sigmoid:1 difficult:1 physical:3 unknown:1 mellon:1 hope:1 sensor:1 robot:4 direct:2 tenns:1 torque:5 window:2 saturation:1 ii:2 relates:1 linearity:1 difficulty:1 force:1 interpreted:1 disturbance:1 compensate:1 arm:8 sampling:1 temporal:2 every:1 controller:1 cmu:2 control:5 literature:1 randomly:1 treat:1 feedforward:1 practice:1 backlash:1 backpropagation:2 area:1 desired:1 feedback:1 effort:1 deviation:1 addressing:1 conventional:2 active:1 pittsburgh:1 fill:1 learn:2 off:1 transfer:1 carnegie:1 threshold:1 goldberg:1 pa:1 velocity:2 external:1 dead:1 noise:1 inverse:3 place:1 cycle:1 position:2 root:1 layer:1 dynamic:10 square:1 joint:4 generalize:1 friction:1 extremely:1 represented:1 train:1 trajectory:3 drive:2 marked:1 propose:1 ann:4 acceleration:2 interaction:1 loop:1 response:3 box:1 zone:1 remaining:1 linearize:1 measured:4 school:1
798
1,730
Perceptual Organization Based on Temporal Dynamics Xiuwen Liu and DeLiang L. Wang Department of Computer and Information Science Center for Cognitive Science The Ohio State University, Columbus, OR 43210-1277 Email: {liux, dwang}@cis.ohio-state.edu Abstract A figure-ground segregation network is proposed based on a novel boundary pair representation. Nodes in the network are boundary segments obtained through local grouping. Each node is excitatorily coupled with the neighboring nodes that belong to the same region, and inhibitorily coupled with the corresponding paired node. Gestalt grouping rules are incorporated by modulating connections. The status of a node represents its probability being figural and is updated according to a differential equation. The system solves the figure-ground segregation problem through temporal evolution. Different perceptual phenomena, such as modal and amodal completion, virtual contours, grouping and shape decomposition are then explained through local diffusion. The system eliminates combinatorial optimization and accounts for many psychophysical results with a fixed set of parameters. 1 Introduction Perceptual organization refers to the ability of grouping similar features in sensory data. This, at a minimum, includes the operations of grouping and figure-ground segregation, which refers to the process of determining relative depths of adjacent regions in input data and thus proper occlusion hierarchy. Perceptual organization has been studied extensively and many of the existing approaches [5] [4] [8] [10] [3] start from detecting discontinuities, i.e. edges in the input; one or several configurations are then selected according to certain criteria, for example, non-accidental ness [5] . Those approaches have several disadvantages for perceptual organization. Edges should be localized between regions and an additional ambiguity, the ownership of a boundary segment, is introduced, which is equivalent to figure-ground segregation [7]. Due to that, regional attributions cannot be associated with boundary segments. Furthermore, because each boundary segment can belong to different regions, the potential search space is combinatorial. To overcome some of the problems, we propose a laterally-coupled network based on a boundary-pair representation to resolve figure-ground segregation. An occluding boundary is represented by a pair of boundaries of the two associated regions, and 39 Perceptual Organization Based on Temporal Dynamics (a) Figure 1: On- and off-center cell responses. (a) On- and off-center cells. (b) Input image. (c) On-center cell responses. (d) Off-center cell responses (e) Binarized on- and off-center cell responses, where white regions represent on-center response regions and black off-center regions. initiates a competition between the regions. Each node in the network represents a boundary segment. Regions compete to be figural through boundary-pair competition and figure-ground segregation is resolved through temporal evolution. Gestalt grouping rules are incorporated by modulating coupling strengths between different nodes within a region , which influences the temporal dynamics and determines the percept of the system. Shape decomposition and grouping are then implemented through local diffusion using the results from figur e-ground segregation. 2 Figure-Ground Segregation Network The central problem in perceptual organization is to determine relative depths among regions. As figure reversal occurs in certain circumstances, figure-ground segregation cannot be resolved only based on local attributes. 2.1 The Network Architecture The boundary-pair representation is motivated by on- and off-center cells, shown in Fig. l(a). Fig. l(b) shows an input image and Fig. l(c) and (d) show the on- and off-center responses. Without zero-crossing, we naturally obtain double responses for each occluding boundary, as shown in Fig. l(e). In our boundary-pair representation, each boundary is uniquely associated with a region. In this paper, we obtain closed region boundaries from segmentation and form boundary segments using corners and junctions, which are detected through local corner and junction detectors. A node i in the figure-ground segregation network represents a boundary segment, and Pi represents its probability being figural, which is set to 0.5 initially. Each node is laterally coupled with neighboring nodes on the closed boundary. The connection weight from node i to j, Wij, is 1 and can be modified by T-junctions and local shape information. Each occluding boundary is represented by a pair of boundary segments of the involved regions . For example, in Fig. 2(a), nodes 1 and 5 form a boundary pair , where node 1 belongs to the white region and node 5 belongs to the black region. Node i updates its status by: T dP dt i = ILL "~ Wki(Pk - Pi) kEN(i) + ILJ(l - P i) "~ H(Qli) lEJ (i) + ILB(l B i ) (1) - Pi) exp( - K B Here N(i) is the set of neighboring nodes of i, and ILL , ILJ , and ILB are parameters to determine the influences from lateral connections , junctions , and bias. J(i) is 40 2 X Liu and D. L. Wang ----?--<@r--- 8 6 ---oo{[D--@---- 4 I 3 (a) (b) Figure 2: (a) The figure-ground segregation network for Fig. l(b). Nodes 1, 2, 3 and 4 belong to the white region; nodes 5, 6, 7, and 8 belong to the black region; and nodes 9 and 10, and nodes 11 and 12 belong to the left and right gray regions respectively. Solid lines represent excitatory coupling while dashed lines represent inhibitory connections. (b) Result after surface completion. Left and right gray regions are grouped together. the set of junctions that are associated with i and Q/i is the junction strength of node i of junction l. H(x) is given by H(x) = tanh(j3(x - OJ )), where j3 controls the steepness and OJ is a threshold. In (1), the first term on the right reflects the lateral influences. When nodes are strongly coupled, they are more likely to be in the same status, either figure or background. The second term incorporates junction information. In other words, at a T-junction, segments that vary more smoothly are more likely to be figural. The third term is a bias, where Bi is the bias introduced to simulate human perception. The competition between paired nodes i and j is through normalization based on the assumption that only one of the paired nodes should be figural at a given time: p(Hl) = pt/(P~ + pt) and p(tH) = P~/(pt + P~) t t t J J J t J' 2.2 Incorporation of Gestalt Rules To generate behavior that is consistent with human perception, we incorporate grouping cues and some Gestalt grouping principles. As the network provides a generic model, additional grouping rules can also be incorporated. T-junctions T-junctions provide important cues for determining relative depths [7] [10]. In Williams and Hanson's model [10], T-junctions are imposed as topological constraints. Given aT-junction l, the initial strength for node i that is associated with lis: Q exp( -Ci(i,C(i?/ KT) Ii = 1/2 LkENJ(I) exp( -Ci(k ,c(k?)/ K T ) , where K T is a parameter, N J (l) is a set of all the nodes associated with junction l, c( i) is the other node in N J (l) that belongs to the same region as node i, and Ci(ij) is the angle between segments i and j. Non-accidentalness Non-accidentalness tries to capture the intrinsic relationships among segments [5]. In our system, an additional connection is introduced to node i if it is aligned well with a node j from the same region and j rf. N(i) initially. The connection weight Wij is a function of distance and angle between the involved ending points. This can be viewed as virtual junctions, resulting in virtual contours and conversion of a corner into a T-junction if involved nodes become figural. This corresponds to an organization criterion proposed by Geiger et al [3}. 41 Perceptual Organization Based on Temporal Dynamics Time Time Time Figure 3: Temporal behavior of each node in the network shown in Fig. 2(a). Each plot shows the status of the corresponding node with respect to time. The dashed line is 0.5. Shape information Shape information plays a central role in Gestalt principles and is incorporated through enhancing lateral connections. In this paper, we consider local symmetry. Let j and k be two neighboring nodes of i: Wij = 1 + C exp( -Iaij - akil/ KaJ * exp( -(Lj / Lk + Lk/ Lj - 2)/ K L )), where C, KQ:, and KL are parameters and L j is the length of segment j. Essentially the lateral connections are strengthened when two neighboring segments of i are symmetric. Preferences Human perceptual systems often prefer some organizations over others. Here we incorporated a well-known figure-ground segregation principle, called closeness. In other words, the system prefers filled regions over holes. In current implementation, we set Bi == 1.0 if node i is part of a hole and otherwise Bi == o. 2.3 Temporal Properties of the Network After we construct the figure-ground segregation network, each node is updated according to (1). Fig. 3 shows the temporal behavior of the network shown in Fig. 2(a). The system approaches to a stable solution. For figure-ground segregation, we can binarize the status of each node using threshold 0.5. Thus the system generates the desired percept in a few iterations. The black region occludes other regions while gray regions occlude the white region. For example, P5 is close to 1 and thus segment 5 is figural, and PI is close to 0 and thus segment 1 is in the background. 2.4 Surface Completion After figure-ground segregation is resolved, surface completion and shape decomposition are implemented through diffusion [3]. Each boundary segment is associated with regional attributes such as the average intensity value because its ownership is known. Boundary segments are then grouped into diffusion groups based on similarities of their regional attributes and if they are occluded by common regions. In Fig. 1(b), three diffusion groups are formed, namely, the black region, two gray regions, and the white region. Segments in one diffusion group are diffused simultaneously. For a figural segment, a buffer with a given radius is generated. Within the buffer, the values are fixed to 1 for pixels belonging to the region and 0 otherwise. Now the problem becomes a well-defined mathematical problem. We need to solve 42 X Liu and D. L. Wang (c) Figure 4: Images with virtual contours. In each column, the top shows the input image and the bottom the surface completion result, where completed surfaces are shown according to their relative depths and the bottom one is the projection of all the completed surfaces. (a) Alternate pacman. (b) Reverse-contrast pacman. (c) Kanizsa triangle. (d) Woven square. (e) Double pacman. the heat equation with given boundary conditions. Currently, the heat equation is solved through local diffusion. The results from diffusion are then binarized using threshold 0.5. Fig. 2(b) shows the results for Fig. l(b) after surface completion. Here the two gray regions are grouped together through surface completion because occluded boundaries allow diffusion. The white region becomes the background, which is the entire image. 3 Experimental Results Given an image, the system automatically constructs the network and establishes the connections based on the rules discussed in Section 2.2. For all the experiments shown here, a fixed set of parameters is used. 3.1 Modal and Amodal Completion We first demonstrate that the system can simulate virtual contours and modal completion. Fig. 4 shows the input images and surface completion results. The system correctly solves figure-ground segregation problem and generates the most probable percept. Fig. 4 (a) and (b) show two variations of pacman images [9] [4]. Even though the edges have opposite contrast, the virtual rectangle is vivid. Through boundary-pair representation, our system can handle both cases using the same network. Fig. 4(c) shows a typical virtual image [6] and the system correctly simulates the percept. In Fig. 4( d) [6], the rectangular-like frame is tilted, making the order between the frame and virtual square not well-defined. Our system handles that in the temporal domain. At any given time, the system outputs one of the Perceptual Organization Based on Temporal Dynamics 43 (a) ~'---I (b) (c ) (d) (e) (f) Figure 5: Surface completion results. (a) and (b) Bregman figures [1]. (c) and (d) Surface completion results for (a) and (b). (e) and (f) An image of some groceries and surface completion result. completed surfaces. Due to this, the system can also handle the case in Fig. 4(e) [2], where the percept is bistable, as the order between the two virtual squares is not well defined. Fig. 5(a) and (b) show the well-known Bregman figures [1]. In Fig. 5(a), there is no perceptual grouping and parts of B's remain fragmented. However, when occlusion is introduced as in Fig. 5(b), perceptual grouping is evident and fragments of B's are grouped together. Our results, shown in Fig. 5 (c) and (d), are consistent with the percepts. Fig. 5(e) shows an image of groceries, which is used extensively in [8]. Even though the T-junction at the bottom is locally confusing, our system gives the most plausible result through lateral influences of the other two strong T-junctions. Without search and parameter tuning, our system gives the optimal solution shown in Fig. 5(f). 3.2 Comparison with Existing Approaches As mentioned earlier, at the minimum, figure-groud segregation and grouping need to be addresssed for perceptual organization. Edge-based approaches [4] [10] attempt to solve both problems simultaneously by prefering some configurations over combinatorially many ones according to certain creteria. There are several difficulties common to those approaches. First it cannot account for different human percepts of cases where edge elements are similar. Fig. 5 (a) and (b) are wellknown examples in this regard. Another example is that the edge-only version of Fig. 4( c) does not give rise to a vivid virtual contour as in Fig. 4( c) [6]. To reduce the potential search space, often contrast signs of edges are used as additional contraints [10J. However, both Fig. 4 (a) and (b) give rise to virtual contours despite the opposite edge contrast signs. Essentially based on Fig. 4(b), Grossberg and Mingolla [4] claimed that illusory contours can join edges with different directions of contrast , which does not hold in general. As demonstrated through experiments, our approach does offer a common principle underlying these examples. Our approach shares some similarities with the one by Geiger et al [3]. In both approaches , perceptual organization is solved in two steps. In [3], figure-ground segregation is encoded implicitly in hypotheses which are defined at junction points. Because potential hypotheses are combinatorial, only a few manually chosen ones are tested in their experiments, which is not sufficient for a general computational 44 X Liu and D. L. Wang model. In our approach, by resolving figure-ground segregation, there is no need to define hypotheses explicitly. In both methods, grouping is implemented through diffusion. In [3], "heat" sources for diffusion are given manually for each hypothesis whereas our approach generates "heat" sources automatically using the figure-ground segregation results. Finally, in our approach, local ambiguities can be resolved through lateral connections using temporal dynamics, resulting in robust behavior. To obtain good results for Fig. 5(e), Nitzberg et al [8] need to tune parameters and increase their search space substantially due to the misleading T-junction at the bottom of Fig. 5(e). 4 Conclusion In this paper we have proposed a network for perceptual organization using temporal dynamics. The pair-wise boundary representation resolves the ownership ambiguity inherent in an edge-based representation and is equivalent to a surface representation through diffusion, providing a unified edge- and surface-based representation. Through temporal dynamics, our model allows for interactions among different modules and top-down influences can be incorporated. Acknowledgments Authors would like to thank S. C. Zhu and M. Wu for their valuable discussions. This research is partially supported by an NSF grant (IRI-9423312) and an ONR Young Investigator Award (N00014-96-1-0676) to DLW. References [1] A. S. Bregman, "Asking the 'What for' question in auditory perception," In Perceptual Organization, M. Kubovy and J R. Pomerantz, eds., Lawrence Erlbaum Associates, Publishers, Hillsdale, New Jersey, pp. 99-118, 1981. [2] M. Fahle and G. Palm, "Perceptual rivalry between illusory and real contours," Biological Cybernetics, vol. 66, pp. 1-8, 1991. [3] D. Geiger, H. Pao, and N . Rubin, "Salient and multiple illusory surfaces," In Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 118-124, 1998. [4] S. Grossberg and E . Mingolla, "Neural dynamics of perceptual grouping: textures, boundaries, and emergent segmentations," Perception & Psychophysics, vol. 38, pp. 141-170, 1985. [5] D. G. Lowe, Perceptual Organization and Visual Recognition, Kluwer Academic Publishers, Boston, 1985. [6] G. Kanizsa, Organization in Vision, Praeger, New York, 1979. [7] K. Nakayama, Z. J . He, and S. Shimojo, "Visual surface representation: a critical link between lower-level and higher-level vision," In Visual Cognition, S. M. Kosslyn and D. N. Osherson, eds. , The MIT Press, Cambridge, Massachusetts, vol. 2, pp. 1-70, 1995. [8] M. Nitzberg, D. Mumford, and T. Shiota, Filtering, Segmentation and Depth, Springer-Verlag, New York, 1993. [9] R. Shapley and J. Gordon, "The existence of interpolated illusory contours depends on contrast and spatial separation," In The Perception of Illusory Contours, S. Petry and G. E. Meyer, eds., Springer-Verlag, New York, pp. 109-115, 1987. [10] L. R. Williams and A. R. Hanson, "Perceptual Completion of Occluded Surfaces," Computer Vision and Image Understanding, vol. 64, pp. 1-20, 1996.
1730 |@word version:1 decomposition:3 solid:1 shiota:1 liu:4 configuration:2 fragment:1 initial:1 existing:2 current:1 tilted:1 occludes:1 shape:6 praeger:1 plot:1 update:1 occlude:1 cue:2 selected:1 detecting:1 provides:1 node:37 preference:1 mathematical:1 differential:1 become:1 excitatorily:1 shapley:1 behavior:4 automatically:2 resolve:2 becomes:2 wki:1 underlying:1 what:1 substantially:1 unified:1 temporal:14 binarized:2 laterally:2 control:1 grant:1 local:9 despite:1 black:5 studied:1 bi:3 grossberg:2 acknowledgment:1 projection:1 word:2 refers:2 cannot:3 close:2 influence:5 equivalent:2 imposed:1 demonstrated:1 center:10 attribution:1 williams:2 iri:1 rectangular:1 rule:5 handle:3 variation:1 updated:2 hierarchy:1 play:1 pt:3 hypothesis:4 associate:1 crossing:1 element:1 recognition:2 bottom:4 role:1 module:1 p5:1 wang:4 capture:1 solved:2 region:35 valuable:1 mentioned:1 occluded:3 dynamic:9 woven:1 segment:19 triangle:1 resolved:4 emergent:1 osherson:1 represented:2 jersey:1 heat:4 detected:1 encoded:1 solve:2 plausible:1 otherwise:2 ability:1 propose:1 interaction:1 neighboring:5 aligned:1 kubovy:1 figural:8 competition:3 qli:1 double:2 coupling:2 oo:1 completion:14 figur:1 ij:1 strong:1 solves:2 implemented:3 direction:1 radius:1 attribute:3 human:4 bistable:1 virtual:11 hillsdale:1 probable:1 biological:1 hold:1 ground:19 exp:5 lawrence:1 cognition:1 vary:1 combinatorial:3 tanh:1 currently:1 modulating:2 grouped:4 combinatorially:1 establishes:1 reflects:1 mit:1 modified:1 contrast:6 lj:2 entire:1 initially:2 wij:3 pixel:1 among:3 ill:2 grocery:2 spatial:1 ness:1 psychophysics:1 construct:2 manually:2 represents:4 others:1 ilb:2 gordon:1 inherent:1 few:2 simultaneously:2 occlusion:2 attempt:1 organization:16 kt:1 bregman:3 edge:11 filled:1 desired:1 column:1 earlier:1 amodal:2 asking:1 disadvantage:1 kq:1 erlbaum:1 off:7 together:3 ilj:2 ambiguity:3 central:2 cognitive:1 corner:3 li:1 account:2 potential:3 includes:1 explicitly:1 depends:1 try:1 lowe:1 closed:2 start:1 formed:1 square:3 percept:7 prefering:1 groud:1 cybernetics:1 detector:1 ed:3 email:1 pp:7 involved:3 naturally:1 associated:7 auditory:1 massachusetts:1 illusory:5 dlw:1 segmentation:3 higher:1 dt:1 modal:3 response:7 though:2 strongly:1 furthermore:1 columbus:1 gray:5 nitzberg:2 evolution:2 symmetric:1 white:6 adjacent:1 uniquely:1 pao:1 criterion:2 evident:1 demonstrate:1 pacman:4 image:12 wise:1 ohio:2 novel:1 common:3 belong:5 discussed:1 he:1 kluwer:1 cambridge:1 tuning:1 stable:1 similarity:2 surface:18 deliang:1 belongs:3 reverse:1 wellknown:1 claimed:1 certain:3 buffer:2 n00014:1 verlag:2 onr:1 minimum:2 additional:4 determine:2 dashed:2 ii:1 resolving:1 multiple:1 academic:1 offer:1 award:1 paired:3 j3:2 vision:4 circumstance:1 enhancing:1 essentially:2 iteration:1 represent:3 normalization:1 cell:6 background:3 whereas:1 source:2 publisher:2 eliminates:1 regional:3 simulates:1 incorporates:1 architecture:1 opposite:2 reduce:1 motivated:1 mingolla:2 york:3 prefers:1 tune:1 extensively:2 locally:1 ken:1 generate:1 nsf:1 inhibitory:1 sign:2 correctly:2 vol:4 steepness:1 group:3 salient:1 threshold:3 diffusion:12 rectangle:1 compete:1 angle:2 wu:1 vivid:2 separation:1 geiger:3 prefer:1 confusing:1 kaj:1 accidental:1 topological:1 strength:3 incorporation:1 constraint:1 generates:3 interpolated:1 simulate:2 department:1 palm:1 according:5 alternate:1 belonging:1 remain:1 dwang:1 kosslyn:1 making:1 hl:1 explained:1 segregation:20 equation:3 initiate:1 reversal:1 junction:20 operation:1 generic:1 existence:1 top:2 completed:3 society:1 psychophysical:1 diffused:1 question:1 occurs:1 mumford:1 dp:1 distance:1 thank:1 link:1 lateral:6 binarize:1 length:1 relationship:1 providing:1 rise:2 implementation:1 contraints:1 proper:1 conversion:1 xiuwen:1 incorporated:6 frame:2 kanizsa:2 intensity:1 introduced:4 pair:10 namely:1 kl:1 connection:10 hanson:2 discontinuity:1 perception:5 pattern:1 rf:1 oj:2 critical:1 difficulty:1 zhu:1 misleading:1 lk:2 coupled:5 understanding:1 determining:2 relative:4 filtering:1 localized:1 sufficient:1 consistent:2 rubin:1 principle:4 pi:4 share:1 excitatory:1 supported:1 bias:3 allow:1 pomerantz:1 rivalry:1 inhibitorily:1 fragmented:1 boundary:28 depth:5 overcome:1 ending:1 regard:1 contour:10 sensory:1 author:1 gestalt:5 implicitly:1 status:5 fahle:1 shimojo:1 search:4 robust:1 nakayama:1 symmetry:1 domain:1 pk:1 fig:30 join:1 strengthened:1 meyer:1 perceptual:20 third:1 young:1 down:1 closeness:1 grouping:15 intrinsic:1 ci:4 texture:1 hole:2 boston:1 smoothly:1 likely:2 visual:3 partially:1 springer:2 corresponds:1 determines:1 viewed:1 ownership:3 typical:1 called:1 experimental:1 occluding:3 investigator:1 incorporate:1 tested:1 phenomenon:1
799
1,731
Spectral Cues in Human Sound Localization Craig T. Jin Department of Physiology and Department of Electrical Engineering, Univ. of Sydney, NSW 2006, Australia Anna Corderoy Department of Physiology Univ. of Sydney, NSW 2006, Australia Simon Carlile Department of Physiology and Institute of Biomedical Research Univ. of Sydney, NSW 2006, Australia Andre van Schaik Department of Electrical Engineering, Univ. of Sydney, NSW 2006, Australia Abstract The differential contribution of the monaural and interaural spectral cues to human sound localization was examined using a combined psychophysical and analytical approach. The cues to a sound's location were correlated on an individual basis with the human localization responses to a variety of spectrally manipulated sounds. The spectral cues derive from the acoustical filtering of an individual's auditory periphery which is characterized by the measured head-related transfer functions (HRTFs). Auditory localization performance was determined in virtual auditory space (VAS). Psychoacoustical experiments were conducted in which the amplitude spectra of the sound stimulus was varied independentlyat each ear while preserving the normal timing cues, an impossibility in the free-field environment. Virtual auditory noise stimuli were generated over earphones for a specified target direction such that there was a "false" flat spectrum at the left eardrum. Using the subject's HRTFs, the sound spectrum at the right eardrum was then adjusted so that either the true right monaural spectral cue or the true interaural spectral cue was preserved. All subjects showed systematic mislocalizations in both the true right and true interaural spectral conditions which was absent in their control localization performance. The analysis of the different cues along with the subjects' localization responses suggests there are significant differences in the use of the monaural and interaural spectral cues and that the auditory system's reliance on the spectral cues varies with the sound condition. 1 Introduction Humans are remarkably accurate in their ability to localize transient, broadband noise, an ability with obvious evolutionary advantages. The study of human auditory localization has a considerable and rich history (recent review [I]) which demonstrates that there are three general classes of acoustical cues involved in the localization process: (1) interaural time differences, ITDs; (2) interaurallevel differences, ILDs; and (3) the spectral cues resulting Spectral Cues in Human Sound Localization 769 from the auditory periphery. It is generally accepted that for humans, the lTD and ILD cues only specify the location of the sound source to within a "cone of confusion" [I], i.e., a locus of points approximating the surface of a cone symmetric with respect to the interaural axis. It remains, therefore, for the localization system to extract a more precise sound source location from the spectral cues. The utilization of the outer ear spectral cues during sound localization has been analyzed both as a statistical estimation problem, (e.g., [2]) and as optimization problem, often using neural networks, (e.g., [3]). Such computational models show that sufficient localization information is provided by the spectral cues to resolve the cone of confusion ambiguity which corroborates the psychoacoustical evidence. Furthermore, it is commonly argued that the interaural spectral cue, because of its natural robustness to level and spectral variations, has advantages over the monaural spectral cues alone. Despite these observations, there is still considerable contention as to the relative role or contribution of the monaural versus the interaural spectral cues. In this study, each subject's spectral cues were characterized by measuring their head related transfer functions (HRTFs) for 393 evenly distributed positions in space. Measurements were carried out in an anechoic chamber and were made for both ears simultaneously using a "blocked ear" technique [I]. Sounds filtered with the HRTFs and played over earphones, which bypass the acoustical filtering of the outer ear, result in the illusion of free-field sounds which is known as virtual auditory space (VAS). The HRTFs were used to generate virtual sound sources in which the spectral cues were manipulated systematically. The recorded HRTFs along with the Glasberg and Moore cochlear model [4] were also used to generate neural excitation patterns (frequency representations of the sound stimulus within the auditory nerve) which were used to estimate the different cues available to the subject during the localization process. Using this analysis, the interaural spectral cue was characterized and the different localization cues have been correlated with each subjects' VAS localization responses. 2 VAS Sound Localization The sound localization performance of four normal hearing subjects was examined in VAS using broadband white noise (300 - 14 000 Hz). The stimuli were filtered under three differing spectral conditions. (1) control: stimuli were filtered with spectrally correct left and right ear HRTFs for a given target location, (2) veridical interaural: stimuli at the left ear were made spectrally flat with an appropriate dB sound level for the given target location, while the stimuli at the right ear were spectrally shaped to preserve the correct interaural spectrum, (3) veridical right monaural: stimuli at the left ear were spectrally flat as in the second condition, while the stimuli at the right ear were filtered with the correct HRTF for the given target location, resulting in an inappropriate interaural spectral difference. For each condition, a minimum-phase filter spectral approximation was made and the interaural time difference was modeled as an all-pass delay [5]. Sounds were presented at approximately 70 dB SPL and with duration 150 ms (with 10 ms raised-cosine onset and offset ramps). Each subject performed five trials at each of 76 test positions for each stimulus condition. Detailed sound localization methods can be found in [1]. A short summary is presented below. 2.1 Sound Localization Task The human localization experiments were carried out in a darkened anechoic chamber. Virtual auditory sound stimuli were presented using earphones (ER-2, Etymotic Research, with a flat frequency response, within 3 dB, between 200-16 000 Hz). The perceived location of the virtual sound source was indicated by the subject pointing hislher nose in 770 C. T. lin, A. Corderoy, S. Carlile and A. v. Schaik the direction of the perceived source. The subject's head orientation and position were monitored using an electromagnetic sensor system (Polhemus, Inc.). 2.2 Human Sound Localization Performance The sound localization performance of two subjects in the three different stimulus conditions are shown in Figure 1. The pooled data across 76 locations and five trials is presented for both the left (L) and right (R) hemispheres of space from the viewpoint of an outside observer. The target location is shown by a cross and the centroid of the subjects responses for each location is shown by a black dot with the standard deviation indicated by an ellipse. Front-back confusions are plotted, although, they were removed for calculating the standard deviations. The subjects localized the control broadband sounds accurately (Figure 1a). In contrast, the subjects demonstrated systematic mislocalizations for both the veridical interaural and veridical monaural spectral conditions (Figures I b,c). There is clear pulling of the localization responses to particular regions of space with evident intersubject variations. (8) Subject 1: Broadband Control Subject 2: Broadband Control Ellipse: Standard Deviation L:i;;:~ses ~JC~~ ~~i.iif. e'??.R ,':.,;':::::::.><" (b) Subject 1: Veridicallnteraural Spectrum L R Subject 2: Veridical Interaural Spectrum L ,.c._:.. . . ~...:. . .." ~I~;> (c) Le Subject 1: Veridical Right Monaural Spectrum '... ~l~. :.:. - ' ., " _ ? ....-.L . .;.' ... .,. ~~.j;~;" ":':"$.. '~-, . ' ?' /...., R " ~ .~ ? . .?. ~...... .t ' ~ i ~: ? ......~?.;;.:J;...;,1"" ',. .-. ';'-"'" , ' ;::? Subject 2: Veridical Right Monaural Spectrum L~~R V~ Figure 1: Localization performance for two subjects in the three sound conditions: (a) control broadband; (b) veridical interaural; (c) veridical monaural. See text for details. 3 Extraction of Acoustical Cues With accurate measurements of each individual's outer ear filtering, the different acoustical cues can be compared with human localization performance on an individual basis. In order to extract the different acoustical cues in a biologically plausible manner, a model of peripheral auditory processing was used. A virtual source sound stimulus was prepared as described in Secion 2 for a particular target location. The stimulus was then filtered using a cochlear model based on the work of Glasberg and Moore [4]. This cochlear model consisted of a set of modified rounded-exponential auditory filters. The width and shape of the auditory filters change as a function of frequency (and sound level) in a manner Spectral Cues in Human Sound Localization 771 consistent with the known physiological and psychophysical data. These filters were logarithmically spaced on the frequency axis with a total of 200 filters between 300 Hz and 14 kHz. The cochlea's compressive non-linearity was modelled mathematically using a logarithmic function. Thus the logarithm of the output energy ofa given filter indicated the amount of neural activity in that particular cochlear channel. The relative activity across the different cochlear channels was representative of the neural excitation pattern (EP) along the auditory nerve and it is from this excitation pattern that the different spectral cues were estimated. For a given location, the left and right EPs themselves represent the monaural spectral cues. The difference in the total energy (calculated as the area under the curve) between the left and right EPs was taken as a measure of the interaural level difference and the interaural spectral shape cue was calculated as the difference between the left and right EPs. The fourth cue, interaural time difference, is a measure of the time lag between the signal in one ear as compared to the other and depends principally upon the geometrical relationship between the sound source and the human subject. This time delay was calculated using the acoustical impulse response for both ears as measured during the HRTF recordings. 4 Correlation of Cues and Location For each stimulus condition and location, the acoustical cues were calculated as described above for all 393 HRTF locations. Locations at which a given cue correlates well with the stimulus cue for a particular target location were taken as analytical predictions of the subject's response locations according to that cue. As the spectral content of the signal is varied, the cue(s) available may strongly match the cue(s) normally arising from locations other than the target location. Therefore the aim of this analysis is to establish for which locations and stimulus conditions a given response most correlated with a particular cue. The following analyses (using a Matlab toolbox developed by the authors) hinge upon the calculation of "cue correlation values". To a large extent, these calculations follow the examples described by [6] and are briefly described here. For each stimulus condition and target location, the subject performed five localizations trials. For each of the subject's five response locations, each possible cue was estimated (Section 3) assuming a flat-spectrum broadband Gaussian white noise as the stimulus. A mathematical quantity was then calculated which would give a measure of the similarity of the response location cues with the corresponding stimulus cues. The method of calculation depended on the cue and several alternative methods were tried. Generally, for a given cue, these different methods demonstrated the same basic pattern and the term "cue correlation value" has been given to the mathematical quantity that was used to measure cue similarity. The methods are as follows. For the ITO cue, the negative of the absolute value of the difference between possible response location ITDs and the stimulus ITO was used as the ITO cue correlation values (the more positive a value, the higher its correlation). The ILD cue correlation value was calculated in a similar fashion. The cue correlation values for the left and right monaural spectral cues (in this case, the shape of the neural excitation pattern) was calculated by taking the difference between the stimulus EP and the possible response location EPs and then summing across frequency the variation of this difference about its mean value. For the interaural spectral cue, the vector difference between the left and right EPs was calculated for both the stimulus and the possible response locations. The dot product between the stimulus and the possible response location vectors gave the ISO cue correlation values. The cue correlation values were normalized in order to facilitate meaningful comparisons across the different acoustical cues. Following Middlebrooks [6], a "z-score normalized" cue value, for each response location corresponding to a given target location, was obtained by subtracting the mean correlation value (across all possible locations) and dividing by the 772 C. T. Jin, A. Corderoy, S. Carlile and A. v. Schaik standard deviation. For these new cue values, termed the cue z-score values, a score of 1.0 or greater indicates p'oocl correlation. 5 Relationship between the ISD and the Cone-of-Confusion The distribution of a given cue's z-score values around the sphere of space surrounding the subject reveals the spatial directions for that cue that correlate best with the given stimulus and target location being examined. An examination of the interaural spectral cue indicated that, unlike the other cues, the range of its cue z-score variation was relatively restricted on the ipsilateral hemisphere of space relative to the sound stimulus (values on the ipsilateral side were approximately 1.0, those on the contralateral side, -1.0). This was the first indication of the more moderate variation of the ISO cue across space as compared with the monaural spectral cues. Closer examination of the ISO cue revealed more detailed variational properties. In order to facilitate meaningful comparisons with the other cues, the ISO cue z-score values were adjusted such that all negative values (i.e., those values at locations generally contralateral to the stimulus) were set to 0.0 and the cue z-score values recalculated. The spatial distribution of the rescaled ISO cue z-score values, as compared with the cue z-score values for the other cues, is shown in Figure 2. The cone of confusion described by the ITO and ILO is clearly evident (Fig. 2a,b) and it can be seen that the ISO cue is closely aligned with these cues (Fig. 2c). Furthermore, the ISO cue demonstrates significant asymmetry along the front-back dimensions. These novel observations demonstrate that while previous work [3, 2] indicates that the ISO cue provides sufficient information to determine a sound's 10eation exactly along the cone of confusion, the variation of the cue z-score values along the cone is substantially less than that for the monaural spectral cues (Fig. 2d), suggest ;:-:3 perhaps that this acts to make the monaural spectral cue a more salient cue. (8) Interaural Time Difference (c) 40 40 1.5 o o c: .40 ,g Interaural Spectrum -40 0 90 180 0 90 180 ~ (b) Interaurai Level Difference (d) Ipsilateral Monaural Spectrum iIi 40 40 o o -40 o -40 90 180 0 90 180 1.1 Azimuth Figure 2: Spatial plot of the cue z-score values for a single target location (46 0 azimuth, 20 0 elevation) and broadband sound condition. Gray-scale color values indicate the cue's correlation in different spatial directions with the stimulus cue at the target location. (Zscore values for the ISO cue have been rescaled, see text.) 6 Analysis of Subjects Responses using Cue Z-score Values A given cue's z-score values for the subject's responses across all 76 test locations and five trials were averaged. The mean and standard deviation are presented in a bar graph (Fig. 3). The subjects' response locations correlate highly with the ITO and ILD cue and Spectral Cues in Human Sound Localization 773 the standard deviation of the correlation was low (Fig. 3a,b). In other words, subjects' responses stayed on the cone of confusion of the target location. A similar analysis of the more restricted, rescaled version of the interaural spectral cue shows that despite the spectral manipulations and systematic mislocalizations, subject's were responding to locations which were highly correlated with the interaural spectral cue (Fig. 3c). The bar graphs for the monaural spectral cues ipsilateral and contralateral to the target location show the average correlation of the subjects' responses with these cues varied considerably with the stimulus condition (Fig. 3d-g) and to a lesser extent across subjects. 2 rn (ij 2 ~ 1 If 0 > 8 N Q) :::l U (e) Right Contralateral Spectrum 2 Control Veridical VeridiCal Broadband Interaural Right Mona....1 1 Q) :::l (d) Left Contralateral Spectrum 2 ILD =~= o Control VeridiCal VeridiCal Broadband Interaural Right Monaural 2 (f) Left Ipsilateral Spectrum o Control Veridical Veridical Broadband InterauraJ Right Monaural (g) Right Ipsilateral Spectrum 2 ISD o o Control Veridical Veridical Broadband Interaural Right Monaural Control VeridiCal VeridiCal Broadband Interaurai Right Monaural o Control V9!idical VeridiCal Broadband Interaural Right Monaural Control Veridical Veridical Broadband Interaural Right Monaural Figure 3: Correlation of the four subjects' (indicated by different gray bars) localization responses with the different acoustical cues for each stimulus condition. The bar heights indicates the mean cue z-score value, while the error bars indicate standard deviation. 7 Spatial Plots of Correlation Regions As the localization responses tended to lie along the cone of confusion, the relative importance of the spectral cues along the cone of confusion was examined. The correlation values for the spectral cues associated with the subjects' responses were recalculated as a z-score value using only the distribution of values restricted to the cone of confusion. This demonstrates whether the spectral cues associated with the subjects' response locations were better correlated with the stimulus cues, than for any random location on the cone of confusion. Spatial plots of the recalculated response cue z-score values for the spectral cues of one subject (similar trends across subjects), obtained for each stimulus location and across the three different sound conditions, is shown in Figure 4. Spatial regions of both high and low correlation are evident that vary with the stimulus spectrum. The z-score values for the ISD cue shows greater bilateral correlation across space in the veridical interaural condition (Fig. 4d) than for the veridical monaural condition (Fig. 4g), while the right monaural spectral cue demonstrates higher correlation in the right hemisphere of space for the veridical monaural condition (Fig. 4i) as opposed to the veridical interaural condition (Fig. 4t). This result (although not surprising) demonstrates that the auditory system is extracting cues to source location in a manner dependent on the input sound spectrum and in a manner consistent with the spectral infonnation available in the sound spectrum. Figures 4e,h clearly demonstrate that the flat sound spectrum in the left ear was strongly correlated with and influenced the subject's localization judgements for specific regions of space. 774 C. T. lin, A. Corderoy, S. Carlile and A. v. Schaik Broadband Veridical Interaural Veridical Monaural (a) (d) (9) ? a 180 Ol- (e) 40 40 ~ a a iIi -40 ?180 -40 180 ?180 a ". (1) 40 . '~'t;' ,1 ?180 ~ a E -=:2 Ol- -IalCl. a U) 180 4 180 c: -co 0u 1 >- Ol 0 ~ Ol 8 c: I/) 0 , u 0.5 N 0 40 Ol::J Ol () .~ a g (i) f ., ::J (h) c: .Q .~ .5 ?40 ?180 15 o 180 Azimuth Figure 4: Spatial plot of the spectral cue z-score values for one subject's localization responses across the three different sound conditions. 8 Conclusions The correlation of human sound localization responses with the available acoustical cues across three spectrally.. different sound conditions has provided insights into the human auditory system and its integration of cues to produce a coherent percept of spatial location. These data suggest an interrelationship between the interaural spectral cue and the cone of confusion. The ISO cue is front-back asymmetrical along the cone and its cue correlation values vary more moderately as a function of space than those of the monaural spectral cues. These data shed light on the relative role and importance of the interaural and monaural spectral cues. Acknowledgments This research was supported by the ARC, NHMRC, and Dora Lush Scholarship to CJ. References [1] S. Carlile, Virtual auditory space: Generation and applications. New York: Chapman and Hall, 1996. [2] R. O. Duda, "Elevation dependence of the interaural transfer function," in Binaural and spatial hearing in real and virtual environments (R. H. Gilkey and T. R. Anderson, eds.), ch. 3, pp. 4975, Mahwah, New Jersey: Lawrence Erlbaum Associates, 1997. [3] J. A. Janko, T. R. Anderson, and R. H. Gilkey, "Using neural networks to evaluate the viability of monaural and interaural cues for sound localization," in Binaural and Spatial Hearing in real and virtual environments (R. H. Gilkey and T. R. Anderson, eds.), ch. 26, pp. 557-570, Mahwah, New Jersey: Lawrence Erlbaum Associates, 1997. [4] B. Glasberg and B. Moore, "Derivation of auditory filter shapes from notched-noise data," Hearing Research, vol. 47, no. 1-2, pp. 103-138, 1990. [5] F. Wightman and D. Kistler, "The dominant role oflow-frequency interaural time differences in sound localization," 1. Acoust. Soc. Am., vol. 91, no. 3, pp. 1648-1661, 1992. [6] J. Middlebrooks, "Narrow-band sound localization related to external ear acoustics," 1. Acoust. Soc. Am., vol. 92, no. 5, pp. 2607-2624, 1992.
1731 |@word trial:4 version:1 briefly:1 judgement:1 duda:1 tried:1 nsw:4 score:18 surprising:1 shape:4 plot:4 alone:1 cue:115 iso:10 short:1 schaik:4 filtered:5 provides:1 location:45 five:5 height:1 mathematical:2 along:9 differential:1 interaural:38 manner:4 themselves:1 ol:6 resolve:1 inappropriate:1 provided:2 linearity:1 substantially:1 spectrally:6 developed:1 compressive:1 differing:1 acoust:2 act:1 ofa:1 shed:1 corderoy:4 exactly:1 demonstrates:5 middlebrooks:2 control:13 utilization:1 normally:1 veridical:28 positive:1 engineering:2 timing:1 depended:1 despite:2 approximately:2 black:1 examined:4 suggests:1 co:1 range:1 averaged:1 acknowledgment:1 illusion:1 area:1 physiology:3 word:1 suggest:2 demonstrated:2 duration:1 insight:1 variation:6 target:15 associate:2 logarithmically:1 trend:1 ep:2 role:3 electrical:2 region:4 removed:1 rescaled:3 environment:3 moderately:1 localization:36 upon:2 basis:2 binaural:2 jersey:2 surrounding:1 derivation:1 univ:4 outside:1 lag:1 plausible:1 ramp:1 s:1 ability:2 advantage:2 indication:1 analytical:2 subtracting:1 product:1 aligned:1 anechoic:2 asymmetry:1 produce:1 derive:1 measured:2 ij:1 intersubject:1 sydney:4 dividing:1 soc:2 indicate:2 direction:4 closely:1 correct:3 filter:7 human:15 australia:4 transient:1 kistler:1 virtual:10 argued:1 notched:1 stayed:1 electromagnetic:1 elevation:2 mathematically:1 adjusted:2 around:1 hall:1 normal:2 lawrence:2 recalculated:3 pointing:1 vary:2 perceived:2 estimation:1 infonnation:1 clearly:2 sensor:1 gaussian:1 aim:1 modified:1 indicates:3 impossibility:1 contrast:1 centroid:1 am:2 dependent:1 orientation:1 raised:1 spatial:11 integration:1 field:2 shaped:1 extraction:1 chapman:1 iif:1 stimulus:33 manipulated:2 simultaneously:1 preserve:1 individual:4 phase:1 highly:2 analyzed:1 light:1 ilo:1 accurate:2 closer:1 logarithm:1 plotted:1 measuring:1 hearing:4 deviation:7 contralateral:5 oflow:1 delay:2 conducted:1 azimuth:3 front:3 erlbaum:2 varies:1 considerably:1 combined:1 lush:1 systematic:3 rounded:1 ambiguity:1 ear:15 recorded:1 opposed:1 external:1 gilkey:3 pooled:1 inc:1 jc:1 onset:1 depends:1 performed:2 bilateral:1 observer:1 simon:1 contribution:2 v9:1 percept:1 spaced:1 modelled:1 accurately:1 craig:1 history:1 influenced:1 tended:1 andre:1 ed:2 energy:2 frequency:6 involved:1 pp:5 obvious:1 associated:2 monitored:1 auditory:18 color:1 cj:1 amplitude:1 back:3 nerve:2 higher:2 follow:1 response:28 specify:1 strongly:2 anderson:3 furthermore:2 biomedical:1 correlation:22 ild:4 indicated:5 impulse:1 gray:2 perhaps:1 pulling:1 facilitate:2 consisted:1 true:4 asymmetrical:1 normalized:2 symmetric:1 moore:3 white:2 during:3 width:1 excitation:4 cosine:1 m:2 evident:3 demonstrate:2 confusion:12 interrelationship:1 geometrical:1 variational:1 contention:1 novel:1 khz:1 eps:5 significant:2 measurement:2 blocked:1 dot:2 similarity:2 surface:1 dominant:1 showed:1 recent:1 hemisphere:3 moderate:1 eardrum:2 periphery:2 termed:1 manipulation:1 psychoacoustical:2 preserving:1 mislocalizations:3 minimum:1 greater:2 seen:1 determine:1 signal:2 sound:45 earphone:3 match:1 characterized:3 calculation:3 cross:1 sphere:1 lin:2 va:5 prediction:1 basic:1 cochlea:1 represent:1 preserved:1 janko:1 remarkably:1 source:8 unlike:1 itds:2 subject:40 hz:3 recording:1 db:3 extracting:1 revealed:1 iii:2 viability:1 variety:1 gave:1 lesser:1 absent:1 whether:1 ltd:1 york:1 matlab:1 generally:3 detailed:2 clear:1 amount:1 prepared:1 band:1 generate:2 estimated:2 arising:1 ipsilateral:6 vol:3 four:2 reliance:1 salient:1 localize:1 ilds:1 isd:3 graph:2 cone:14 fourth:1 spl:1 played:1 activity:2 flat:6 relatively:1 department:5 according:1 peripheral:1 across:13 secion:1 biologically:1 restricted:3 principally:1 taken:2 interaurallevel:1 remains:1 locus:1 nose:1 available:4 appropriate:1 spectral:50 chamber:2 dora:1 alternative:1 robustness:1 responding:1 carlile:5 hinge:1 calculating:1 scholarship:1 ellipse:2 approximating:1 establish:1 psychophysical:2 quantity:2 glasberg:3 dependence:1 evolutionary:1 darkened:1 outer:3 evenly:1 acoustical:11 cochlear:5 extent:2 assuming:1 modeled:1 relationship:2 negative:2 observation:2 arc:1 jin:2 head:3 precise:1 rn:1 varied:3 monaural:30 mona:1 specified:1 toolbox:1 acoustic:1 coherent:1 narrow:1 bar:5 below:1 pattern:5 natural:1 examination:2 hrtfs:7 axis:2 carried:2 hrtf:3 extract:2 text:2 review:1 relative:5 generation:1 filtering:3 versus:1 localized:1 sufficient:2 consistent:2 viewpoint:1 systematically:1 bypass:1 polhemus:1 summary:1 supported:1 free:2 side:2 institute:1 taking:1 absolute:1 van:1 distributed:1 curve:1 calculated:8 dimension:1 rich:1 author:1 commonly:1 made:3 correlate:3 reveals:1 summing:1 corroborates:1 spectrum:19 channel:2 transfer:3 wightman:1 anna:1 noise:5 mahwah:2 fig:11 broadband:16 representative:1 fashion:1 position:3 exponential:1 lie:1 hislher:1 ito:5 specific:1 er:1 offset:1 physiological:1 evidence:1 false:1 importance:2 logarithmic:1 ch:2 considerable:2 change:1 content:1 determined:1 total:2 pas:1 accepted:1 meaningful:2 evaluate:1 correlated:6